var/home/core/zuul-output/0000755000175000017500000000000015140111447014523 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015140135627015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000443175015140135537020270 0ustar corecore_ikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB…2|w6b}Wߟ/nm͊wqɻlOxN_3~𒆷7̗8zTY\].f}嗷ovϷw_>on3cvX~egQBeH,nWb m/m}*L~AzHev_uαHJ2E$(Ͽ|/+k*z>p R⥑gF)49)(oՈ7_k0m^p9PneQn͂YEeeɹ ^ʙ|ʕ0MۂAraZR}@E1%]˜(O)X(6I;Ff"mcI۫d@FNsdxό?2$&tg*Y%\ߘfDP'F%Ab*d@e˛H,љ:72 2ƴ40tr>PYD'vt'oI¢w}o٬owko%gQ(%t#NL֜ eh&Ƨ,RH 4*,!SD 1Ed_wkxdL3F;/u7Taqu5Ոӄp\2dd$YLYG(#?%U?hB\;ErE& SOZXHBWy|i$Z~hal\t2Hgb*t--ߖ|Hp(-J C?>:zR{܃ lM6_OފߍO1T~̝G?ƥF%QV5pDVHwԡ/.2h{qۀK8yUOdssdMvw`21ɻ]/ƛ"@8(PN_,_0;o_x+Vy<h\dN9:bġ7 -Pwȹl;M@n̞Qj_P\ Q]GcPN;e7Vtś98m1<:|a+.:a4nՒ,]LF0);I$>ga5"f[B[fhT/ɾgm\Sj#3hEEH*Nf äE@O0~y[쾋t=iYhșC 5ܩa!ǛfGtzz*з 55E9Fa?Zk80ݞN|:AОNo;Ⱦzu\0Ac/T%;m ~S`#u.Џ1qNp&gK60nqtƅ": C@!P q]G0,d%1}Uhs;H?)M"뛲@.Cs*H _0:P.B_#'HqZY PTUJ2dic3w ?YQgpa` Z_0΁?kMPc_Ԝ*΄Bs`kmJ?t 53@հ1hr}=5t;nt 9:I_|AאM'NO;uD,z҄R K&Nh c{A`?2ZҘ[a-0V&2D[d#L6l\Jk}8gf) afs'oIf'mf\>UxR ks J)'u4iLaNIc2qdNA&aLQVD R0*06V۽棬mpھ*V I{a 0Ҟҝ>Ϗ ,ȓw`Ȅ/2Zjǽ}W4D)3N*[kPF =trSE *b9ē7$ M_8.Ç"q ChCMAgSdL0#W+CUu"k"圀̲F9,,&h'ZJz4U\d +( 7EqڏuC+]CEF 8'9@OVvnNbm: X„RDXfיa }fqG*YƩ{P0K=( $hC=h2@M+ `@P4Re]1he}k|]eO,v^ȹ [=zX[tꆯI7c<ۃ'B쿫dIc*Qqk&60XdGY!D ' @{!b4ִ s Exb 5dKߤKߒ'&YILұ4q6y{&G`%$8Tt ȥ#5vGVO2Қ;m#NS8}d0Q?zLV3\LuOx:,|$;rVauNjk-ؘPꐤ`FD'JɻXC&{>.}y7Z,).Y톯h7n%PAUË?/,z_jx܍>М>ӗom$rۇnu~Y݇̇TIwӜ'}׃nxuoỴRZ&Yzbm ]) %1(Y^9{q"4e?x+ [Vz;E|d1&ږ/0-Vb=SSO|k1A[|gbͧɇد;:X:@;afU=Sru CK >Y%LwM*t{zƝ$;ȾjHim @tBODɆj>0st\t@HTu( v e`H*1aK`3CmF1K>*Mk{_'֜dN${OT-n,'}6ȴ .#Sqη9]5zoX#ZVOy4%-Lq6dACYm*H@:FUф(vcD%F"i ' VVdmcOTKpwq.M?m12N[=tuw}opYG]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%Egl1$9  ֲQ$'dJVE%mT{z`R$77.N|b>harNJ(Bň0ae3V#b,PY0TEu1L/]MTB4$`H6NI\nbǛ*AyA\(u|@ [h-,j7gDTÎ4oWJ$j!fJrH_HI\:U}UE$J @ٚeZE0(8Ŗ%?O"d.wEр%}5zWˬQOS)ZbF p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))OrT 9KaͶ{߰+ednU$YD',jߎmc\cN#0"",tw>]rύW -a]hx&gs7,6BxzxօoFMA['҉F=NGD4sTq1HPld=Q,D5I9qGGIi'ޗX7w{:.B)ƸXM3“QLL FN+\r]IrfWoۢ,"~BRlnbu;Bx`C:gev\g7 6܋ hH+P5co.Q/cC@.sABC{تI=;̶n2¦l:d9 ΋pydqZrS6U A~@Ve Ȇ*d96 FuQ==ƈkmb]/sl pVpvUEM.wtZ6]( 1aVf~xN>/!~x]y7D7@u邗`unn_ư-a9tz*෮9E=զ,i:xI_ˌvg&*5F>#q * CƂ lu" yo6"3껝I~flQ~y'H/jh7hM=~ ֟DɈP͐b7j{ *bmc`  SgkmOl7^~xAE,Pmqs;l};Щ۸l?28Ć%ZU.]5`s=r&v2FaUM 6E]_vE P / څZg`9r| 5G;`.4&XkĴp 6l0Cз5O[{B-bC\/`m(9Ay@/ڠ --i!M5mjozEƨ||Yt,=d#uЇ > l]չoݴmqV".lCqBѷ /![auPmpnEjus]2{2#b'$?T3{k>h+@]*pp桸]%nĴFԨlu |VXnq#r:kg_Q1,MNi˰ 7#`VCpᇽmpM+tWuk0 q } 5 ¶]fXEj@5JcU_b@JS`wYmJ gEk2'0/> unKs^C6B WEt7M'#|kf1:X l]ABC {kanW{ 6%#mϸ.6p5k0C5PdKB g~E#zmxfU S 3_`wRY}@ŹBz²?mК/mm}m"Gy4dl\)cb<>O0BďJrDd\TDFMEr~q#i}$ysi*TRJbż31"qrk3S/0g:PSX~߃ʟ~㍖›f!OI1R~-6͘!?/Vvot{? ǣro_nն-2n6 Ym^]IL'M+;U t>x]U5g B(, qA9r;$IN&CM(F+ hGI~Q<웰[,_ qnriY]3_P${,<\V}7T g6Zapto}PhS/b&X0$Ba{J @xS}NEij]Qexx*lJF#+L@-ՑQz֬]")JC])"GGzpFL`1CS$Ǥ46i*#zN9tT :<XK*ɤ{ U܋N5 l͖h"褁l^=UF^BcAw`g*7R(#ғ [K&#Mp'XގL=s5^:z7y0^} "NqK$2$ Ri ?2,ᙌEK@-V3ʱd:/4Kwm2uZm8pnglVj!p2֬uT[QyB402|2d5K: `Bcz|YץF .Jg< ƜINs:b zĄu3=Az4 u5'og^s7`Rzu-anOIq;6z( rx„2Hi{(2HFE?*w*hy4ޙM^٫wF(p]EwQzr*! 5F XrO7E[!gJ^.a&HߣaaQÝ$_vyz4}0!yܒ栒޹a% Ŋ X!cJ!A\ ?E\R1 q/rJjd A4y4c+bQ̘TT!kw/nb͵FcRG0xeO sw5TV12R7<OG1cjShGg/5TbW > ]~W9dNiee$V[\[Qp-&u~a+3~;xQFFW>='ǣC~방u)т48ZdH;j a]`bGԹ#qiP(yڤ~dO@wA[Vz/$NW\F?H4kX6)F*1*(eJAaݡ krqB}q^fn 8y7P  GRޠkQn>eqQntq"Occ°NRjg#qSn02DŔw:= 5k6JvUL*t*\!j=Ã˼)"޶*̈́\)F^jH Gr%ie A>;^u'}8H0]+ES,n?UU{ x~ʓOy_>??~I&9s$ $"+ 쩹& h'|?1 ؚ~1%dk􂗡Ƭd 8AIڲhn?le\ZO1O`E;\9n@VFB0℃= OWv/XV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKd9A7ޕ6ql?N/e1N2i%e"xm뻱~0GBeFO0ޑ]w(zM6j\v00ׅYɓHڦd%NzT@gID!EL2$%Ӧ{(gL pWkn\SDKIIKWi^9)N?[tLjV}}O͌:&c!JC{J` nKlȉW$)YLE%I:/8)*H|]}\E$V*#(G;3U-;q7KǰfξC?ke`~UK mtIC8^P߼fub8P銗KDi'U6K×5 .]H<$ ^D'!" b1D8,?tT q lKxDȜOY2S3ҁ%mo(YT\3}sѦoY=-- /IDd6Gs =[F۴'c,QAIٰ9JXOz);B= @%AIt0v[Ƿ&FJE͙A~IQ%iShnMІt.޿>q=$ts,cJZڗOx2c6 .1zҪR "^Q[ TF )㢥M-GicQ\BL(hO7zNa>>'(Kgc{>/MoD8q̒vv73'9pM&jV3=ɹvYƛ{3iψI4Kp5 d2oOgd||K>R1Qzi#f>夑3KմԔ萴%|xyr>ķx>{E>Z4Ӥ͋#+hI{hNZt 9`b˝`yB,Ȍ=6Z" 8L O)&On?7\7ix@ D_P"~GijbɠM&HtpR:4Si גt&ngb9%islԃ)Hc`ebw|Ī Zg_0FRYeO:F)O>UD;;MY,2ڨi"R"*R2s@AK/u5,b#u>cY^*xkJ7C~pۊ ~;ɰ@ՙ.rT?m0:;}d8ۈ ݨW>.[Vhi̒;̥_9$W!p.zu~9x۾vC;kN?WƟ+fx3SuKQqxST Ζ2%?T74a{N8;lr`$pZds=3jwlL Eڲ t|*n8[#yN SrA GYb8ZIaʼn8 #fg3i`F#5N 3q_M]j 8E!@1vցP7!|+R@;HspSI]ڻCZUcg5pDcIϹ,oN-_XI,3\j ]ٟ5~' SuipA!C厐$&k7dmhz/#"݃,YqCL$ڲ`"MUbeT>Xuv~4Le͢ }UVM)[A`b}mcE]LCEg=2ȴcmZ?E*-8nhױ1xR2ϫCya` A y!?h!9yL%VLU2gr26A!4vbSG ]ꧧWp/ &ee *w$-`J\ ptǣC^p#_`{ К8EW>*(D{ٛ,[fnY𱹞M=6&$<,"lX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?.S[{'g=Ҥ):d8h\y6]t1T7IUV:;.1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[l:h¢8wݓ19\:f6:+ .3}=uvKc ٹeS<>ij(o'ciS<{1$E[nP b?8E'xv[K+E{,Qƙ1*dcs_Z'407|qBOgYU|U--sG8`u! qGYܷw;ȌCPc_|(RaIBKb+{P.T! =ĦiTob d<>SHr][KqWs7ѝBYǭ~RR"p9dFg|K- obY_vM 4>/]e/dy,8!xŋ5 R<^mYo 3c9(F?hXf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZ:}O KT?mWO/dͻݙD։nI@'MJP[7nw-x @X$i,:)]Oȡ*L%sPmV$2;ܕsU J-n$ HPR2 xGUeCS-qRT%U7k= UFTa[漹ӕ0wܞ$~C9TiXtϙ?X5H2'R<*CUej;f|YQuSA?TUi"iQdP2= PϫAzD u(|+Nq| vGy&5Liޔ4"1 s/"?6m´/l/L~xSǴ 7QFUsЂR,Pc4FoWNNS "kG#33T:(Gz[)#c$; E&T}GDe278cUb7|3OVd!sЍxkڎz\qfyseq;,Μf~6OPYюy G]4|=(Qow "{lm]E$YȼDY"{4|0 u`BCk*XfTI ˲H@ P|"U LF,zRJdXդ^m S2f&Cr5KuQ={trd4ervOweճFԗ%K<ߒ^W7Tg&$ɱԂJ-'Kiem!I V=,?j "n&X_ 9) RY1I6-_>1z;K?G0z+{8_P =O21W [FQB7 "MG0>qShL6 of5ՏCwSJܼ:;%!v70ʿZT߾>簖 \%b+)"Ev"JwI31ozVsUܬgTw?EuBJ߀}j b+0R" k%O_6w_s /$ +O拦w|~"/AI`["xi#,m$h1 ҫ?iPv7@_{wq"LհRE+OB#3LIX=ݪ/ g2r~eΗ7 p_Lǻd^3MED']8މbqw?,!N¬'\9=TM ?ѳPsÃ)6.! zg0wjzyVx(}!U?0LAa㭖耒>ٯQD <漅|M> 5L3VJ(S\1Pp+"FI-puI)q   q9nua`tyS)fXƣW!,v 89獈.QOv6ȳ I8!GPun/.iU2Bs\z] G3!Q’*Akkkwg |-uꛆPȞiu{OOmlsM­tI=%deۦ0K0aڒ u>x#*!r;-YU 9=뙄ὺ:!R8Y9L7Cg=h ƕ Eqy|&'0-x̚dcHjؑ4BW镸\шo?=DQE4m~Ccibʪf`fx;U"E6O]$Jqh X]UYTV v=#چeLL|hp|D#?N6[OEM_|ԑL0X/t5QX0lCRTjEmaȜ5w$V`8"kw?H x6zCZJݗ\&uQnw\ig e⺛̣etkX};|(`[oK=* uuH-$@G|Ə#b"Su)"K7h-=fPGp.[#kfFX]xt .I!(?ٗ9>ȮdFmDK+ص 4Ug.7Ji 0P5mAaB u1>f[[* ~yHta@~:#Ub.˃ YCt’M4G:Ktö΁ǩ}*; ڼ8)> lCHXisLϟB5. !(@4-#7kkIdm0^"Ɉ\s㵦[b&T"wjfc*SfeWcIca0uyMM+m$Z* 'j:s$0=O#IbChiªnfw+Q([jfv M"zK6alDxƈ+U96\ᱯf,VJ%CKQ T@tMp-K-J8:+/cI.o,;v,q]nEσeFL&j` ֆ1`ulcK^]PsƟ.*6:oLO<e `ly`eOd0E<a\N;tM"] U#,CLڬ܎aE,OG2)" ,]&8FsS80[rQ+ Ծ)6q,*JF0[gI*M\gc:m_+oRye: D$J%͍E:Lh;m]F߷ l6ǰWӶ^lKO^2a7hv\k5Gn1X[ܨftܒ<;z}qEFH9/m+w_m z u D.RmI"!GHDĪnJVEk0V$ fH^?J ÷%V'?$ߐFrFK۽ql mD;~  Cg{{Ɓ\ȧ\wK} wM2u[ 8O-M-۟Z-vk</"U{5l"/X&X|>O""*2ԝtݵdt#I.m6lm^r2U;.y1i 6]qv.٤u{es:b V[16<#ـsSO14WqLCʾqnF*3uKSEZoFQq1H ac7k<7M/pi{׆p<Ɲ/aq"Eu gn,H-Ii,S[޻[8"}[Z"`'_l U<[!2Fb[r#6a5 -3ϒsHs'"|mQ"\otfkQgYR˗F:)r|:4>?!;nZP{NUQ]Ey'tb9 V7E[Bu;szO4clU#%)|v.ʪ%/$ջG-u8+QY5gOfsq8E~&F5ڤr5ȗqKkDdQxu[ni?]ȿ>藭[cixiZkwhnJ c>M@nJc] $]_  .վ!XꆽK`:cA8 BǩVqD,X\wm8P" Ӊ5BaK~ہP+a=ٕ/qhp_&Kr;bXF0ݯ}K`p;cC8ى <혟't. QF+8Pqm(40EbV˽G @~kPgCw:x46]EGNJd'Nӑcw'F`qn%F j G48@`p{X4-⪑t,Aiэ r-r8|N[#nwDGu$C.ȣ]Na TQ/-u%  nG A7r_83qġq۩ ҕHm`Dp2BѰ{iSOٜrch5D g/ 8H<18,*1[y .m 2fZJz!` dV'G'䖒B^@ \ߡya:LQ ㄹfe)F-9IN%)%u*.U%iW>0uM2 1n̖DQ/fnԳ݄V<0v9spm.Q#vS( x?y^/p^/0qL;f/!t|d3dh2qb6ŭ:{B"=Ղ _"_'x~3+aJI&~2lJ2K?IC@%<1V৬dCV뛱)on~ ̮v}xL1P~P84xԨ$Y)xԀe>@})yzIf9j>eCI naH_{gE&4Km'̀]ʧIVs 8c~>9B>Kvq ](p߈n w@ON$+0%ehYcSXMRܼ0?=>Lzz2eWƕ y~-?>?k~S 5mtwPpaP7QߤL?u>lMzy1Ʌ(aX˜{A=G^`V/E.u@a<31{h4)湽ȴ͞mF>F0{9P|pڬj57\Ԃq i{2Ц'r[#wpQn`5Pg3Wnjq$Q֢3u}W^*hGmy8e {3=_ FʵǻkI'zSHgy6fLd#4ʈ9]D7t=agkGf̾ϨrB60}_Q7!!2\lrmox2k8ͦ`( j ~$ G0%@b$alIYyHG] nqB`Xv]{Qk^{t[꭭: Tڇ:v[r #gt}m3jGa^6G͢tʋe]hF^#yK{qpAl1er0gcEAyxv<9$GWsOtFU?QE|M5҇ *h4ßeyV::h}%n-6Tvqu^ՌQ=Ia QFʜOihdu~Ѡ)NId?x<# !tX4_ ||#4ʨaIej 5r|K5 PvlkZ<ea x ,_5R_. k"禔4!£چtES|-e@Z΀yrQ4W]a\;˂ɇ+n]B_|e"8'J0xgY"x;s苙 "w f4wi@fw62˛lc(fT) /ˏ7炂#UCa1} -4 Wi:=8J'ΟVBLg[tS&<> '`~&fa &>2`+3*Btd|e͂Z9Y !XV UCK,{=1=齄hjaV%giRK')H6 F\4Lͅ_B2־^◔g€>;\ˢnpufsƃ"A$hQ|t-R1%Cr^lڭ^Y_GUW'Q6M}ê ,\q:EOaqwU Y%8x7_"K `%#j?D:X3n*Nn@\SP)GF58;"'Q)΋ 5<т(bֆ:;okP-NBu@F1ϤB\a9Un92D H+Y~ W2__-laQ.|\p XN?cԀ0FŁc?Ur}25"#f9HO?94g6j>c n/6QZ%>O,JYGvq+RVs]2IVd# J|SC;?9rSuQYU}u㩟W$q1AiX) ݷeǀ:4 @7SnR4>2 XeZJ]VXK|NҺjuXuT. UƚCE-L@T"މmu@fyiz&Ty*tB-k.RGlb&וx|7B-4Bt:|2wH/U3GX Nǥ?Hr|zYT \8fp.,~.cSuFb@]s"׳ZP2AСbل谵3i'nqZäh"e=W* Mbm~+#ZI:*GiLfn@E)OR㼍Ըz3D'dԣq•AGVG0ĐaO\Iʺ.f$gb-yW"?&у4I4g5Ua&y :E"'-?~E L7#vJY9}140`4UE*\jԿO$mGse-U9xqd~wݱ?@'|J>MXauX끀뗛\= ~Ū|왣WЏ?~gt(|͟}>GE)>5n'M·y[_7q󗽔{PU7#:o1Z8[9w|7ѡ/.o%_GtD; lϵgGvR: s}t|8hac؂fm;,uϒΨ*=ɶW= 0#o+q!WOB3$5 fb*JDK?^#T;&~h廧h* CbM`+KUYflA,|I,HY>NIyUM y`ɤ42e- F[.V_7[ϸ<>XXհhPAe ¾kc$veqz~$Q8&!Ե2rK$ri̖(}h%eWk$ǩUeDc\]"Ąi/y1(?;ր453`9%&j6Ud ö'2ëXE'n-/KbJ] yS8$B3\NjR5,9Y pFTpa$,HH>Q'#Y4ɱZL! gk1.lm7$8*;+x&wy?(ϰ{uJ,UaMn9[LІÞ4G莔{lwldK`SvN[Ky T!AB{}޲zE~m4(q2%D(x[FZ׵);;8 E#c86LRsY'#C Zz4 n-!QB3 pƕ E&86Dx`>\jd-U,,XlΝW6Èa:s@88MrG(PZb b$]x(phG31f=Mâ0npxb\tJѫMﺵ` >F҈Sc@g(Vyezd@qBY b$18_~>oQ /Ʀr+a}]C7M,wb$]܆#qaq4}u>iu~ș3 dzM(ɵֳ>o^wl{UlXc 5hN,e-@HlKw1خ3r6$\&ͬ`uKDJ&V-M6 w1.R4 v9ZbeVLoЁ BueG$vd1$fjT1LwO:Xuf #B ?ݥ{Ic0* "OT B2?BvӲ Ѥb22P+x)נLe$0X^숲9AQ\GI`ĀOs;d#<*"$6!4Mv1obd |6P˝AFInV$qtwf6\Y 29R%-Hw!LJfSHp4jVٰYF0xDrgղtV@T[ּ~Eӹ4"k#yUcZD<eqk:K +F #w ~bp7 ?v_$ ANY C<ᑤ.%U/ĢzGԤSG( N%%烳 *HҴʙ{E0͂ u5#T#\ ,U, E!JԬZF&aZFk OLvu;k5>o-<Z'85қw; sZt 2[XװRҧp; ,@ g)``$]Dv%)x齖&xM]\h޾J-D#2G%g՚t<ը$jP 6V:ܽ/nI&!᎕AypDg#qHqz'э[6pY],K<`tj_^F{e{:'^o"%2[ ʞ쒷q'b݅ͪ'!]B|lrT\Xe!% 7)p@HC#Q~W>\ܺNxռij3lPwER4h?p '/,8Y^ RG1$YIj^MTOKƼ(ҴĈ1|lEL:`/J1!Ar4TTkB MLkV:Z'<ǂpx .آj"޷%Z{6ԷGqlFP5#n_Hp4jP-#r\i쭔& nɔj9 "ELYAҧ2:Hbjz x) UըXR26qO5dXD ]KPa $n?8Pڱ'f8tUqs8o^` Y0Z9RȞA$` #锡gFkw/j1 t~^9rs<8 ~I@ vVI&s[p--MpT(uXmHpݖWH`G볬-t'ϼu*,{%M#$ f6;qoaB{:3c#`\4Yp%5 /ꐒȵvI.$SZ׶iMx+0Zt̷OO$V~p9햆Vghr1.m 9m 9YL?HaY/`SNe[wzp]ڒޤS{TƖNfXL + 9$+F}~^ٱ)C"Av9}rO¨k{OsCa%`wʳ$)$SB.+H ]X&8k8hk.J>ݒ9O gw,&G _?iP*3" tbxzOp؀8uI N]7lI#F%%C~i4V8ʕ~.k'5h `cœ(Er4(.b5ZɑƐ\Wcll !2m>88ox췌ןIܙm\;|ڔa;-ql$!o%O(}s3)[./>FXTZd!xJ N^[+9*PB*3g+}Ixweyd2 [^Ijk3gQI֋RI/$8z1$ŏ)9Ղ*n:q#xkNH).x=m1^g/ v})FHUq)_M2&{{??hp4='ZG,Uz~墮Hڅw uz_7ILnVFukA`j6o;\-5N363z^v Vfw-[d GĻm=H76z V R(A_xf+ W;o IC,p9nmprt֎Mr}af <}Or w#6#I#t=SIviE9f)i J>tzv,P:afEJ8i𚝜ٳ2C.A+ u~diOS& ͍Y/yZdP D9Z4 g[>xaǭ [uˇt g-(p2YʠZF-_-ڒƀy!r>Q[ |w*`2wia $]RI&(5q_ǻ]{$8JY ?onr~^.;Gd9Q-IS.-u #ڻ|0u}jrg8P9rvTymPbο#_o= ~^Mg;H,v?.ć6n"ٝbHv/uKd8GXI:LcİL\.~rgE;1Ա8j!T/|Cy!hS7RN嬤^7Sɼ_T$H@kC_ҊY." uy>ܚD$zgu}E[fxY8YY>޷[rVG%xTv!U_צ'0ZyI`Bk"qDD'+P"P'ojji:@=wmX toއF#ItE5*)iOマUHrɖ6`R㏏# Ki҆&s uz`aҠ$/)Q6u`7GtO4U \> F ]g9hhi H2!th0Ax#kBH4H>s\eK4xp)_$`ÍH! HsClHNw0Q5Kzs\̇Psl6Tx1Dc4:#(tL oZ/Aތ ar٫?*;$xS(p l}FcX5Z2 : .OUʥGyDfϏ̄'/9(P'  i>7e_tMQtxXt v!Pz[ʎ`QƇUQ noA &#cIdl UCE N}h՗` @':u_=^s,k„\r4*.Y\aɌ ஄]`~As e:)@ŕWUꝍ7.qflJ406^zf.LKSh(K-n]GZ(͙VP[Ĵ1Y^;(:In:+HCP'/p9?crʬkFZ19m Sar1o4U6 q*$I_d2(t DɀTDq .0.=k^m#l̺]vzȑnlEBt-o7H ɊԒ6q$-PhZ3Gavi$vx Rm@0iRiA1ź]@i7P[ Gaxϔ"Ԫ9C h+$DvԌm\$]o b\'n7RfG@ik}H ^XQGk91*Sd{5z sv_(X1 X(FYp˹!8'GO4)l:wFrcx#d +~<;&M2lP*ʔ1s;@`< m_NBJo~g/DǏQrxWey|~գ3~4rN@wH)/),xXU|+i=Vzzj[Vj(lS1*[@]VvԃRT[)LE}V7-fyv>e $.xNc@NZ, _aoawu8ޗg"pﳗ2uٻ'QP}ț{f ~6}θx]4g(M"lȧQDց|ӏ4o *zH"}'Al8~ v#ԧh)|=<7rqK%kӞ݁`ݞX\[Z-R6;*(@8B) ~4h M/6 &#`vdi 2r;(>;,W?ɨgdz#J,Sc N UVs#2PөDRf'gHKmwJQi\SV6{Sdyٴxm8i!5䂸ڀ!k,Z45M [~Ee&jJD72-T{ᵖ^)SrM6p!]כVԈכVހ׵vǙ3mgx|O\q9dtڥ)ZbgOVFT12eg^zC4zUoZQ#oZ9Q;4l뜱dDf>Td8y8::aCq&<ӌxŔȐ)aV挦5⌦7 jN9c,G3 ` L=2)h"QH%RsYF2JƱM+jM+ ISeseafVgb\cօʠ+}R5vC(=QqO^F}ćoJ$@r(*~ruo}Iq1v'!^rz Г`|{#4V(kOxs?`خp g^9ûmtD3~>?Wz.tBĺ"ScB$Xƺ$Wn7ֽdwu#qTFPp8i[C (jӅcwvz4pN2 /BWWQ28fEhlGG^tdHa?3VvUSUy9}EDKpA4; Wˋd4z`rMAe9=#޾s3ՎM4h˾}5 ^N%m0ϒJ|1Y`\$xԼ[.^;'flNPޗCEdsPr MF@PcS?- _@D#՛5]\Ц1Fr!]{fuest`4>o :!Xi$lJ2YTu,Pv,қw?O3?Wӿ/}+̏|i;khk`J uk?J;3k cpPٯrH"(/;S|jF0`li}A9g,XYCwwtC(D+_ ", /zW{l@$ 5c7~8@NAiȇ+Yc̝Ώ9J1g5QUEg'y{HXY2PJt [nC|e~0=)Eb"rj$hлAJ]FS7B!N7h=R\__<ɸqxdw'`Ba؁i%oX[95 c lk$A)Sy* ?K>`P)q/2HjHFҌs2D'c^ f4B BH&s.{!2~L"Ԥzם;p vniZdbu&-;+vud"IqK83F)Q, k s߄ڦT˔rI*̸T6YZh!W)1!tMm {#6 )cZ?AjɌF,(cY4!P±L5āֈNR]J 8#R0CRufQf 7"A RH4 @U  >0`^o8F= ߈w4m;B!Ox'3͵gF^;iRCeTb"S ߐw.$RmNw P?)QğXFqj{tBpXz"2p07 7;+IdeX8`M'Nj lFAYD6EK"/8IE ȕ DrZNȶKeC6CT$ 0Qx汊A&CX܉ 6J/n®y&iV; lch +ȊԒgׂLp&tY'3IքbN[2\/W;&ݤ;ݵԘ{¤%Ds(Я>D(SKFpjnWP4pf4zZ@2 2W8%N)ùp65Y=\s:0@nmf=2 z,-(gR+*ō4M#.K& ::1 nߘ#u;p=ٶ=-v LS-)R=o8J=oƒ !͡Cܻ" ЦAKb\ںȢ+JN"]ꃲ%-S EܙcGPdZHR%6F8NmnUGM0ooqCWuޑJQ 0!ؔrH SA2X@eB4x0BØڒZ?dDwH(M#$Ɗ@bS0/)C.NčE(pM0ImH[dR";6jO85+ `j;Ҙh XTtI,CQ'Dv'$IR.qEt\\3-$¸X0-KĴ+UߠZP  03a|W S!G*ST=Um@rfSEr70,"\&h3;쁦&h Af> #J*_/VorXV"*'ilZC$&x.e=ڵVNw-Ne{֋{31<5ʞo?JkhÝt Gx0CEoUTuƮҾ0}Ruki%ňpx-D2w]]PKG$jllHM0y:a'$5`zjMp[.R*BTHF1cTFHѾputhq3XP)S8@ꐏrXj9IJ$md"!oHxH%-)W޾u,3*M ҩNƱVIJ W #Z ` %'0f;Pܯ,C[ePVG fLuDL."0\_$(ɡw @,fj/| CTj!j_ j:2`gYڏy9dcbmf#_m MQsiY;~ 5fFHlJ1wYĥqA]Dh}N}VoL8s NҘg0sgzb1+t/)qW@2DDտS{̐/..:H|Yw`\O^r~R/ɍc1gOc!a%j HSojq@_:ƶ`=|ji[RlV-X\nbcg+o+ k0i^Mpig`5(w6ro|NTlowVa#oIO=$e +&`rI*bt@VexU r>!3v^n/޿ ].ĩHF-<KB&gcWˊoy/.nᔑlx\StJK08.ٹ [&c[*:K' +Cfޗ18?#z?`%;+ WV-z+^sXGYD :ÊYL hx +|,>z+b{mݵQv6VhQ(ݜf87Oec}tw&%d˻U;.@j w'҅Qzbc=o9 &ZJ-k$ӧ\-!2G c 2t#XOC '{ ˁ5鐎ѱ.GVxf1 t[9 a=%2`#F`R%?4Չv.6ss i3XC~7ÔMFyB&+3mt]W|>Ywi'sB{agŹܶ|Q9ߐo+4<{s Ìgb(#֊hAP8,hn i+zA.1eC8p|ukr2,al3 [3ElTPޯy4U0l_WD?ѐ1 Y(IϷ>zѤcVS4с].&j n ^r6O,*WPY)gk`g [x9/Ǘ>ֽ5<=ya245x HkʎrdtF/}/41;6ܐu@m 6zp5(w^]Jpw-_e0HC* 9h_]jWr:_ݲg$!ZZ (,YoVMȬZ ˬo,)Lx#%E*0ȯ|MyWBy_w^_#i23@r톼o`U6󻟹yU̬L}pݮcMjtե~W43E_$Jk&0&ͬ*٣>idHmSHQo";Mđ䱕4VR5EYky[/B:p ;>HckXĖX:Մ} .& \VB\;^oX2/A}kv6o?o,infh?OumF8Rǡ@n>E֟O&yWfY6^g6 W$|?p3x?v`8I1},޼ynj.~%b6M~B>חn=, ,]w΋Qn?z /4}_nF6=,d1Y9`ffo| kVȇ~|?6;Ћi5Q:> ~Vfn(oPIAkX2/(0^m̅L~⮆_x?! JalvIF7k2*hc8bb ~g2,jo>X50)TgtÒi.OpJ$xrO23;|+xOU}}{ߦ_iy{X[!|%7l^,w-9{a x/4Kn8kdP>P-"mtR<-V,_w[Le+3a\|4Q2֮&Ps} 8?K^Cۧ:3p͇L')OͤDXf.Y4f3.>ϲ/0*=i/ne!wol +F["⑖77'Tnu6%hZ}s3BdL̴-FZ?}z2L&ngS3ʱL&rn׿ٻ6$}iߏ p{ȗS$CRW͗F$Gjj8@BXduLMuկkV_-0pOС@87tn T^X $84hrZ$[_шPׁ+T{w \s;\IW1@BQB<ڄ8"x d9Vl4d#,O6r@MT\:Ӵ?saZƇ|1KNBod}{|Q4I_T\,)?Ď`ŘTלtjNkNkNJӚXkNkNkN.횓9j*jN`ގ4z%,F"5Cbu|:`_GD_nv;0& 뙱1$ˇWLD&$S#FkuNTIY(+JW/γ79<ߞ9E3]LG׎ZpA0_)7{_7]!" lcö#Fkır;1aťEJYPqcPB?jAI\ikΈmEl_b{W p#U"^ӳZIz5_X%i+(So5M;Cn|UICU繥/(FGτV()5y8DL!šNX`"EخX 'muq_ j 'cIe|kZ.3> [xĎ#1-!gGbyawY҈$4Ÿ> %l!X+@1Huzwz{kfzrwO銸~COZ|fl~k`?)3N0c`pS0&eIۀ4hyg$sTDuDt','cT.(56H SqC| 8 A4@%3_I6w$jxpyl]l]l]Al]ՕRu5Pu5[Wu5[WZl ,"G=P[pHgmQ[pLڂ"ڂx-8j Pj{rNzNtZk$LsCۿMVJ^mɭu*8@^[xPֺd[9 (cXI%xUA nd*~^uvbbJi F0,"JvB-jyW՛Pg量hm[Yi(YeF"GDh.r5nZ2!2nreXtu: %cޭyLjXqG2>$$ڤs膘ȻsyDz$IkMW,FE$$o2N x45Ibta*z? 7dH*9kgx"Z[ٖlȶl%Mq%N btL q/rJjd A4V%b.#(fszw⓸(7$HX!jpnH0PN㚢-"*Z.{=Gxbo!ɕ*i!w7 MZsV8O"--F>y/*r@JX6HyGe+ BcyexA@;()G ECe]sO031=S#:?7V6|c mb"U_P? P p Aie!PʶddCee1u4-Np#Jc)cd,$.$NI9v\E]E`H9(Vyso%48x~Y3AdP3p)"*T9JKpp9y񰶂drGrs"B5rkEnmdw":NU~z"7q0Z܁ŋ\ϓAn)E~.FEi`ˈ}sd?\M՚:=-iOGwzK="ⶇx2{;orSHq$2[`hOnckJx4L^yG'CFFIpEjıP&Aeš }i)!(]Ep%AS.PtH 0N X j"BX8!@2T$2VEZe9!ZQ9"/yyWskʍ&vuPol@J +0pAE覹("*f)z~=Vfc;h[}y.fk2$wX:R=VBG)=R1e&qNȩµHdhCe9L#A;e!8‡ P8#cu;%r-#*W/7>뇛 JJʣZr6ˁNpBd'e!o_Tܣ܊a> /A<"{&!݌˵Z r?HqCjD&މF,9v+EL<\q#h$M 1:%wFl}nV ,'YeRikRΆ<"b.T6*Ee]GfGyu7w:( .4u)\}zVQ')/ !@\kbKjc}Q}QDUDŽ 7MKS^wě͙ h9}'g}CVtfh@¼31"| A5JͩywUW0UWDQV> {l[Z~UU2ǀN6 9B w@Fu]bW9CԀV.N&DKeJfXDMRxj$]ʋTZ?WR>8? AB>4ƈxn,G-sv x4wȻ\ihnVE+m bE!:8dNY,# m6E"F$@Mzk7slE^£!2"gЂA xz=j1T pG%1!B%s F0J0dPTyWL>=,O eI phց&wP.l8P_*icEXԶ/@; L{|⊨./kOg25 ,戚GD<ܿxXVeOjK;D<8gQ(cE`cEV4py|7ˀqźMЬ$,?>\1qz Ju +ѸS2W,DF#jc^^DU?wad3{`s1ߖ9=M.x|eDy W/jxmubM`>'Jλ1/#c7$`?jbgpn"b}|xoQ#o`.M{,{âJ޶Ӈ,ՌnOz^N&&Q?`_ytEΰe!D`Zo_ TM#?:!+66Xu%(GNp$d(33lH' KȻ wxy484:+ <3ɀPbxU3)-oB _QD8ګfJF7ջe'^QqWT}~nuZ0䳼[zk':[*rqn2nWr;&!K%=L{ZLWhjprg^KWR%L-&\ߟ-֛"& Ldz_/@[8|i8}|}߾]ӗeXֻUzmo\[jJ|.tlWcX^_3ݩ?,̢Wt?qAlM<,sw ;T?3u*~n:.}d}+v& ߓ Y Ͽ҇vWB .4usni|0v{a[>/~;`8-cT]%K+rۧz\y&t\*lv{`߿>SO$'{pꛯ?$t9XmqJ9L5?`=vX; &~WeXa_7ij%ղ7"؃9'GYpYG^ ls}@G}+zd&P5G3.&9lCƌ.j#o9 Cƌ1x7A*Όu(l0)(|8ʨK<%(a*7 ֥P\W&>[ IN3wтthy$1CUZ#3]g1TYislݡ^*p$rz@˨kBIF*b!r6&$ڼn Ͷ3ɒ4K/ 8artIo "o j< l(9μEyC_3>#6!;aL2x"Ys)W^i 8S!$H2u}/9ڢ%uI@k6 bEaL`QAj>Ӕgd3H !*;;֦<)wI1]6k VՑ7>p(fUsᓤ07d}H Z0ĝJrt#'Ӊ4Pd$v-&n0|kK,Ո",qFϞ/ :)CCGaLaJ3z<=,uZ^Zt\Dh"!d&VV uhAiO " cyz&cSb`)8DE!.[Q++׸P 1օØą7 gzInCogt!٘c;0 ,~ñ9(Gjf7wy+REϚr4R^P)ќD.^iF^m\#]ۛ WSWCeīvD=j+6TT:&(͊,Ϸ"}k),jOZw1ʑOrmђ:QaD_.`Xs!^~k@m]*V/ xPY[>y_nnfK8D E MdV(blʉw^LYsɿ,_;Zf ~x ]^jt8jEO0E4xS*Ti*TyЅm!(}݁0?wM#{{`Uq>blyq1Z[~[kt/?/z ]"tXчEH-er_l x-_~ 2nHT=:[o>mݳmcTd\ ž]MQŮ0_g1FuX!c BYuE\;\롏X WElcUɡPd%-7Nǜp^޹ :2r` u[*n1Ot#Tͬ`YKYˊ%Iq78PG_[c؄ɯooڸ=sWϷ+KRţ-E[f6~ x6ࢲ bLa9wΓu3v3@չћ٤Z􈠜RCvp's .q#裌:BhGP(*q']HH|AzW;bQg'l+blG>P<»~&q |EyߟK_Qg5{l^nߦ:!KfwD-cvSB8Z@\aIƨH0 koϫwuTk9U1$2I "%1y{l rmbg%X [u~0&)TQ"=~D4c$:|K&dˊ(;C1` w+Nad1J@J5NXD"PI3(N$%^$IjUAv˧u^-9ae`a;#ĻJ;D!uqCpn2n_m?$EG%uG?!F0&Y  U{6-*  TjStf@/K>)f^$MXǿ}/^iDYa 65=7mɀ&8Qt/|x0ӝ3E\Qx9bA]t_WY7h_6"3xxʇ#5p 1֢0&QP/^wq=PW)r&E!F0=a_GU!I`aXt;SǼ/CąJP1l,e+r!F4ɽªK ãņ#ŦDL?;ۯNjgk)I!E{+X>L&P!+2*4M!GhܣN9"r{'E;򭴋y]P gYj(K?~h @E+:`ξAF!^?cSxM_z8fq]vp݇@\e39 #E>TBG[V ݣc'̈́(BiC%oͼih$Phq$C>TiC\_0#]iw&IIHRiƙeb)HjJS!7GK8Nǃ~Kh>%v ے^rQ'J\55E>b<[ԍ,Zאv?-pwVZ4,D]I ᗤ91QB0.}8-{4<"T#B;r^x ̣m"]y®]6F}$_/I\!7otcn]pՌxfP5RGu&xEiSpFrV Sf zxEDS54^+JEDAvȨ[p!x[pȇP Uoxr 2`2vyWkҽ'H -2l]s]Vw>V`oV+x7>[Bkōω2^,\1J|?)L PiäqPxdؿY3ʻ\m" vEh~>uU4cn.XAvP"c1^6ns *B18̸NG'?C9ڝ0qc&#֮]|{?V|w[<6G^gX/SzњG+[؈Gu>bo،*=WgX9cHL:c\6#m6ȇWxj;T5tS nFt)voW_ 9IR:+ &!0PXsY ?}@bȇ*[RUuM˽1Azq V*qy++; ;dc.KQC֛3;F'0MBOQ[(Rz k=PCnԌ\ I!([a"O 3̓-VXGsAzۣtI2X-: 7щV(WpnIuuƐUW?/ץF*-^,lڋ%|֥c\:H甭Ʒ7Vc=!]>%]ʬ9>E%+S$zM:dQCUn}wʹZA 4Dys,e棌&uI C>PF`!rURIqUubv>8NĒy($ )&ERϿ%ڿ7RDIoe c8CoFt#Ftcy|S_Qܟl$} wg^"Vp(e@}M۔0w|1*0z<6ķ=e!|"*QHE IRl1{ٱ-e0hO;׉ʘE:]U˰ar5l*76S?ӂ,i2=q`gyA8ҊpݡtC0vŁ@z 4߱~@6!5.w'h{R.Kb e\R4 8J0 P7qAx*ɧ˂̦ffm3`{t`!qhrd})mrtnY.p)_?q1lvva\>'>>(ʩ Μ0.r6f_*俋lMOaQv/6^/u>M<#JKYyeq`‚eu?L'qq_4 %JWӬ Dޟzęe,Ge6Ef< H3 g&t #pRpa}gϮפ_'Pj?K3FLA@'"O2,PUc peA?WI3z \pk85J_z]+X`<,} ż E 6 1e4%*$Uq-HI9O,[xr$ EįgdBZHC}cQCI42){[UAx6?̮1A Y 7*)SͤRL`j&*c)/DQ %C$TV9ax5s.WLdA9%*+a4<OCLh,#5NqnVJ(eWP:h8ffN$֞&46L9cYq'J=r"đf+% oQL7<00M:h8fN0q2xFS=q:~PæJ}8 RibΨ? 11 ` >iwq`n{z n>h043'geg,b9jweǝci9^^F 4ZMTV;5h&M+%wP]V9eb5qC%x$xcJT Xk@Y[W433עGs>a2]-1B$DɄ5 Kj{ uΆH <݋ A1spS }h+IL75p\5 2a@5Kbǜp~~=Q TC<1*$K%Qy)[(z).}Sv/ŏCE8Ur{*vΜ2)Wsv Y%21Fܫ>¦K;bnG׺ RY% M֦#3E;9=X%VKf8$+H=@9a4L"a9iW|r>`LW>xsƸꁓKC1^=aMO|#~lAO],ŭj!aSͦf.>LB 7'H'043QLL,lST[֑'Q}3 r%Q,g"w#%[Ƀ'=.3͕M(Q;M(rA]!yY7u5l#wO'٣ ETkH^ iI"liO43gҦÅ}zϠ3@ lifÿ1+('86f"jueAbc04:nYp NIfj!Zmf <֒@@Nfq&i3yс'L;f-rFTtbj#pq|*L>43gde^;h^vF:F1g:yu>'=/n9 8-=E9 'Ůwģ^0p9;hPzeF)-g QOx1t!*ҙV4KiūQ?(4x) Fc۫?%d^ sQdtL饪H&e6bjU(2o]sXSzHclAuVlcbʛ+MV)f7642'*Mh`C]x8ǂ9hI'F,31(&M9)="=p&1C sY9bt%93бK!}8P /)ǕJ#{Kv욆cDDP%X-531ɌIdlґߠ6 d$!vXX O\K#rL9.3wqc`Ɍٞ(*iF#@{c碇!Eu9c?Bl A]# Ou^v[~`;făɾO.̓ORr+%ƹцp=*QHUOfu6 BRŁ<*+tΣЛAoWbyzcQY'M Bi؀HiFxD# 2et3h])=0طCw oҗP5O:rֻN4wOh{qzTܙ?1jX(B>kPemҮ`%0 =0mm0 2h&eD$cjhġpEө75]f te&<᪝_ Cq+Pp@;_A}"z\8E劈Qݮ~xtQzȯs+?`~NmQ:!_lq]If60D2m5y!gRÿir)wu?*|5/}27"!(8J7櫇iVߟ zDgMwMK:gKu8~w&1'Y1w/07ds.r,3xCggx˙BDEIC 10B3kp;llh a>1dNMp8K7]P[7Z*<>T!vɩ{i3;]$S@;O/˪g <SSb=ƭ˛<  c*A"+8"R!H)ybz9iQ]|ufY=Է)O=&w?{|QLTmz#3h]ۥy@uE ڳR<<bHu>Rγ EYrp%mՆ ~RIGye;*ǔMU>cN+u{D@w1Ozm8?*e 8Y1y]žEA?~,/ 2XiZOO~S10??K;N+EVA_vKEg/:/_ =_kUSis@g7vTQ8 $<[z4ӫCӧyI}?nͥ'ZkMNKyb:|;K6#,Ӣr훳|/< ^ޠQK6_(eӳRZڝvk$/˪u4ѷb{8W!Wn6^ /FiRˡ8[=cDq8PqgU`YNB81x O'O9Ӆ%Oݙ-"/Y|\1,lcxVC Д"LxX|7Ge,g ^.F6( n.x0r ;6Lr\ǰ߻64i@9qc"uBk7Pyʄ3V5 ukV=7xLk47 mwrYYpzom=wToPvyk158^/_OKݰ= ۻNϦ91NmQb_ݮxL^z8 󸇱t_DZjx*v4 Dw9|i'tt \ KZU~E.gPu 3'׳Wg`g^|[EiLwbvy(3u#aY$qE:ZR$+EcK^"3&+Qq4Hy~"/n$˷ LsU8]nf8x#}LEMQG b =[2Oثv3r D|c{1CWp@̆ 0 Hbpljr:?Z]_倨JvloSgLc7eds:e$<=2H}7Ȧ)HKEH I"{AI; YR"Y$LM3I~2S'2N`k%)*I_RVS 6w ~rwsX1~$ 5 9a3eTMr,Oє^^^p6xeR^ ]MFC-?Z[,T0CQZQ" Eea vץDٱ)+ho}WRn< ;@*UsqHf(1هZPg3]@Q?[. #xAv=P|<txN6!lmo(:y,ĵ*8Pɸ a໘d+< 0dw*<䙯UA:/~=}BLz4;sSଅayHHD10tMyjH#U\cz|4dYXcp@8QG{8F'Q)-\N4堨$8D m 9, 􀩑8H:8pG<@j(6Y7w~S)&YԽpS[P,w3QU~Ք^\Q,)\QӬ0-\h[sQJ>hҲP-|/v8ʶd(E; ;?d*931|IB9"Ahd7w6<@)~'GޫZeF}%76bwRy3{b츣N>X<[#fmRB!L%d</~gem __ghuQRxeIn@8f>ƒvNG6L /.Kj\E.) G(!VrO7;ŏC{1Tot2ozdЕPtaׁky+9*u6y{ă>&u-v_9mW9xL$pArLSD0 .Q4)hxXWKc/RPYFEQG $ d(3v!R+[i,=E)RPxH$~zsƥP/wWNoHji@i 7Mgv ӌ %\^3'2 #ƘNI+y𺤊^5L#~6D$\ ?uW y͘ 7h(<Su.-PNt`mrXiW~/B;s#r,@@I+emuHqa뽉P?4y+"R߽=Ͽdr.n!*&i7U,YAiDQ´.I.]0)jѭrR)CiZ 8zKCKir!9U$F;Bz αNJlflV@p 1j0Y 1P 韸;k,G9մIn@-$p,+ypo>aN³ܓX6ڌ._3D!{ce tDsnUdV?ߧ۪{ta'mFbTVj]v1vEZ=®W։墭>F@H83ccI/tLDÅG'꜂TIY: cl4mFJxeqvysOs$60A܏=I,v&I׈8$lGWf o/j>6pp abX4NFY"* >zN(j#飿jK:?ƭg+M.:vk8|,G`+A"(pJ"Fy$@ev[mnn&1<*%X\ij BMI6Y#FDT%{(?687ѧf".I37;yIdl06)bix 4LIܼw ~d~Kqm wM-,bvv3nZFOZVCs7\ߪm oU%T /Y=Lìͤ~z2vj63W?]~oWkj$0f(Zf`a)~ fny ͍,7Loihr"rhאǫIae:(.5a7 uz8 z*$0Z}rӘ4=&Ղ9A.7 oagRǗu󣬼=3ੳ/lV;,ؠ,[4R~~N>oU n^9'aaHЇ*Oq"0LO酒OmUn14 8TO>y`8:,@H2lVD9)}O46b4ٯ_ǟ}e룢.Qŧgz8_eش%h/N1/ W/sf/4{,3Wm~8׺]jG_sQ|+%.6S:Fn%#J)*o`8: 8Ux4bl0crRT)jۨn .Pjd3ֽTTe{xD.oi=o_7wh ?{6s#39iggQF4xc/a sUjW-,MD7?|<(_? ~C/RRaYE6 L (R:%Psl+h^ZdE뱥\ZEE"B2`8y=hw",AsחbqL[7Wqg-@ YI1_ֲw=^#QF!I"xD#1؀j ڳ)ҚK.QREPr9B=dq0r,K+u{q9C8*P&iZ".9CGrtf<~|15_nztc@s(T* Aρ,}F̑Ѐqz(`@\PIkeѵa{]h8Cezd638*HH"\${ ]%BEtO Y_ar#(Ǥ /!RTJgq՚wgv!a>^`!ApRd팥J\H@3F-FrML^(s( CAie!Pn3Cd{Fgr 8FĒUj>`<[ހ _!@p0rk4pEzKԠ$zfM>@x:iTtL#mt(pAF!`{r42]9J08'oeh:kMnˌ aZZn`_0Lq`d/֏CS9x_׾Z"lkJ'9v.94ArJ[hI] ʙHNY8 9TPdS(|Mr99lz{$pU Q9a syx2gˈXGLV F$^Jo C׆ 27ķ|szg0Q$lQRg!.z X`1TCP[>t'D}lX0$D6jj=YޱP4 6sVgH½6%#cE:41` !|:T{jHQuO{D$8 <\ F0J0^UuR嚳$6F1$>>د+ʧٻO F6x#kyG-~a(+`QN #0X~:c ҽ{o3s(>*?rI}3Gw`zA{cRǪ.8+mw>U*~ˣDabɕi/^.+^\=sC.nԫsMw#;0%iN:vr`~$CA(";Hy̱=W޸Q;jٟ<ﻈ # n"I}0@JMI6&=GI+vG Pv<a.fc_M@сܢHJ&c߯upq k1#A5E^1t{DYwSth's-eg(vѺ:)0B3Ĥ R{( =sBK%A7CʍT M[zChMUX7in1r0!+E69 Vp@263S Y6;ҋf`,xj):D_yFF:,HDHh`Ptg\n "ͺFj^҅eM"=u{p(Vx$°6jFݧs3-4zfNw[ށ ) MB<%s48Ċ  x;$쾞'3 ;$Rp:d@Z"cm3sTYߦb`DnM SX`Qy;`IobRPRiSF0P))fLmTJ<=3CGWr v_W1?\Lv.E.2b]d )qn%HO9\c٢JQ~o(JegrkLz'Ol25 KfDιTHA~2}6e7ܶyHH޾ ZY!Wۼck [=bhKt%~]:.Ӏb}A" ?2>!o޾XQ ނNY}/ jyLsW\?wX$dʉZ" Q/[!i.A,X,6~Uskȿ̼F ?j]?q!AUlDi0QBj0COO'j>9V=/X?+2{q5i=)ѪRCj)K-=:n[`(Vz9o4zɛEm*Lr I]j U r)jE/NF|1謍n˛FweiPI6E@HxBcp%*:5j|N}?ÃH PsaлAkhjw3Y:*~f̽eVɅOHBsHg8 G^C6V$rEf{E#Y G\9ɰ"杋$~TDɾnp]7-BpbZ`0Nr7 iX|Txj6Oh;jcre7ZM_h!7|1gl-H-g\ 5l $\bTX(;m];e0K/&\a==slf t%I*=wyTlI^[2Qw&0At<ݶ=ov9QEn6B'fĉo4N4.s苬O1:JP|޶Yqڙm•\!:2TS)&3td&6nGo{(WۍCnDRzo rWaq8s+0b,!!)xʹB[hlKAPl_]@聱&ŇJ邅Z)87mF;YQz $ZHU#y& uMm}ɯo>_/Y5KEfbį8* ܫ\#ߣ5s:<*'N'K R"e܋BL mvRJz`Q阤`lStx6OnY팧YJE9 6_Mf(fl K:StK X ].yEdVn0'jVWb[;6_?"X^R2VtIiœ!Y㔬6`\&ִd)%Omӛ۰Tkjv `7`/d0.|X2G;х/rg)ZAl##Y@$>r kA"+Ҷ3DʒbiꞪꯪ, +Л*0w )PBqgW ImLXAr&C,1_= ^::g)Ƃ0Dl%,31%!hppT0c|X51(Xxq!N|BP7M}C_msGXQI~`&`Uc[)溕b[)溵s"7JyE֘'/B *|!ks&YQ?kݓ{Wԁ2Es2# zw.%n c4Oy{c.YiYh dw BqŞ'6Cq}0oA^TO?6r^)3]_7x;пpUWhx^PyW=\iT8ӌW߮Ϫ|8] 'MOJ^3k|l$ioƼgOWf0 ˪` ,f~m :ؠ=MFFqƢ  =PFx1UE>u_Ss7 5:UNQ C.rEcS>a*RQRkZR5l39)w)O|%M!fRNOZW}Nb 'JKgk]f" JWtu~NRr|W\T*fInEYkvcM͚ϲҢ[5y6kt́$QG$r B(,!0I(Rs52:_ur>^C3HTxGDJNcRYL4,!ϝV %w-' `F^+cPL6ktVɃVrºtF{0[@IcT*8s2˳?Mi^l)}XZ7xެ&T,M S"94-|*qr\xj,Qj`^ff2v&QgLim F/@ BVCڙG|ޮRO+7iNJ׾U{HrtoR280)͒6=P>L${W{r1r`Й^=T#ToβyG^6qGGh(M_n=8rr}2(bgLT"kc4bG^ J1umo{|\z57(T.:mA",@ ]Wa3KȮ;LYݍc24Ĩ,8zdx-V."lf /h;刡++`ɲEws+B~_t:+mW-1pH,&fk 0u#qT\*0'%%y&y'|eWݓWβgf+[ȹbrms [<  zd~ܮjÂ.rRG׫E+z Oa.,Z{WbX]0!۴CJrm`Bh^V0(6e*;_xʽ?U+#=gM,Л1hمeUb$xAL|bY3_6x=T*03P]OuݐnG,)ae(sRAr>n;28xmmu>Ⱥ^kǪ^u%$u0Oq N*n QnbQ4j^66G.{/_z9&Ǔӗ`|}|qu$7 o#`9'wZCu ͻ5?gKCHoL7_Y ,{=|<ϻKE/OQ.L-̥DWl~fM\Љ*1`&]m/@ENzKcLȢi1\tTHqJ;a%o\m_g=W´!̶iyI DU]]>,oіwfm'䖡ADHjH$6r,C4Lt;!(hg@ p:UЀa2ɾ+v:Y)WV na'@`ث)*xg\=yza66B'G٫Tr`z3ЁBtT?dŐU EejeQlD+ ~8<(O ݾߍ{Gg|Ey. #,F:d sgzDtj#3`oHN! ~0lqQ[C*w/1osч>pKP.t(ύTu(k[o3K`\E^*z5*Cz Y=@Oc.*5O&T *[ nツ10PFq߮awHDm҆RJ{ti+BJ.,[};$-/u6> 5&Ѿig;lχ%?d ܘ?y+V׬aYyXd#7W,ʯtcÂw?7˷ӏd %P dCW[-.p`53t`ʾ}떵j|SVDԸ(4]:S}w9pjjKɩr\ަ|{v[pؾ+pgLLVÁOL>mÄu oc.L]?,_|{!vql=#Ck>3g8L0kMƣ!@Hi1|KKZN)Ifx8&0X&۫z7I2kRҘ˻K*Ց{a sﻃ▱׭xLc` sH, '1beP J\eG_0Cz 8$*8R`*L @3@cĜE NVFq_nկI+܍3.9g;֐V?7e @zQES?AxZd< bj'#QAEH)CEPFyf (:UgrZRMm)6e2YQgj̑2'6FBkNRDX11mIISFx"rc8@S37we`nj4VMS3ij45s9M͜Q(9ՙRGLq-wN3cYT9}wZ!샚ֹKFT~z8j>JA3QLFFcTX$ T!`mprIST?9\86Iw$.ܡ Q-:$q/q:T`< y ( KM$gvB$9{.ǒO}dO=<18Y>8#G g)O$Z'Oa.7G;aqX+0{ 0E=BRPk<}ۻlPvJP.贅/Piu Z\ u糌Ga0Ic,Ϊeh|S|d6K0-gu&JQ`t&vy:i`Bhp5qk k,!a|R0yhtV>܌g{SyY8N.fNUb$xAL|bY3(FN ɅTS g =!jYfSCBG YL+pлOtN N(^+A{ڱjikIx]u̖wmi$temGNzî;d@'?Y݀",KnU_fe~YS*(ɫS=4o }uk by!~V=!h +^,u~G)=mQAF*8ОsOApgUp9$]2Kh$/Qޑ\V@lK( "HkXKBx &hN E[T LHjjGW[s뾎dcʄ{! lsu4׃Ām~H=OqT4AH$W&\ɭ!9{&6xan SqL|m 1KĮ%?wGF7{ضHkOHKAZ6ul|-_ײl|-_wPO5;JbvzG0dchɣҥQ tmZcca;Mk ݊o_}y|̒5$N,I-Oe!DK4k<SћTO Sj h-(㸡>P3jM89Y .цoRIeZɗ32&0:hT (95yTD( ~"vq"ۨR_m5p4Eih$^fK瀖? 1k0ŀD+%U>QNmQ*J&Np$˅x=р瓉BOL2 r 8outl E~ǭku}u 4n3uhuݓ׮mD"rJc.ږNr (ko|΁2D I$bh dRD*鍒lF6 ``G4V`+.5i ,uWǡ!sB }mtk 9ڻx~l_<+_pWƗȯ&sskCByO xIȍ˦Ӆ- vM NCDG fKI SS4"!"m~h9NRtK냠D ]T2Qc Hmuz6GKO }\D ەkf( P\de6GC7,YNg 6֝q'URvޗaB8v^uN7rX7_AHZ $ rIʔѸhbRG Ch|:SRH 2H=(ùsM6G'ݑ:DA;5o!~ )J~|=&\쒆@8-g!KU"q%m;%e۠kC=M'Ak׉Pk(!:eD%\&bԸ$wB1[RHImm;n.Ao Bp, DIA\CNJ"FNx$h0@2[S/%YrA{㐫%ZdVֵ?HG% ꨜ~:ϷP(ns6 ~B5=E%;$'Op/\pIr-ID(|JiABȧcĴNH}r{uJ'K߆ [WwGV\j+o{/_..op(A0v:4o'O_40)Y> +<+&P>7F4~wQ=]^_8gQ ug.쟝^v>3 zfgE(hwWʱvn{ө+NȲ4_[X [ڌdi30^ڣp,擼~LEgtK[eV'\ұIa0{DCroϏƜg=?֞%( ߹5:? ꌚe]b.cl3sMٻVIӠ:M&G=uUMyIx+'k@wW7a]Wq/{۟|t}fΐ^(#mT:A3,UvWM{/3b3/\-_^lcq W6f`TPWzh '70T`qC@{C #WXy{kJOۨCeF \SY]5I-(Dk KDk=wd??D(AKst%T ,K%!l4Є^J€D-Bu&$^B5r 5-wu_Hc2q1BvLr e½ĆE:A׃ x[3ԭdžr)?b&9ɕIWrkdilHeγ 1窍?4㍶o~\-A~{2nkoŪ-Ǽ3k-fuGcZ6$rtPRj"Wqhgu:%*x!T 0F+HZ`y;JV7>7u*H+ ލmڭUCz{zY^ 4hgy5^y._=C#wWZʠaNQz7O $`,-yT2WQ tmecj;͵]݊p_}y95$saŒmyl7%*بd۬ib"lp 5A)`BqL l"n<,YNg 5֝7{'Uoޗ_(ծyUҫts?}!7d|MY exxKF05N, `_&)YmwzMi#>],oYnu;0o־ٓ}@J(Kϰ]8(y'j{#K5W:p~wMk =^q#y nb2asH,8)pp5t)ǔTk[mF_vIClZ- ~&Bgms a;EJ"q,ɇqg5>S@&O! ËsHJY"4Ϥrwsx;dV"ߘ/{KɬX;T%ٗW׳9xqx€ B" C\YQK`g/VebR 1zӓm_.__]:9TWyeςk E,h.H_ ˮ=.FM59&wnS5zc5jONs2\'뎞_9}7=uV-u rSzy;FZ*FVŜg:i+@MNLob}{ ,, ΖO-~o^o>w?{7?)^O~VO|`q&slW?tu̘'UzBPm!swNӥe](|=Η9۠ݟte3RM,O+lVwL$X7Z eһeNM a4`Ӻ"Gm~iyMŕ79[#S։mtp&/G6\M̟]@ ds0߸]Աj r;7gԌ&jYvI]\$`p=z~cB]WUg61E#npl.JeŧjT]7lBVr/LVݢx牷"iݧXOOOOOOTP*Q*:'eM{y˷śL8d&^ɿ78b%2U$=S,1=-O>bH BJd91g t M쌗"PgIx[>ƵΕy;ΌcT]aznFX={ %9xQ%VnxE 5EXJBNe*{q~G=x=/~P[s{j%3iWgZ׺Zg.dYR9 19 &Un&Y}F$ %@!@GUJVITE9[ԙ’eڭ9[ӓiz~Ne Zcf,jfhGZkbe1)`pIUN .6 qMACB.EUdӇ 篲:x๳Cb1%X(Ue i;@/i H.&jH-+4ܦ[2,vTC^CSz?:JK#y672$}HKjgѠh^vH2}Gk.c^ />r ?%7I 9ES 6p(bqADk8,Jw|kZ|HE[na&V!hf LJTUB^ Vld=*=ssխ,CAI.g~֭ 6O<8%4{vL숼sɎ5:E'J`{u\J_L^]ûzy@yG/$LC䗟_urvr|4wP(3Q \nQW~0nWW^O0d2Of-Heq`cYb[ݥq2+\l ga#{#y@x`ԕqZHG'mMz_'_jJOvŵ* eܵ~\|/Kehҋ'֟N]98͗3 `Z jX\u\Hi|x0' M==;m.YV/vW) eo*W5'm m7NG>$z=cbef?ʏO#y̡mJWh,+flk.N۝]I]5}9hc3{CH[ލۗWVU7::?Sm}(U^RaPż/G3 /5!1i3RGG!vCKkV \5EM$gKYw>lj,9K’}D^?\B 3x6meYg:++~L =+ u^J% ,Ģd$b%t]8 |JU B4+o;EYzgC;2ʖB`Ͷ]D@@rR&)xZtkkI"$F[Tt4~D =%kVLde`9 0hJǔbEU4xb~JfFJ@:?3xݿȦ:j/BJ`TQQw{Aen%ftIvG5;>":|2߬4,Y]2~D =7=DТ "D;eezwLgK V@J%ecJRok=x)q0Y I3 :7RbFjReF" Y=jS7.;ԲŲeҘڸ63sL5% jBem$l:f". x}Ӱ)Jc('c:3!QBZv9'YJY%Q-VcJ޹<<,iB:b, x-L4`ǔZ^p#JZ,]'?.6uкK2jƔ<ܛa-LP섴"bRL~L ]lmTFʓtU[)x <yH&7cɰdk?.'+vH"S>LQb[,R΍8Y4'3K 5wO B @FArB ˓$rfT'>&ښr9`=4>H8 .p^*iGbEd/CC0 .-n9PxeuUl4(!FlqҶ8vitud"cI>_IRT*%bA-V]oXm{>2@\Z.h0Jb KXK.- ^@ulOy%dҲ IY4fxdK U6&k\hwequ$ji"]ZA$FM<Ĥ ORf& Nl!C 5Wh.ƸdmsTjUץj#ix>͛?{?|wGVN9` >H 2s{4۫J=ΟzzIj, XF5RqBxFTpu{DjfOPUjÓ@u`@=?W[}퟊t@] m`ⵌQ3lR%;&Vnk#+ `+Ʋ⣟ |9CNl76TA spQgZJ,9e.E'6`Q V5&7)I%a"-81!్h%5 @/P9flϸI{IIZ2mN>)4V1҅bsZ͗;Q ~CՂ=8ID W@Y :R =f:Crz" L.i3gW7ʚ{^1o޾y"ƓZ5]N{!V <UV_o(;U'I 1*`\T`"A3!L^rIn(K,qG3Xi" yfE*h1I 9AG&ʙg)$ԦІ 6㘠Թdmj`މrVm8嬕?z|x]ОE4^96dOd̂XA P9FĥB"PS>>_l'"Qb#fJ6G 6 4`yF*{U jqo'v҅O??\!=[W|gh2:}uӕc4p卍d9k & r#ݩ;2GJAqT8flmcc@G+ l],=kſF@Ѵ q[{p5 +BW"):Yr4՘q>Zh <2'>VDvviF =">^YB%-0xElŐ0hAU=gXjS^U 0HG-t@e/lVk0o^iW+~l:Zpe-0{%0;nWuF7HדCBlD&V1;,'T>ٿg%yIÛchM3*ɍM^(IP!sW'iz<"i3߰,WxH^SE"Ȍ>%'TΒlC#6@:_uYq.},P (Q FFr%Z'UtdqC= BB 0Ro[؉U["=̓l A0t>9,JlytIIIv>jz,E|vҨ0Rf@f>b)&} gozA+H9W>AΧnp4΃p5ʙCɍm&p6a xG\ *g"tp[Iv1dK/pq |Ta6% _S_󟽛.'vx?\ʬ`9YNf%arK@9b S6<4Z.f(χSyųfF.J24Gypsb&d/RѾ/|'m2Wا~0ɿguԷ[bhf^h݋O5RpFD쟎g%p:´vj^p%$zF|JԚw+-_/njg{\ߜ~X|j_>xta>]A1ip?q|-ӼڞZѦn`c7Zݬ[Ô'E3Xfh|9ѳ>GG[Vm.FZY #K )+b?ٌv/}>y]4:QMY;&<ueT!R1H&b32ȄQke=/]B_BU-Z]):f㼄S]f:>;Mȸ(k/RU7Kldd *8X BL=zVo2g_{IYKQ7u{勈#dޱRΖ:I>==څhWK^rr>io0.Hc 3Z4Y+Hdil2]:'LǣxQA Ȋ * c`"(!IXdQ0˳r 6Lb\oB^RvHRguERg;A{gځ؀e57d;Ab+WƆ-X< TH` z2iK ^h9{U2FB($6RVa>Ff'K15LJ^vHN}GOέA[>t':RY_Q6N.6Jƫ,%6[ېSM7AK%Fpo|R/=V`٣ &d&PQ0d>LV0mttF+AQF!pRSOَnTq$] tKy_zӏ\w5R^<8vTnnŃ*DG#:Cf2D"$8::@/׿|V| oyb.e>F+ÜLyS= lre˂xϓѯg9^{Z 8:)+:,BNC3=98*Eɿwrx :Cjtq2Q{ڡA_ۗ]o-z-3Om~fC #{1Fʌ׋uyڭ]-N<[qD86za.ֳmyGۇ,eiG4͵Z/_N*N|YotpQ CI<=B{B>'Q>/r֦Ĥ[~ز\%~+|ж{ڷS_ib `ZVOؐ +h9jRO~?jp't6'Lt==9MYr0X|tVuEZ k-Y]]IN]b[{ q.EZ{I]-89>>moB߬Hc`jztaޙgJȑ_a%>=뎘u|Q8pJlS$͢l˿~UE$2Ye8 x $^ێʼnk>jޡ?e9骎Nƚ >$霷vgEgR=;@'߻:W}t-KsMnNvMj)#|6)`[ѯLގ>ێÆ!ی|_ݪpW >0ԛi(-}g?ܠIST\M^+MkƥJ˽;gƌc-Ysܰg3ɣwpr";.溾M]麨4:BvK^Ż;ʽmګe;GbXiOdij:R I_w1Yhc4 ;ukZc4t`$a߁:wbK|_|ͶP]QvH*6GTBH9!B`X1D0aBm#!Zg`F^., S8wi>-)w 4 /Nz*bbNs>@FOPK~MFӺð.! _H*L8LTooF(U,N4CoG_择3DuQ9`m> Pp4.ʑJ0/ӛ[Bzx ?5qh*e2I %,VrBԿ+_W\!*gvFjk(Wfj_7דaYG@_~jPlIygnO{䏏zLٻZ. 0 ,.C&X"_o8؇؆I*V{o3jSgfyf]fI М:mGa+gS*WK9N# _'pf[;IDKymO֤8Xg+#I@Sޛo_M+o)Y(Xn8::G<)āwUb.5 QE:1<4끙>:q0۲.sM vX+Q#Ň(#ߕ"S(92'VʜX+sbẻ9dN̉yz2'V̉92'VڕA1 FʜX+sbeN̉92'VFG :١~;~;s~;ΐjH%+2~;c~;s~;_ Ә(z>{l!AVJy\DdRS{DyD,#.DJbX 9\葎:zGJ!6xJ'qp^tPPҠmPV# $0Yf_gJWi }JQc1_$_qڊ#X ţ#mU/(mDm G ʃ2 )̵R;6:brN:$d,h.˟v I uE" R*BE"G*1F`J8'W>2(* %B# 0fDF4QAJ(M`;gHETB@Ѵc;L-/*E]STɼ &:L4\R/ sG(ਅ,yu0njGch&Х |z),[ӏ4*%POjИ`,#T{#(!JVQVgPMO~>OWq""F*>Vqˉ$*ick {''XSuKNQxecuֲщ[7h}/ե;N_̷!f6Mီ0 gk7a,GM *0@vD2glmݝlד|*`˧b|ȮƇ;6mCf)e)Pϔш*䬷1 NA90؇fԡyXF'c]iȮA:=^icufpl[Q<*"EUh,"Cs HpY@B6y`[14ѿΉ LӯKQHdƛUO.Sc"E)S(HaF.ϓ'me!ڦڥY՜>_~tbpj5$\a(+l(=$O ZƼEmNorĠ^3BBTLVGQwnM0Qdnf]._k:'_*A[urU+V@;,$52}4RH^z^1:G7_p[R\TNVW\| _qӷ~L={?cޟٻ7 F`Z_%{g 5??oߴ$]5 ͚m4̀oӮ+ڽo>-[9}7Y+7C_$0v4].EJ}34wMT#ºN=YI|(@UUo uGڇX Aq%0Ubaf4-}5bϑYY<y4H)ʝ/+'LyOo߷SJ0TOJS$9F&*D6:C%K┲<;u}:2[3ySx1;MM0w \(@3NQKN.V<W:Bwzvgعs<#l<unG,C1Xqc$8 +D"h$Ӕ!"T(N`2%ӰB91/C',ɇ+سEc=1j<=SRwL1( >)lt6NAo/$hӛoRȆļ^ y(pR>R*`>}8'KU*C^29 H?.qVTܲgIb GoAGO=Udo\f8"ՏMP\mM\*5Wm )j)IaeיFB͵BLϹtUV#> uϨ0>: ƽ-xٖ# sRpe4]zԼ0޿zz;2߱QPot,RS-N_i@0~v3t-h۷zׯ`*׿'p\S?Oa~]&s$pk{m4 끙m N;͸'&*g$J(zX&\\m$&Vb7W9 ϕjyXyUt2MF.ҞUFQ}#!L\(Gnc^)2t&Dәh:MgL4<3tDәh:&Dәh:MgL43t&Dәh:MgL43tv2t&Dәh:Mgg}soHJ3KH&Dәh:MgL4+Mg3t&Dәh:MgL433o+.agh`2}#tu,'`P}kO }<B1y?a?4Gje|s{G8foƫW)#V}DrWœa§^Znbl)tP6<{MA>FuȜ8oWT&2>Ɯ!'}Ir҇!'}>!`l<&ف5 4*сR9kEZV9kEZVK@2%aaM%S*2 "4̑ ((SJ/7nm?J>?!D!e!x0k5f,`ZFL&Z iDtrSQ[ j SD r6+n}E F?+oi[7UTu&lnvU0j+lk!׊Vs}s16 :AG u1ՇPUS.zay t}~4 d%F[vޫyQxsM5Jׇt ٧Z}FGø܆7T|+ɁoFz '\5gݟݖfE<ǰ2{ylrYǜ.Jnᔋm;jVܢ;靶[uN6a=O8=A_~d4z|zO<"Q@{~ٴ!Oc"\_՜/@LίKx|aulސ.ueLj/5 P(ǡ][sr+zcfKh5ǜTrR4,)GSu{z(h80K@V9Fch^ݠ3NÖҝVTvD%,70e } hbnQ[;UTᴱ&X퇟Po9{Q\!=>7.M?F/W)/"Bﵦ_CO@WatrShRxgCc|riQObb1 v%Q}<'"!sQ ;rAz䒰yn5][Oȃ᧥ g%<؊*:R&ZJL6E;)b6%k3ڪ~D ) $O-**Br c+ Ck#igզ3N>F&O ($urx1|NA"&^h9Q#JQ^ L A5]ۥB=b>X%KR^*SM(&#%Ptq %g *mԪI-Ob:Y>(@0q Nti;^{RD:j Gją CM<&L + gQ,LHB _JU,2D(=dy[[fE/N{OVv>"ȹ[Ԝ\$΀QtDbJE (2SI-Ut5[q< e!U;ϸ k +n ]Ԫ ]:5)_XVrzaΥ(}U90h}d kщ#du9B8m/nt [^,1xC-#*뇧_57 )o5.z7^FW :,/O={w7I^'y&酄]=3B6zQy7k$0gw?qc/w*{!7v N2#GoN><`uUH6&#~3!%3}pZ`t> vt0tт4l?)))iU$UAIGJC2e*ZQh lMؠ"WV"CЈR&$dEGrV`9A*h7(BXZtW+x:kTҙy̤$vY4r$K9I ~S(?Uy ,b65@⸝~랫L_  ggx]Q9.!*O^A0 $}$6ODd%yM5)c|s ᢡZeR]@5]%?=rAt՗I5ͯ}}x|?Ӳf!/9vt|v-;{!>5B i.AmY?)j/ԫA#A* /sSq&wosLe  ;Г E=C,baT |Uկt8+gA 0jmf=K6dj@5d?B)')S$3}he:¨׼v^y kKEx M|Bn!չb"qY#tUD/HYBWB(^GK ћtZ|_- [1Oc'1vc% UDonDHI ԃѳi @(ݝ*l_ B[+Am;HہhF֨U[kTmQ4湍YjkTmQ5F֨U[˓U[jkTmQ5okTmQ5F֨U[+6QlbTrÄĝE9s𾥟gR~͊Ƃ/\{~وDn»}RzwJqR\HfU2߿_[c\`^UM{23:/1E NGG{lB"P~ &J$HP.EfQgPWO;崻]~ |7QQdƏSD$~蝚HJr\,f;߈{OiX+,4OF9.8=qB_ٹ<2^N*j/EBLxMtJM7N<@Kbp Z)7 qqӋCz]Q2u^V~Ӹv t׫ r+<}ȵ-VpCUe1Z]-`nPOĶ[w8]eoroS6jm Z䭰dSVTGmgAϮKG+P~D\0y轨,v>CN)ay!r=*l~,Z8kuKb9|Vakś7k Qa@ҵj1X=2I-=MɬDh$긗o9z-]Sw={UjPG ic·\:%%'nA&rc1FQ X@vFjO O=hIK=4;!}SE~kOT]o}>c!Զhj:R:c>ߗ.(8EE АRCяCC2xPf!2-䈳^-NXgwhNؠ"Wjُ~8:|z sǘ˃F/Gݳ&-E>!͏_ʑNHdž,Xp91 AwuV$ +TZ;]BT `H^OD^Evk-;k@Bٙ] R R9㤝-Uʜ.HXv\ZYMZm>Geg%.d"JZHɋX QHEžzŦS!G)5JzDk;o9p}upvӀZ^<1g[E0];qx9JFdu7 H4 8n2̬$AKfd w  ^}k+5Sz)ð#rLH]G%1'CϤ<&+}i;)z`fߌn3Q`SKEB*e9AΊk8Bzc(U og_'ߕގ.Sb$Sn꺚YgvPVx;*mצLm7oBM=){i4v#8jfU>P= f0b:{Ea<>;7ђ {%hyA68V-*Zec Is`.5};+f"~Oq֫|Ov(KUuq:k`ǿÇ{~~x{~wKL..?|N $8M$1 lF&s\~ג57*N׼QN!7T|${ovd-ҏُ]gX\nɍZz@Wڮ WlR[NW7QR9Pn|(@/Mv]cicol L3ꔍpy7PNec, }QC{Njm~`?z[%*u[cD=gt%æ>:C%K┲:2ޟRԿF{@] D${aDb#2DϤBa'" @uBj 5 xn=0OFf:Ļ\A,$HM\Q_G'~xj?4Òm ̢-4|öq*rFBB An)1H-=^]?71X{( E"h$Ӕ!"T.b'38FQ0 D>MEcWDpCLDXYh6&8轮[:/+# x1`]3`wWgds˦-l7O_x[7WJ^18 QZ9JfW)) Pxb42i=v }VFo_ILn=u_Ya:nѝA3% Qٗ~;c>㺮|4Kv웟xoϳpmRl|1<+W"IyeK8'߿;9?x"WƜ9erɍu\0'Ԅ-ni'@ˁ'6P#$ "\g NiXjA$V"ɼ#R V268ġN`s! yۨ&T ;lb:Z5bHa)[רoSy9?ۉt['pOY=; (:)_*om¬5Nilm"6OMC˛N~RS w1̣voOo˛,+"]@.ͲAx37[ov>iՍ5Zzs9? z%퍟.h8uxWoOdn|Eu9vr΍zɝ k?~uݻ n:w;«"G-|---kOl I$DuU2?)}3EJYF3f?jJ#b{En{3ܾk~Xaz_R"V16uLWb >o0C[/%HURC3YU9 nU磚'WWi_bQ Wa0˦X90S#O҃T| >S陳 6?}ߛD_O9w]pԜUZ)aƐ$eP5(qCvY`iI(ʰ[)rmq1v՚<ξ29oϞ&Gk$^H1 N9ť93YZt)]3wF)!z浟Vœ_nw G=R420G JҾ@tY$h) Ӂi]I0)IXBg `*u4H&$!G=rl4xX2vHƴ (QHLd$&A:łKsg H!c Fˣ%(i۔-EIaf.8/=UFOS( Tt [`QbIIN7r(Vx$| qc{0N~]ۼG+nq_O⪅{ <y,B:<ibe8$6ۺlx#cZJ o`)h*̥ O=cF@mJ@cĜE NNYqwS5BE6&?י`cחK;>Ey$$u i~&.ZߊOEh(h_T㞧~v;Aqg:b s4yL-R$*VYkGE`?%UIm Njݻ**  PIreݟ|܀P ӲKիrQi`M+u;(,6)M ˊJ*9(|_/ zU;U#rs3ZN5\JhffAgȄOTRf+G329ca?r np^5>fpv8ٿG}Emgqлz_żO&8M\zܙTn{3~ J;sv%W وnw43 6!9տSOՆE`v(Tג)E>i ?çVR(|y;nxrl=}^kS^M3pjQzTT ic% .]WԵUUh-ؓp z](Y?Kj+yT66淢m(e?JD>),1IJsutEB>c|~*{]nfSSqz,v$ny9>ՠdǫXZ` +2A S[*+r/ޘz2A[q4!) la>Swr[8Vzf:QfǴeC^'iᵉUTi-QC@;qq-HϤC^WdUm `[ۃ^]ޠie/Wc)iZfz%8ߡ":r[Tx~"ץ6;Egq3Aء)mW`t3֦^#Mwl$~C^cet*TbKXlA ۡV %,( jcs\x ^`mhwMRǪ^/+YiU*Vè L3S`Q/ E92z7Y̔tCg25Es6ƳU9+T;N?Z$L}tYZ_ExHIRO.&#A l%\J/5C&$-nMV|% k!S2x-ŵG(N|;D3RZS r'tw]>,n>lWN薳Ǚ.7hH;{`NJYI\fNڵ'ZgǠ,LpeKA2xdΆ;ݛ9~jN T$?d}kuc9|mͅ;ygn+Y"٥U!>כC'v7f߻V7Xwۋՙt. .ӏ}X%QP.pө͹?\ />>oL %~˻R3UiJk}qSPy(p|Y=j3˳I+Yng]/v](/m]<6 ~.zmT K$3z\$e.8- &0RYJL(۲}Iā'G|N*ԅ`ic!F&Kk̜lz(CAs x}g}@T}1%9EV}Ji|L4Y'S3{|lřal15 { lzjC@lr쒧1TesF2v3htCjGۻ>Lf;GcwqNPEr!5t^ efE4BqSdWt  n¸~+; 20ABxY=`o̮:'ct'xBtc7GV 9Hp e^JAd`5<$נoZyGME[u^쬩w^9p?}XqF;%'~T]Zs=OkQ<9^tr2,8OR$cVA➑T`'rߦ`>(ndB 'OLj)AƇDLR WŰwvgD] $C Ǖ\13˒/;mЃµT oP> R)M!y2L$ee|f81Вw t̄ X Y(jE))),[Wbi} G"#S^YRN̓A@Kc<$niJ<1D4` BAW!0S F[bFGcTkJtxe4YSd~l.5Ǐ̬>rIIK Y(Ȓ.x6?'De|ڝ7MuTEnXEi|"<2JIc^'mT]tv#v1i_O+LblcFd 648  %EKc=`!@U2,xZ-6gc? d36䎄z7>[Yb@\\1+ʡ8I`=RTT.+pD9a9Ⱦ #;6,[=H"F=F!YbǻC-)!ka$Uр̳(yi80-pdn .z -*``1\qLN&acɀ|ƂBą P$BDMK WD2# B匀  m=V`{)-Э(2h`G!cLS$ 8QъwIJya&S ,G aH6Vz؏JKxєP1@2.؂P9B @ @0 h!x*A6:=2z*vL  a:%xIhwXUAu3GWrв h)A8>YOfINGαK3l,@4M 5uESfmymdA*}& wQzT.J@>V( =c mիdj6֑'P?_`+@\̽sp \փDNa|[Bo:kMD't+ʇBJ"8R<U1u-Es!S6,ѱle攆6FK1rSH$pUeR @r;lW`ng$jR۠5ҵ ajQAْN)S}}F B%K;Hx! L)6PtĴHJ&P f@(Dҋ^uO0%7a@F,xS)5@=EPBpqBO*t+{K Lac YKiCUnԺj"bQ3HhDS(l]=F LBF1treU}"!w*YmV3& 0@5aVKBv7.w8X)yL),}eޟ;/HBPty 3 gh2F~[]P+k_>|qi m6f9[frOͿ2K7}חeqFsJ7; s3eYDmUA2JUdhmrxŝKi>bqzyu0j85E?~gh֝M0|-X ws{7jUSo/~^>l>ԍ+_ pw{=Spxyݡź+}Bvڬ[WHӫJ|zӷwˮMxmХN6iպٝpӵ_T^:(10W7wG\sX:f *ZOo{?jy{Goyp{{O\>)1}A[ꑆ?$)S\?S\k^S™3RzjWI]~6iq|&-"k&-Ic☴8&-Ic☴8&-Ic☴8&-Ic☴8&-Ic☴8&-Ic☴8&-Ic☴8&-Ic☴8&-Ic☴8&-Ic☴8&-Ic☴8&-Ip윴8d?-hq8G7Z|v-Xi8G #G]x;An Ga_ay{CIkS`P_|ΚoQO7˶[t䈶.u=AY(9ϔ4yruF(w7>Fˏo'u#u#=_ߝO?.ul}wxuw~yuZ|Ywksw>k˱jBY_1_`w۶Ϻ WcxY~GnoTFpNgΎ~4ٴ|W| XAGAϺXѿ-YwAjgtc0V-~zƛ|ױ:%3Cnpֽ Dкߕ.\ͺ/^!et۳6My/8I3/X<Ęsfr 1;=˧׬uOk|7'7%"_-+ E :+O4a! Msև>Z2 omg)6>#O!ݍig\]ӌK1ۧB?=ue?$cWY;ϺDkh:-:ud+eFJ=YYEoM}kMO?Fgxu=_#ÍW'Ld3/9;f6RR55&|||#c~÷ww&w !&gro]}r/fqywnu{.;}y|D=FġIYrNWTu:ŨX[kcmq-ű8X[kcmq-ű8X[kcmq-ű8X[kcmq-ű8X[kcmq-ű8X[kcmq-ű-n)8v ϦxWR[Hؾ/Rj-"kfk?EPB\ũ~C8UqoT׶S,jkGF 7<1+Gz0<;I<]3@4*\nT8+^ratjnؕ i9l[߯erK9˂LkN/gc;.6:ypw’=O}ݲw.cZ}·6-}m) ovxA푳k+T\x ?Y7λ#">|S[Z5k!Qb~xHӋ 9S{; ?. #nxe࣋pge^J.f6 hIX!hЦ9C_e H*bJE I1Db`#/Yl2cv6bAClMme- - -m-l F8Y0fZU=-nçlëxu E&[*dHdHdHdHdHdHdHdHdHdHdHdHdHdHdHdHdHdHdHdHdHdHdHrVEW{Nd χlk j9j{O*! V_"jqgU74oqL5xcjOZL-(R4,S{nV_G||[edvS%:McI~v+۵'(=,ȓI{!bȚd뢕!TN2hHR7ps}r0}. ,#&ee zӇHv=aD9#?Z^?N7[f+eO$Me[r%K?ٶwc949ŒM{uwe~`*}q">Tɲa_=@M' ;8''娕ptW LgoٵFPZoaI4@>Z++^8nȚ'E`RħT}XCWj.Ï:wT2%O5spZpJNヰMuL)+ ]tl}ki_CFoPjz{9v/+ o.&/.h.0gsor{=KEh/& sS7X m? ^ BZ2%!o鬭ڌeUyf`Q0eJ>=nw֛<9edk<꼓m* X[H(G_aU9;N b>P:arz秠߽w7ǿ7߽?'ISM%7d 4o\iͺjڛ7͍XiegN.{|fGMC 1 ?T?^|~3ơP_׋*z͜rz2MbMφzj^ro2o!U8bA<  7p{+} c<#c0[lLÜ+JiZ&y~1?n;s~6N?/x$1+#1ˠ-KfȍO #ԁ2JWS% "G)+$ї̃+c|'Ah#ЩB#nL}܃ S̢9"//;L$ 3ncx!8M.7Tu#č'l?]B($A[QԌdtF>=忞GvO<;o`B|D0/KaFed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^Fed^~9wB%jeh}xۂccg\? {Ri}JRY*Q4,!qx6D &{O RD/(ZQĝEOX"f^߸HTEBRۧҒ~kuor_1G}G%4en Ş;p L'iRuQ fiI95,.$jB#MEZc^6z-jp(MK)ES>ɐ\p:ZZh=-"!-C[CZyC3]%j V҉b_{a2*/,B dܬ\SZ&gO(HΌ<+5zjRD II%K-+;I*T;g5UYhH?vc?QǕ+t{OQ#΀J4_s"ЧK#ׄ>tLib3U(u֫ژH9"QYI;8EWZyK:%0#h񣤇NZ9]Zw%jV}beԭgs+K 6ll|T(> z|'ea.I&d"JAJBcf:BMbK&+Di,T3*~{dEβ#mHs¢#{ØEf25z|ʨFujg]Othg_YɄ.K\eDu($*"2<H&Х |Z6Q1OXy6p:̘ Q@L%)4iIe4)fϚZ 䗆h"bVzJ/I@ z}Os;.y<] L%(7>agb^5_&a[^0=t6]m&4[!hǟa+ߟv3i ٮͤRzh$S7.GL`;kֵ,4CffG ( K_`)vnn Oy1,7` 0uTU͞D5}M켁_9h0NG@HLУ*ôWύ WAX] /F QGO0/W]~+f ,h*vѠղU35U|W^~8j:`܇Et5TrJ3sqsJǂ'ylK'm`13s"CT01skaW%zcpI~/yU5+e CVM:ؾD4u.Ifwhۋ7A3nO/nShRx}GIȀd,Պ3ZM픃!IWByLB,uIOdyI!,)vI?9GXVFE 19h8M0i~8MOmt@4eϺ\[/, S&i<ɢ.EF;ޟL1.~Ϋ v"WҳV.$*cc2:KMA;x~XDpҙ@3'&i|$Hǹ(. ZSeD.45Axgˤb*4jKI]Gu%j[i h){d)O WHxtx,3qeԀ$!SA +,&4E=B|CnFIm|&@Nyô˜ _"1\,j)#[ҀK'Uj 10)Y \xZRXn M[P mlllS+mm8 mc6lkGg[F[ɶsZ/B9'Ev2 k댩aO^ )LS>D/Ctb>;qd1eN2qI'zc*k V+=JY9)!]E_XN'aO:cr(B<<#ga>,(ghGZf,B HQKE}Yb$VKQc'˄L.A<`xqts0ne.-=[ ie%,v(Ք!̷J^"XHƺˑO'ŶP\oz0ȹ8E]C>έȱnE+|}JP.ƍa h;_v`MNq3.V佀L0ϣR ٻ6% %X!XyqI6g}J%R!)C")ICHdk_QY+냉KM-a$9G!eLDix{= +݉uy,xgcӺAy@n|cI=<D F%ճV)[M7{7XH)ǽfhpP%J7bi1g;Ja8҆$)ȼ3 IZ[ʟFI穻*In24Swj%/>Oyz60rߝf.N1;ܝbE֝6ɎS'V[_N<Dz]ՆItQrm]M 4L-y6l\,Ɲ'+77YEz[PZcuv&IݎW7Ty+7іiڃz^__BQNٝ{fъ>glMk>U*gV LRYV{jEgƯUiK̨5*H.vՀEvҔNDN:XUqw~񔍺 gF݊F_^s 38^w1׸|{ww;ws"v2'N~m)Oi{pS91uS}[L696/]Oy ?^DV㙊(zO4hd4FHE-EAb_My6 ,x1mA9.U׎.!kg܀}X6KJTy!){S5ZFB(,!0ĂR:J*ivwu>u;"E Rr"̰ׄ4N+2 (B' L=T˔pⰧyv^$}-6@豇Ӡ1DQ& H ;5#$DOt2[V QdUB $&V`ָQR5ᷧifstҥS>Up ~'Oa.GͅilJ4c HS`)"\ J1uf4 Wx43xu.u^ݫVrE-|II|E"j\ 鹳*/YW>L૲Ɋ| ٵ)LI%ݽoij D/ިL#NGVҢӝτ@W2} yeuFQ. |T+_q0i^(אkqAc1&k-|cQ:E+5}e ?kZ&2K?+!g#)E|8Y?4@A Ɓ@6e@`D=gȔϦ:C%K┲ zvy_q>pv-x1;M?S8  ;!(hg@ }"T+PM"~#Z{ZNe\B $n,E=b1w/=3VoŨjkg>«׿l"g4`+$T ]ή2n+0 RD,BF2M.M,v0c {䬆I`ڥ0-7BOa-wz|=QLy5X78 QZ9JfW&Pb42i=v )o+[gRͩzbt;`N3(v&|7;o8CG'I3ϗ)eu(\Is$7qhdJ.y{qN};'(=Aii?z,(9 iϼԂHEyGId)PMZm Em*9#"(KlVJAi6Fl,t>c-NN>-%YtFy_N=Ltd:u 5bb @*YbYΔ-Ч/>%b}ŧN?_'SGwf&+zpp5nD0h2+f6#'?݌Z.Km}v3*TXƠeQ,zVr&8 DutZ⺣W Ź\S?HCm d-Tp$6TN_Kӯy_Pnv]܆g740@r{O/ҭGW7p+.> VZR3\" SƂ9E]5= &ԟW^~wd&#QHYu2LwZ D佖豉hj4BZ"Z+Uyg ک{ 6}x[rnس[&K5ôlq}:nkTʆUkt"k^ȍrrg+Wz\Y]ҕg\JZK'LD~ )?#-hJq3V npGRrܸ Uu9^0Z#Db3B [J.vz?}&Ly ?C:Kkz8wwfu3 Sc}{my7s"4*kc&PfTemT BԀDL"76h;X|Ab8^Fa͐ v#dˢƗ,eQ.~ HURט,*qG72sXlL?w(?<L@N/dSY`lgEI}X,\E/s׋?񢗈w“nx9BSX˧K+*6-ǁ!Iʟk:<ˣoӿ|0arroDʵ!Ǎn WU=&mh-R܀+)&[)Q[W12#XP1g&k9t:操 tD{ys͝ۑz3lhـ򸝢 >OP fh/#Et$LEt!"'U6`. @eE.qVq3 A2 ('׆# Ҵ4roӱNǞi@ QzHLdu4(&9bkC&-;M4ӴmjAk$1y`-v+Y']u@`'?CA: ZJ=?Cr*EgB`\1W\v0*Q+\%*eg^RD06Spz$C6`TWN7Sh@s#,CSDB'7`\6MϡzЊO9^Ĵ& N~{3B;w3]%i qF$lp PV;)uBh9ѤFY꭪{ƹ`cxx}{/k^@YRgL,Jex6U6ػߪmUq@T{l?~|H{ɱ6$Qs-9˝R 3h\xs_&n#u]V1V0#)6gF\Dcr1ehmWLϣe_ڝ.U8▱(7Yt">Mė<`#[n5]泉Am ";H.Dr\A%Kn{rH.QĠ/1KqH@B3PUvt+2WxsE|@PP^Ⱦvו `Hi9,e!@x$97\GQt*eSE?^EeFHO\ !=@/ 8O%!E(p"|tK ťhO ]pLmB&9kKb6uT9v +I"EͰ >b[DMQ gFLT+v5bó  YLbC(ș3!Sٳd<>y7ai009 %XGSh-"0ʹu*481&EF"S [du k F1M84!1Si \9z:eh=GC6bTI2J5Mg'+T)%Npbp6D) %Lx NܓoMV䦙 mdIe}JZsE-$F{E+ITJPYf@$Ұ*Z&m2quR{^>hR\|qSY<[㏪"$qw"L7ܞ )GV?Cٌ2WY640}|y˦O_~??U4C;;t2t#2iG/>i핇X_ʟ)l 3g|4g"SqJQi883ˊk{ nn$$>/ׇ6Eo@&L{3]{o#7*@&Y|08dg7% |ں%G'_Ւe[-i$ʒ6W7EVX*{J"}z#/O ՗/J3T ߪPʆM-ͼ+B}N/B $NΙ pz-')/ic fG3P[rpbl#r}]4/[gBBYů>+l]XQ+vVe8h`:&"Jț*(h*wx! (r8c x@{BwV#݌#\ዞcm>&vsvYrV5o!pp68M npI8EI!&)1cmk#8d[`F'$+VoykIU7fk!MRvtmK[#Sj Qf ]ɁM-xD7;Gl!%~QigP\CTqZКRO |$R1QqygLD.0*.Uv/ l}`0>EDvH@HKQ4Қ_]gTv E#bZB,*gN^>'bĂIUl6m;V.ь${0A @IJ"I\Bǖ mz@`l $ TR$NeIr.<#I3Zks @b=gY9{ FH+ g^];||_<[c%b@i-),W9?`|RfrIy5S"b:TJT"LVY9OZWD墶\p-ʺ]b8Sp OÁ2D I$bh dRD*鍒;Mw e%RQ=x nm:j b %D'#QQˉ #5.I1PvN9.فK -^ץ!\%EMEsS:)EN9Qk(d F+k>݂0ןA3xddJ H+TCSny҆ o)#!JY kqs0N i"ƒވ$zdzX9΃O$c>Hw}3+Ghh˝ S<__] ;?|<>|~ٷ}=om\kǛ?'Ӽz'o:Ltzza$rZ??JSd9U2KF h wf`Xi#K>e0URQNi^ g;30n`'mcω;Y[<@'F5 K~l'?FWE3mԡ2J#^UpOApgU~rނBtɼ. tJtFS{GfsUcyX ,K%!l4Є^J€D-Bu&$mBل:b1qAzs ]|py\ǣٺN0@|w4]˧]O_DoP5YT3^F05bvR xdb5"۠RyrBd{ՋrXP"KmgȾQLwo7a89_H< a+NJ+HVFVY䐸bTy8 ;YjXx@1-DM!HUF2ErDY0Dpe1"Ay)pFIē\x4,Fk&s @*# ][-OnΊG*=57/O#v XLGBࣃRMιL.>Od+T\_a F*g5EioTO|?\?Q4NScp :g(瑲DB0*QHk1Y|tLh\ǷusK4mْI&WٲqI;M\g]\Fծěw=l/ҵɂiL]ue>'=9A2ۧlu<"͎ܡGt,gzȕ_|z; 9X`ǃHƞhdd;,oՒeE-6LjQT}U_u>=]UO'=a-zlDYc+jR}p\7 l#W?WD:έuaۭ3e7^r(Ա) 衏[w"i,ؔ={⦥kN:48ku|$uwյu?#H}{r\CqLr.}?mƋ>Fdu ״j j$(h,k=;Q٪#OJ$O3'G+ QZ ٞWy'U"b>eq–<=s/,Gi#ⳬxr/^}.|{;-jqR:QrCZguI?k^~ T Bj :k` 95c-!E( AHQ4 FE+\$rI91ej-p}CX4WDSɘrETe![J{C:2$fr4ˈcoc8c*"I+]6hH!L2Dg[PT50ֳi9"mDڈS"vPT'- w;d_*J㬯oLi]JUL_U8֏V{4@6ʳh2 ƒ`*l,~Tac V8Ge2,dIAZ"1̐o!)=ctsLI$iQu51 tqI%ZdƎȭ܆XCBgguK$)B`$樲Qgi9g<P0D0UqY )c9qZ2pIzoA\1ȐcB;Ygզs$ΆדUACb%ZMG+&I",F+r+y@M<z/3@& K++!<`PBks )[srV5wOȃm1$L&p- l,A`Er&:GS K@$sŨ8(aB%oz?dǘi&vŅ&bRȴGY LhiȆHC3 5h#Bs4/z1UWw <| /R ,FSNR8OqyLGcTk2֗q|tԒCTFxbXO#iHY; ?%%d0)Og-5%kxsC2)5FQd0ʺ)ehwq>39PʎᶯUT]Sh _zBBw7Wi$%#S!ĸDbU~\4 0:=;m=!B\]Th;Uj]b/Tq@bv!]ސg.ݙ#=3OaŁg>sOC6x)uV= ԍ:,FcqXRK1eb7Z G<[vR c]&Q[ }mq7PZ'o {.3,D%,]m ^-Kk&KQ`VkGceƪoNx<07?2ldㅵn6#kԺZ7iuՠHny %1L+o5?\i1RF% 7~Ť%i cj]{N`8^4wi ޱXR9G'D6`"2YB/bL(y9X*E]$"ǐYiKXIr#g94GZHiA{bD26dҎfz_d8{u j8$G8$͑C&2Xzy xҎ^eI Sr}Pj rE,}kJpAJLL)QpURsdDdDʀbO IJD%M<2} ;4N8$,ych&҅fBJbA3k 9dDa˞dB A k/U+ 11FZ$WmbE.@a$ H4lZc7w3{̓9?[;eh]s ::?4p2f9b^qJ]F|ɏkF1rq3H9qrƧ޳_ =?ĸ~ﳅ1Ca͝v(6g͍]wczշ[[#Yh􍗜5Cjul|1H<>ݽ֎[΁9"zE1sa֔L(t xjbV)G2=wW] c3+z,pw4dv$P$,B\7@0qn'<$ imf9yn22_;t#x'z8.{ _[߻X|>qOt%ѳ%2we`A v/r@ӛ/ݨZ AfE3,dGdBhT0*ZJ G',Kʙ)3PClBXz "MƔ+"(%0RnF$h1Ȑ8y@wb#F!9&$wX.U" T2p+ JoBPXi#NF @Q1"6/,^ߙlq`蒥Y˳qa$ 2H >aY\ ïto~.{fU撝|¿>VvuUoVvu$YG];E;3$ͬ$$$m⬵O?C?/?; 4](VoaHUtoBO.-*"/z>ݙ-juC?ϓ_WҚ?ww~k V_d:RLo0_>Y}DLjoBc='4!i|z;y qΊ|g|R%p2^zՓoo]bdA̕/ҠqY,3]2^獼x{ibcԓ=qqOWMn,aG Y'&z՟nԢW7:dSc5y3  ` >7}=+ްrlhTO^5/=];_>oo?yP}o?N Lᴉ)L?FSNݻVYL]m]Um߻bB >~,$tVgAWޮ Wo?bKF&|sU( q0.>VزQGZ~Cpfe0)ΫS67E}y79~HQ 9 8xlDZ3ᕡ?6vrp8;P?)}J(%$2W\\SY 8όM[r%PNN{nlF=F{-GhAZ$woV1SB/% G-@uLHsBℚX[G(3nc I?mT\ z^rvC[*wo%O1nVreQ݌ڸg8Ԉ8@Q$T".%WW)3KJ&"cLVq86>Q$ 0 ‸}\ń bW`T Km 8"x d9 Z#g;L=q<<[:.KY4(n,2vÛ"@_[*zW2u5s줷Ez}XP u:0ˠ^N7x?{O~cnd:U@[N*kz(٫1Ý<&$|ۚ+rW8%˲;2|p/ߥF!To"ss3'@#xLAա{[72W-5 =b$"Y?tA,4~:RHe~]JUܐӭë BԞ CJ(_,myZjG;_x8r QzT/ծ6Ꙩ_-U&ORfag ]i<bZpm1- 5 䢶I.Q1B;V&r0%f)BʸINFvRJz`Qi:qh0q2 񘻠 6KsQ;c:OZ#r{iv ~h}CF/D^,3ZNPu?u~ꢘyi>_b>g升c:WjɖmзL1 UzmD%s4m$]ԡD#Eဌy)'CH68nvIYŽOqxn,+pjQAA*)yIL09L*vL;\31k>>)pf!呒7s V38r)F@)&ױo,pܭ*4!E L"x'?]'!%$ fIh\X:xfX;WjG6t n;bie~nMrѴi[3w;}r[_p}e*N*?'~]ÌNLS^Xvqo>Njbw7Z?3L/Y [wn[jL*6ѓ|<톎fqaăy]tXg"@'z,Z.nQ$:|, _toEM:%I @.0pXrZaN(|K`A%)1PzL!h.]ѬVfeC@pk7(\aU7XXrݫ]da"|)ݰ0^] *Um_/9οʱ[L9NW)@'*/S{:7ΫÜ' 0k& $SX%A }XsJΪ>wi;L9h|+SZ,DiBTMy?߼G&Қh_jmXfGug WY҈$4¯> %\怠}0r2[ SQ4nr͚C `58ÌqR0&e  {" !Z,q^+!9U$ {L{Nzc0e7SQk,̉jA ZiYAn \OpJpI2%}# e M6{<m6EV,ϏMR= @:!:hT ,< VXIbX( |#ʱlUVt MRH(ŁqƙsgՎ1eT px˲9;F܅ (\yϙ ̂Qk응IC 8Z)p6+M|C4B,u݌c+J;Otul >.VdhR3b1Pξ~\2FsF8VqX38LAw0g5cNƎHƲ&}۲]}fC mecP8SYPL)݂ɏ㤥Z{m(NY,q-MhO9bǜFG~Jţf81I, >)(#"Z9x4BNt5ZE2&g`KB$NdVN.<Ş8b>QJ*RWlZ#g;G3p~r羕//_ #?wG11ش1ԟ˛%TQ G:qa w8"x)^( F1jqt\C=Ȅ$D%6R̶ /U6F)X;MLq\9C8&QMf -pY,۲B =Avt=D=ZCHXk[e얦Sxifyn \#r#iUbL }ғS 1Yg~Uo[ jl 5:n;-/^n#}қ!pmnot{p;H)ޫni8x5,94 9kΆ?{uWKsȮ:U!˚[H7o_Y ޻5>wpuڸ .bPYd" Cp|tNwt#ݭ('mX % 6#Q} 0~_ll b{B%R!)^:#K)5WWU^9I),Z H6S6BF# 0[.tKZ23N0ETJxЋ#B !*cʝs5^iP_7*wx6!NJ]ͳy֪dRښRYK%$0hIٕL6gZzL#v*p zQe%~j{A/8Yz?8|:2AX}!)\D AC2•ڼN2tY4 p}ȁ}D6>8MʔQ!fJf@a&]KLH|HY%&G}v. ="s3 $wY. #Ԩhc$sVp"Dǖ7 %5Sti-kٚ&➒-J93KXz1B*:mUm=N|a\|TL(d_~\ DDO/'=b|eR:ۑ;?^5n!vPmq0L𤼤RjC冭V[\t yGo3G~*+wI!o_X Z#RjPڅk8&y@$(UHZ/Up Yӿ6`<h@F^!Qt"4tNk[KOZ~(й٪VE:)?N=1e;37 f:EMVӔV76x"J埊=[r/s"j 3PGiϭ_=p(EO[n)1u$Svh|wdGv E Եcx`.:6øw 7kSR9 Y)p.Bw]ft)6mM_pxk9DLfQDV22U29AZRuHN̙jt^]}.P_wY(Q2G0V&CTt֔#p}uMm0=O=⾙Z2D P,21@},̞c>$!TgɲP1*,܁_g3.>b{FKe/ۣ1m

~jV'|/ՑoDK٧ӿ&kdΓѓC5HJ#[Q:l\DN~ĴvĆ14ܝRmb:5U{yY ?pW͉gɁy`T|̡~Y];.|g=S`lHT:1_4aabKY$Ć FF &L=;/j騂]>tF]:W:*82"`ˣG$rU.m^ sR;ȝ"G ]d RvENF8>>z(;[n̞t lcm8ڝ'=rzܤKZ,i:OH+Apju59/?$P5L0Mj0J~|uow&m7ϓ< 'k$2Akm0Tr1~aCH4C~QƼγSe5 n'k}GWaYpDBzXoH6wE득_6] ];ts\Nw'+&RAлqٲ6 sh&uZ6TZu ~£_@"eր^A _k΄oTJ 5 |[%fR&řV FހcNjmE 7)O9V81R(J nW)J=p<8WqH&+mkҝ,ig?JY~poL8iv0NvplB=&#Bep\J/4[`Br6 NbKwd6ؒ*n\EXmЊmط=E_|pI6 a9@Cߜ_\n6? hwَ8 L*ÖC:U GPK^qJm9"suPYRJ6e`Km_K7ވn)Cr, 2d# OL.\ɂc\+52W[r*%KsQRʬnI 5J%`u0(Fk*}՘+' |:}7!}tټe}7O% 0RfRͲeewEfzN'ƫtjz7u}ꞻ)=Y}~n` |}|f$NԒm543T)Т.n]krcM7B}Wvx_*_;ٯO \pҟUt[t: JޅT YkäeVvnݹI=R 45G /7dߺrٳ>ﵔ,F0QO쮵8v8u ^Zydu!gЀo184I1[*}Q`&Mq =cVa`b TJyeҚTX5kRuj]2RlawH)PrH3R4a%е%Tyoc)KS/igBeY#H4.8FC' ?$bU\k`QCV%:AGYz;v iG#ϡ]gH䘪Y@BO"L:TL0:\J*ng*UM}RIkqS8|-_8L%kqZkqZ^QM&֤ܚ[rkRnMʭ5)"&Vդܚ[rkRnMʭI5)&V;&֤ܚ[rkRnE[MʭI5)&֤ܚ[rkN& ~BIH0g$frJRnCOT Sr_`RN@qCתǼAR3}*G b#цG+-/ArjVy *pDˋ*&kH6r>9O)ƒ2m (4X4TJy1$--Z$, 2@IBWk"t<h!2Z-Fv GY?r= cvJ2RQE4*$` =M@47h ")\$P)c$W,G8mm[9QmEk̴^B:Ӵ±jG{Ҏ =$M97Ydx&$4F xOq+b"6'kf J%`RX(qT9*x XZϊtF^ Κ>v>0yokOl~Mzf:I/̧`!䣊eWd@ffdYWUAȠؘ1uw^%NۈH-6wv -qeLOZ݃k4_⸞rt-I @ 1O<#)I8k}NiqEGL[+ MIs=wah;?p31ٰ{1j9}Inn4KX]@ӧ5HЋ;.F88;}ߝ MV84qmВn^̠MA2B3>f?C% pu)5B6&@-8VPA}1ZQI!!PBt2:UJ\W5R0ft1\k))Ÿf[ў]bSnIIa<8I,=кl`݉$zdzX9΃O$c>H=|'rǝ׶0DLq,|^0gP gMŒ:yu2>%D`hiq&Rq:?lc;3%U•nb7oǧgJk6(I03.ဳF:0~38tJ|εMLnFusɏoǟx$FWĜ?|,]p!9մ}b;NȺ4_;G:]7 [;d0abОQy${]=7LNjvTF6:QH QGr!> }~OczD;ÍS,T /M^O>gdo~ç~}O?}eӻ}[: E֑ ŽIЊǛ 2(92*95ބWm ڮK@;~&mBf7Othwҕ~Wl SwWTm*MT!5mmnmm6}J?{-lYj)u(ɻӾ4Ǚ2 `7Ӻow gm  2ZgЎ><}#m!@d*f *kuM-(.%4Nj=wdv>7-]N/麛k@;Y3^1)'`Bk&)d\@T|&Eiky}lo6zM"k~[1{rL[Ar^ &U ã-w晱ղy.|#"HHɞX<,8e5ҷǿ#L;;zoܴ^}SIJlD9|ZpC<'~x;uv0Wn/6cu_-+ md.=\5c|&yߩol^rnw@ݔ6-A/67Dmĉ 9)r-,m"ម gpqs3n+q<9 _~BuP`j [ϾV8g;‹^W|Zec]fO>Şu@#~ʯ/+uu)sb>j1 EE>gK _|?NlYe>h߷ZxSGWx؍YgdGtoA)+@"`\p ?~}9/ @hB)'m8oTkˎ>꼟z!;iw p]@XKau [ uԝp? J<ۖT笼W5ȕ+_Sȏ~vk?[ˊy3 |Zq8_.ˉw~^| CmNDgۤKUn9C܅KGʽɬM`̛|v7IѾPR )'C-1,ܲǛNr |8dz~sAn`}~hO 6ekÔ^fɉ1'zqz {Cщ/6qj F4OGg]鄙j 8Ir_B;>[XT%|hs+GjZ%PD(+YX% -8 灐ePOy5gAO$-VeqyaH\KS=?vpص/Zǃ9õG $dZ›C &dj`r)Nhِ2as.] $gښ۸_ae U~$U9ڇ] WI1E*d[>~3C IC, LwZ81 LpO5Zks wHIo,)A ];\2]#$,2rSj|>sT! ![7)L)+<[|d<ɢMJ'h^rb<+K%y)vȼ;a^( Fb4̒n@\8>o5dF%*Fi&ruPHY"!(O 1ٳy9+#4%ˋޣg/rŠM=uDzBKj5uwKQN>nKnwջ_}\ҥuFm:Z^꣞ SV̭u'W+ع.]n-~oQu\]>]^x~vKGOV1N'h}'03;4ᶲ "T683JTn dHŃQXJ.@ 1B"%IW&T Hu@2`)!#3"ALjM<(ÁDVHu mm"YY\mO-]Of?B/#_G\E-Q"R j ѤhQ9XDRST)Dd@J؞%7rKv lU:$g{+_k>Es!WgS r CcuEs ??ΑpjTFGZ\RX/$š, b3 i2|zy28Mjok7Ǵ Tq >DO~>讲2Ɨwe9kJ^jf%;j3cGdݚw0QvJ Jygܶ[R&OCuLMy8g5.j )O4IqrEBp1i dm 42,WԦ"㓋9_-Gp(7HZN=ǽ\DqO-n.nO֬6o|Qmw|w\+/?8^+i^EҔ2YZ0mUJ@Gyhp/wW(J!$HƒZqk ;4cU Jn)`*5co,׌gхqƺP Eх[Յ. gGy(o/^Aeq4v/@!'@ kf#9Yh\*/IzbDD02z>ٝÖvxI6QT;-(;ft< )D%L{V !~`]H"TT!a/{A2f4%D*@: $( Xʂ*h3P* >#!cEҊIۦmA(zD؋= _Qa'M>Hd7\!%]OmPZ*=cꟿ}ZݳMW0& NF (,>^8vbKdYy}ɷHov%u`#&,0`_<Vl]ܣT|d{5,j|qu<>|W-sv֗UtqT2>)w(R Z b}4 TI6|y74'OKOnAQs ҤUF"sVzIY/4.D!67lbWOKŽ RgF;X^Gy4P*l2^=*3b2ro2V4lV))\!C~?2L9K1G*{*aI>zmr:Q_v^KoqN8J;K)a12Y'h 67qDN}}Vó\|0[fԫU"%΢CsAnH(xHA 記pFΠ>1 >ڔd%08Gt̳J xޞK/q!"< fA>)PJ@i(Y4 f,PJ@i܁z64(P(Y4 f,PJ@i(Y6 f,PJ@i(Y4MT4 f"cJHZ"iJ@i(0R9ľa̟%^j264 (5BV&@-H뎂f#}[`fӸ9A(~Yztnmae#[qbЉBB XNFYG*Q?2G#N(f^/t ̈́RH#biYm\D"Fn3h R7rVhutLɩpkŕo0b!u;^d;^˭n.WEI➇X%hFvn"I'D@ jmJ*Qf̡I:~P$GbQg )3z.OUUR oTDD%YПdK%p'{< Nrw/./n +{na]$zdzXGs'El$Xz 0S&Dm ? g|5=ִ hM'i  M,QȤ͉,diGFu1*ԯ z;zÄX9tVcٌ_sg.3Ƹdwu: a Ss/sSKߧY1709*Li|we>}rk~KB=D'apǝ'` T;rSd`rj M>~גD4O8|D[ANɧcČp rښ G`T~ίcbh!9V%*ʾWW/?S`,'tUF4`.WWSAQfԶ41#Xysǫ˓7}b3t'rssEh߮"955RS(@< =r2x$F휫F[8<@"J|"~H`\CbSat*'tv:j _~ѻ7o~}ѿ~߽?Lۛ?_p.Ǎ᠋)6&7>b} MVZ9vWW9qf ? ~(=pt=­o͞ـ|dGadͪL3k~*MfTPzf@׹pTHb}>m[qt6[6qJJ۩;pz85F_/@} v.ƾO <1"ʰ:/$# )ߔmQܫ2h *GCύM[r%tN=T?Wٌ2 VqsIov SBQ/%aD-ꌛLj/w%'7X3'~sJ_*j;X؀%y4!oK9|?]WIUH]V قB@`+stPuu "[`x4D ]T2Qc b|bK 8 8L FL$АRb JqW@p&e6Z,YֳͤYn&NVu?=lIjj".L9ߠ>qSOǯOj@5Yfhf0 fה;C9,z\^Sk"d:,9*ʠNA+vkطſy&*ohL'|=ŠeҊVFVY@KS-aŽc+;Tv8CJQsB( DLQQ<\t FBPƹ)p*)cP Iѱ5)bJ*S6bo,QCW4 )ϑ/v:)GU{jɟ^t>tdԼo|yLLEO"crAT'[MJ'h^rbɠ~o!K$f0KBau:r`&IHL2a#_Ez5/#< %K=it>nKnwջHȅMcFs ]^?_>eөuUzqɒNgmqūYw-+l9|[E{;o .\˷u7}8fΫJij|Im; A33֜Nt՘K_nQ?s+3ThP[ oC54IhraQ‹!8"|CDL3Ub":2bItC:-@Ze Xf.:}(f_#I]̌<" 0~} д%u7HQR,Iن-ʊ/㨈ǒD1Z6C:xY GC1M$&ȨCkgɃZU+*ITn_WyJ6zfY'w&]=s#:gGt΍9 r,*bujbHk+WMy g0`E,MIbdT`r1Mf7Jl?k&O=ٕʝYf<ׇDamvlEqמnRӣE/ڨzً4or+1 `Eފx{xp?i|aaY_U>0S{~{8y.",ˤwn ~4;Y2h?-c"X~F-=g˫n?ߎE棪fC<F\<ō[\Ͽrr܈ł7Yc b>1ˁrT4AeÔ1rDl_9\Ȥ}I :@G5vh=f“J\mCs\tɢ㯗%eՖJ|cJ9UE#Sh=8E1BfRrA@RuarIXב {5'iLĠ-iPcQ:@b-r T!R&ZH5) R%&mѦdmFgAP@xV"Y s^bsX@qL-q6e#7(*=J ƫfCvJ4ȁZ)) IR a5!2\_l[ Db/T$T2b78#MƒB$$U`J'8iehjTbQc6NhTtE"il;|Q "HA:dTV:FjwzۘiaJ*S*RS&$!E5`"NO"@,hka*C[7z:5SO6ߒsx n8.+NFǫAQC4$sԍQCx8c7@G+ y.:TctVÚd}=2L3GD- z5ԁ;V|OyObΜ:t\.mujV;Ob8Cz7ժ{_nz 54CD)1#;E ى)N+Wg+0CN:rm*_Q]޴ۦiwڽݶߓ:h{7:!U;O/M0u#k'SIgNX)D`K!ȋG3$-/bRbRfR{cA,DVXm$a! ֩ C _݅ LKxÔ5|qMy?xH{G"Eq$AK8s69SҢsv}zg uҢ˩zTʠyOL0 JXN1=9TT8yT i2LΎ8E i88m^'- yNcGc /mNkG}8TZEjӆt!s yʒTުUNj,8PH'' __O˘v3nv˖39'nO?|dBij_R?i:n:nF??eU9׳Ի|=e?,}ڢ~ܚѩZ|T%L/t>p9;2@ff္YwbϠM uL| z0;Ք>&ԓLjk$>.zm{!Y{\⠇|}.;8WɜQ4/:+^eAE|}J p~qr6sqEZ0.;,V[|kUt& j'ۖˏxu*D {_+trq% 7EM݁QAnrs0D@f]GA)QJUQI !gVԴ`}gρٮ0!Gc?L79nc!{֐nQD ^jR{Bc{|ڰk^oD1"Q tq8T $ V)FU4]x(F%˻I9PVrMRdfH%s Z:@LATNh3;r" mp˔\ 02rF{QM`H6vF3&Wڵv~69I6?8<]vm5i` 'XL2k)1Xt &`e1 dZ44mc9JJ کJv'\* 1;/"3ZeE.HGʈ0EYD^4 2 ȡ`TW(ZI-0<`g r ].ۅF[)qD IJW\9 K",P!݁*l-M.Doи|6f1LqaLs1hQ%|֢j67"$>bUY5\u؈0vQRM{X{׬֏ n]m,dPO(rd׿?_{f~/~ywHGJ}^Dѩ!9X/ NuZ9]0Ek|@7C,hh䠗K(=.Qm\":H!(rdB HJhu6yg lPQ &1w6 lF:n<ў- ȵ|qR C548&O:Gc9c}1hIkpyq>i$ q UԠA;}JӝA1ץrNz^҉9:1ǣ+z֙8V_&c).a\*/g3N%])M+v@r 셰X4YYBX\Ȥ}I :6: C'O)F@UǚF M'mF;hRX5Y;U /(xx |Q|R!THib6E4>E1BfRrA@R0%uOv>{(y 9*Q)-$ʚɆI)dhS6c (SH-*"߭؜k86Cc5gkz_ofˡ\K]{ $$W·NXIiRR|_ :z6P@K\_l[ Db/\T2b78#MƒB$$U`J'8iE/ySB*(K i1G'L4*:E"{FI( Q}hqjM3xmg< V12EԔ IHQЕ)Td)I1%m2",Ep5ӱ<|1-j9vq,_q u5E9*=o:d@qHA2J5Ǎ36oAG+ y.:TctVÚbdć Cn:lZ,H&*6^Q3fؾf*ҌYn͘,=6czA݀E?.̡EFh$>d,@9)]1x=*Yuށhf + ߓMQC]cu؋˪a!Lx]Š'S+;tbJ!X"1D^<&.Kf]}ͺPu6낲֣> 2J$kFBA)ʥh~0F*P cb:[Db<"&R1:_9Pkf<_݅FTb_{I/|A<5z"HQ? U+|%tWY:_I+H|%KtWJ::_I+M:w%O>M\ԇ-}wDIR2U"O(^"[KJz G+|%tW҃w$%x!:_Iw:_I+BtRĥ}E*&?;.UZiƞR{'?. :>?~v}:ܹufjg&.ӗK% DCR8+fu S"A Kvӭ`̸%nI 2kHarv|\ [1f0z"{'ZbGc /m +ȣV0i1B 1%UAY$0ɁJW xy].grf̛Vm.m;|#']T՞}-}j77 ;{YM(q#&)g ,+͉ר9Pʚ%Q(.9eHnY>wz~Xq?m->bT0_N\Pk/8&v\3#dv`v`+uG\}<0:( y3]o֨֫٩1dR['qOmUL/zkz>˾ˍi[I!. 6ԸDRݒߦϼٻ6ndWXy٤6#h\]MRnڗupiX)R!%1CR$^$5-K3CA77liFgLSj{ EZ?uzl<378?Y=sRiq~͟ڽ;g~;oj̚ebgue{--yu^;J+u! ׾}gn7ށEiQ>pے3nA33`MRbv!/LL0IF)OSR3 e#/1!fT qNJmTwC-MUO5oxb芥P ]OO= 8Ok?kKyJL*h+iƼSrh2" oMD9|uu=G?x|603f3>04dW 4r0Sl$rȱᵬi:DH) 2'M%K=f *duـF'̮*l!/U$eG6gh(Q+L%d2['UteP*`mwc>w8`iGaS3^?di$OMR 'bȢ-.9).Zt.3J"gAH,Ȑ=# >$!i$7jp>"xC\q;<L>q$K_,a?־bpkQ2%h|t Z Br(uhiˁMs+i KU^7Ҽhy]ώ^*g׉1%1'pGŵG-og5֫c?^NI6`ؓT7tnVkbÄI-h,|2],&zvdxepڪ 7qNsX:+\pHXdK3^yPsp^F5ܧkO!o:y9ڠZ3o矾"YGJr$xX"i!OlMT U[ROR-B,iwS6,ѸDmVm6}=҇lE&98+onEq~}u: H^=/!SY *'謟?:bP9F&W!2'%75Dp>C`HڐHPQz# \լ灡"7E!+dPs9E&Z3 ] p)>JS_;k7%2j^zYp]$JKyq_$ eV*L IN_O[K*dERL+\#fo\& !!J`}|+$neܹW'`K|y4nˇO;w}&VƼ+Xm&Dlv )%ݖf9{X G˽뙷ko_HhjԆ"ծI Jq| cbYP*`,2k)U,6fh mҶFn$t&;d.-q-mlm(~<+AgpO-2|X%wкyr>/*C ꥐ9r} JZ+%z YǵW/iB9RuHTP/jGq},K8A`TJC$$By^DU`CLAr),%g2;H4A3۱_b9>^uX[?-!ӋH @l6D8Fݙ\Y6.ӓӳ9S( [Ik@ uթ/nvvDdпy'\[xl^19nk-_>'duǚ~.Gw4<41ЇC_ ˥}@l$M}̙>,Ĺt(&OdK8h5l;hu!CVjAt=Z]V?h5mߕ1M\Z|q\!KǟOM F0Y%F*)Θ{Wcj]u5z7ʁ[#rIfD1i2W9kf>;3ZZZu5 )$e9M3p ɸe.^ td\ˮ8 >V|Bdhl/U3ouCyfE.G<^dTd` H*p/@3iʭrpt$-%:4:εs{-:U(A2Qf3zn,rden|q6M"{_kf|m+zG9,XwYdU~3X¹pZo=%b`Uz"?`FZi/Ъdr*OgS%F2){do0is$ږ[2nRD2* uG*-p?Y~v^<{&.fqy*?M>/\b{ţ&!ʚ eĄ(52iJ8M.J>xpFϕ# k`iZgfhozͻc~<8+TAu㊻ )agY[Ũ<:Ș|?* d^Qz|V&zR~He  0E!)}Z2 Jz`gL!MZLkw}>|8-؞Iy/ *l`qFC1?8;@a:@a:@d*24gq*"xCdȓQYCBu&KܖGoGWEñ IzQe^;%& d.Xc^"^2+WnoS\mڍ`%э3a>ۈqu2B14֘2`w@U `Bm*345m/wHs35"ESsao^'ڴjtmA['X$:f %Z.dCDk{IS`K2B__㯓rlfd&ۯ~oki0W?3O=XKh#-;r!Wl}Z ] +~{F[L=˵I\&{:pZ|zլ{CCJh7kDQ:lv8?;xݱu+mwܱwesl)PݠMޱR:,sNNN2[m"%Lk5&ǔ vNYvK"3`Dq`\YN2#Mg``%'c`wG+^ %H#MuiTgq{I7wRQlzӨ/D p ɉV w9蕷g`r 6ttPpȅ HQ7(:PpZ8NΒlWZ#gO*H)ccPlIy%DN82a&f.H٤WV(]"t@]X+cC]7rn0`qNZ d_vK]{o#Ǒ*Id~?0g>#vcO^9}R$!))Yh%驪UuuXw P|\UQwc-o\@h8'QQKP `A9(O:phqVcj53Xdl\:UeM7Fh}J[[Ai6\\>[2'#3HၱXhT /lPqx>􀣏a˧ t/;c !eGC("d"DB ̣!#&1F: lp6HF@Rz\Tkn1LU6wr\c:۞~(MIT"ЕOW<=Rgأ#+ /cј3 &c(H^rТ$Q -9+R&$5/ry/7+Dy \XW"/H*4\&etNm]J{HuH%`DgTqׅV(n]"6)"p]DѩC7Zwx!8;qb4H qx%f>L_j z>%Lz?+&tș0bQG"`"RSFDD bIz<`pHVpy$s mKAټs4sW_30v~olR:+y4^B. ̒.妗N ,Rg2D'(DX,-,xGI4 GrMZ{YNKݕm 8]d@`In0SWf-75mm1Rrw/$^0O'ku/-ru~b♅bkWwʱ줥tQZRsM 4L"39t\.BQ ^uI˭{MInD5[F[4D]73 h ƷLL \xydcy{ ErRo}9ޝz:mDl4<2IM jZfA6=fU#Mmh7V]n<ت6? oM;#]]W-{)nju,}0?Q=zO1x)k{pS0cFsQ =6dI5}G D'[< i\_|3Q"ihZ2boM 6 -TI;%\yyrt0< ,KI)u{K"I XnK &a4+rC,(RHᑖԄ u=ן;"E Rr*9̰4N+R B' L6;Ac9 UᜪAGS ^m@cO4 e̻(Qk 1N{ S2=]Kqk' L_a"8KX ¸4q j'#{A+HOqsf{+nNގYUgb?/\:cXjZUx`o#_;[$=|'rfɧ0Qa@ /5cjġ;ӻ)}2Ry"Hn]ޏ9ζ\STrmu~mޫ.F-BKHm2uW$9>ÃI@NҩWp\_%~Oyߤ?}퇯&ӷ_Yp)S L?#9x0ВƛMgh g\[}(>oS vVD/o~jug6̺8yHAf~ .8,颢okT) ̷]J(@Om~oncvm)]1HRcIө:Mp8_njhcd~V=ϨG!=Rr'5S~jTz@:Pq"P'M*"Eϙ52e?j3T$N)HNX{#s\Tǁ^#!ȹ ;)0v:GoBeلTW6xbEy㿞ZQ]EuV1`oux+k3U>\MwJ>ʲTp>ׇQq]?J#f e[2VZWY}U`/UP|ohAL & m #_$/sc'XF׻GշAʏx>Aͯt\ 0xwhMa>8x~6n܁5 z:&o{L^ңXYC7~ZN?fsFhPH-SXJL!0"^}Vb< =0sG)"!D# i,vL(Od9:q:VP+~e<7?P'yۏK7z6%=Mx5``Gˌ]r WRdLphfLZF›ExcTkh[n|wߴ2]oŮԜـo{w̉1iIerc8P E$s R`42%to x}2Yj%g!YP "sӞy/8!bXI$J56X΋)PMEm*)#"( KlVJiD h̍#.V4t@bڇޒ,Kקߜx9Y:?`j>=< T`?. kyy^VZR]0\"/)Jc-_:/y%Ah cbĠ4BTG*Pn': ,A 5Y x$ )[X1c2b=6M%"3FΞюn~ rݙ] WᎤu"*ΰevG*דYg xLFV$ ;GUń}Nud7niwrۦAs$Tr;?L'-ͯN>19.9J6R]rl?4iVafvnߧG~d{}@+vbgF{2e" ?9(,&!T Y%iV"h@QAD!M6e*$' Utdl8q2cۑŵ+_rlSԦd][',98y&~3+K9癨U<<9O&,TÌz~B:)OU.0`*B @[p㸨3UOc8]/2[+|0t>߯<Ƴl?VYX^:xJC !$Ie)S !*L0cRlV_]`Hj2]cB (8]`T)JFn ,zYre㓮L` LJ1Zy.w ;ppRaQcUFqa@e)a-4fJFHq|0f`)Q cd j0DPR5c6rvk($0g2ǹ! LT[pI"༇Q]HQk %ImR!`V0/N5EJGt`ZD"Pq.rvkl;%+Z;wtEk^kvG) V:^4 "b6r1:x-)'ٰ[ƒY1 !f`E;$$ǐ2!c}904p[VF4q T#wday▱Ņ(.2XsrHxRR9IЖ}f:0l'贷"LLryå )3jS2#,RFXpuʪ\+L&"$m|« $>E3hO&Po$W~_ɧq/"Ѿk+=`|sq[xyGue0X׿>Z zr=3x\TYeGfږPDP>7ԫnQWȞiԝ)5<7DM9+ eHj}@XdO eɐʒ!;%#2B&r0IFjwm_1IDQ``lxi:~yq%(CRQ$NX]IJclg{6;n#匇> 6IVeetGᄌDQN8E0W>۫Fù^ާg |n\xKř]q0>p.SRmvT^7`ʱ/^}!{U~@)`e&%k̡R⾧*RÔ|ke^2̸xZQ(p֫ƘM&[$2Xm"7 mB%Y*nw5,l *K0Dd闬#sd6Mޖ %:&%I0 :(ixH!*'P+SI,'a"qt_kClosZ{X!Sʯ.a]1 {y]NwwT6eG[%Vz.f Kf]n@ R2$Ku hBЁ)In5mȪ>΂wrvFh+"G zù C"seF c򁠶 AhTU9s,; i+xr4Hcs3٤R>!s+Y5rV~b2&F!Kۈ@J.ܛ.j.BBb2O@-Nx aD|V/d%`j/8'.%hR̚HYˠBՙZL0t8IK`ɅS5&fF3%g}ΠwJm$Dh;QEqL|_҅Ď { |6ͽd]d9)X92<#f*YQfskYzbOɽ9UEv>,|0mvH,{K0 i52zÂ3aId9!ͨx,NYCv^Y:5ejҴMwPqP֐(r.6fl21Z>]nCZ~/}NS}HKB4 z%ͬADgUc] gֺBwTuFZV!$&Ρ1  2b ڨܖpG&P['iw }u)Ÿx2hW6,;6di٧|,ᅳF/0j~,QI~e!d!. aWO5i'OF FN$D|'Wy?N G ڣPβ4#cV5LVTlR 1`y)bd [ h],Kuf*@YcL{MVa "~GM-O;:Niě q:ܛ UWuk߽ vHe(e4R#-ȃAj\<V}|Q:="5ZRӲ|^;MgtG2MN/ڋ^)No]LJaLĥv[ݝi9I=$9E[ȵx(]*t;TU!U!רCQWZk]]FtGueWⴂW(uɜ֋2j?}6CߋEW_1�Ǫ^1Wq@wWZbe-&n9ܰDZ0|3{׵hjKޚD1 ȀfM.jC94?TY2oT|j'B]?lu-_þgbhYA4 ?{i:mKQK&dZ&ÝL 'żgkŏXtDZO˱lM<#7,Tv1) hb@ƗV'N<.VI.DV0wN6Z,z|']&#m|el,i=DL>ES( ZK 8golg&Z+KZ~8ٝ>JYebsj"eܵdJ?,~WnLFx:mрXd8yA{)m!i|reJ|Sۮh޺7k^]ٓ/ެg(ኘ3?̃ʢZG1׋i K;HcF2bHB=ٺaadkYew$1 G,XvppX՘t䔑kGlIu\\:)i!!e`IѸu_LH{^y?#fXT *gǽ??%vo;yǟޖ߽>Ͽt?ߜ5:]f#AG55u554fhYsf\]skƽ->Z\ͷY[B|a? #rQ(i%W{fT>0WlePIwRT [RbB Yރ ]/#]%VaQaVfsXy;mE.r'^ucsF22cPm:nrhtOrA.ҐP% :b0<N9Nf 5Y<3ײuY=c**(rN,eaaFp Glp$ds1!~UH]I;l.E2D^̵r|ͧ?},&a4#w5}Ȑ< {2ʰg++Sm ?J`JVTJiaQ02.Cd3Zb:ed XQk${%wj Eyv\]V9Iǣvǣ᧻-rmDTd-Ovo^*Y(n|ܥa&WY4ނ$5;fdfR^Zɥ1d@o!=f9;BBV޹-4g:\׵V N5+Zi;߼QhT7q89n_H< !(/Uc4F7faTWd<'ӊ=Q&NxHDHIEJ" *hϴ'D˴O^ ]:uCIC74>[`g)~my}>37/<1;%nv[LGgbTpOr9SZ#TʮwX|Jh+.x_~LvhaXhLHh%D~"4, Crа 4D8 ܱ,,chy@$^InA+.]AuJ.%%37x6yoA<΀9YtHԞh9[F;t󷀃e&^w/W2bWfhxkkN8=ߢ~w45{Ö> fot=߅zt>m\{5=\\r-IԆ5]^?AC.t5'ԺKfzs=.Su7f!le3ݴzj|sE+-7x4x9odmopeq~#=y4wܘԄ%k.hm ĚihyNwÉ՝;;Qlv'εC("ΕBD˵ du 7>kx\WL!FI 26.[C{;쾧IiAct2{0(wF'*3Ȝ3A+ g5؝ 1*%J2%Bf21Q[\eVq'xrT`ƹle^-ػМG_.slڹOs"nOlCͶfZf|bz:kb?e{jxe!8㼳Q!. ńgA$+Khp\U%m ѡP:p 3B!M29HBs%9Kv7 u)g+_r܄3(Wte<񩯋3@qv*>i㏟k$xhRVM,k4v.%]jVʺ7@}ctUi“,w` od7ؿק黝kYbu[JNuC_> 1S( Z u`e1hnЋRzK '4̅K%*MiKXP$%oXf]LIxO1kL ~]o9W '|b>l n>|cDc$ىgqpdDzdZN&fWd=6h9ݘMh? Cf|cUhբHQjEg=ʛMrzK9"Kr(e)Q:*D1QHXJiؖGT3Zt3(jNI3 CWV) b%Z36#gf|D6]um ͠ ;N[gMn'ita42:L߸F&Z2+CVu&H$┉A8HwdۤbT:FUUںd!ZN=R*$XN*nEf//s݌;ǓZY-(+*KG2 cN2&SQafmK~ )2#C E']\D(Lc]NDLBflև3~YU*ƾhfFt54[(+Rb@KOI(* [t`J2TI`ŪͶ^$mKQKaQlY.SDC فɔ0 'lX1C(Yl*PW"&XEWtXG~fh6xKNtW:ȣP,8$]g- v3@Xmi"itFBZUuպ%{RB~׿*oף<򱜌.5d㪭6!''I>aO?6}ŚJcmrn0^O|bxeguwg}!o Xz*YeCŴKPPPϋ=dϔn7o9;QO77W&E^w%a?BN8L +KI,ޒJ|d m`=%[MypcX<蚺rik9L\[2.0rTCkJB} x@xmI\%?w:/V{cgeh;,;bc$}7- t75Jxj4΅LKdIZCtl:k̔'O)F_CtXiۚE,:~,)`,X֥E/Wʩ^J DpNK?uX3rc#8ՋV }9)O%O}rpCЉir#*eՉ22ِvR u ,ڔhYSZY/$3rLZT E؜@RCX^7fl)g3 ٌQ/ϞbT>(I\0^E6;Sf'ZDvJIQwM2@2K0AZ[BO5!]+奒1EPY4KIS( 9lVi;@fP+MG4Iw:;%KM(Mpg脉FE]$6{϶(/H Qh8&ez}՗fOwwJ1i5nItZA!SrRn~SoOk^ǐWd A#JČ 䋐A2) :'HnX[T`4dɂ[vo mrhܻCցSR!nWa.x^ʸх"#͙B Db_Ҵ|nœy5\i} V7tt̀FtuFp>>Ny}%N?t9 Jњ}zySW>05T,_o_Dwd.[m,s=./#W~"j~]JR5LY.fYk=r@9:ilET¦+[{LBCc.5$MO >mL7,ݷz3d~3͊,P˳'(|z|1+?ɼ5o~bwcU9| sԫ Qn&>uǗ#Joɸ-Jg]j 6iR ؟4͝h!4STM"6aȂʹjL\ҲdN%Kn| Ee k#x/02tF:Ϩ=ćW^ޭ9[]O*ۧdV1BdAZEb1;HOR0$-P  6&Ɵ HDT'FMԌN ,"C8Vl{2HJ$a=AuJm2 C׬;^y D JH ,M~ nB!˒c"1It%yćD,YBG!W҅F;Ռ1Eꗌ\XhQ%i|FjO*hIK= 2#T[:mY m ~]AŁv/?p`TgSgٔZɺ` u> vlL .Z08-Z1c8īp0xPf!r$Bq^%';k4GlPQ ->lQ:z鉐 8ӯu1voXm_5#б80jͮٟF-@ΰ:"wb3K4rg=O'Ǜ҃_FGƔPiHV,ÿ^^^^o?Ep:,uysѭ,?gL;84I_+LySu+~SK>_>~8-leJt? ALҵ솱OOV3> (>g%yRZ&O11gc :!{,) h=V=;}<_O xɡBF*_ܣc I<=53=o"gzׅ^TqovRҨ]U|2q _~wE5'F1ǓIoٯ*$‘e|5&И@90ĘK PVPUY$.vdׅN<^sD;Uߝv՝[<>-̓2㞎|=w˕kgѡǒv.~z䦗r#)gpD.T^VEI+vKI.J%'(ӕT ~׬)^wW0K(~yVof#Mn'^f\t5s|d d9g=Hzw`╅dX&f|0amW SM&eR[@XD6-t\ŠǛL.2\܃%A=& 'bBRUIɴ 7]NLS .-Iw?6Ɇ{7]2f!b=TC22!s-Fmf[ncv»c '[T8 OK"FYd$$oIFk!LzDVSfl'3}jbkwۋ~&Xy[竧aa:UlA̮t(h' FGRɕp,;)Z}IT;)?NWƨTfʴjmWG ΓֱmCFI~S'B|G<քx. ݎrғ~eqCYwbAUp4^Q=GvKG..ty{'3P.yCBO3ޗv3+ięFٱV@jb2\6-g4 ^k3xfR]E "FtK [&1cmԞ44p[@1Vx_$q DD,PJ!u9,Ew~_|S&6L-ijJp1}d1! 7}-8]HcQfrکOӈҜ )S6CgҀ8;;(a(U^S;Lߗ͛1WQˋx +9U!$Jj$$$?Ek&ȁ^6!LR5ĖJIV6j\Z?k9lu:#m#M~q/hU Iuʶ[;?ޚ پ+X'y<}ۖk}F])omI|;(yp>q7 9e 9 qA0i6!^pAbrY( a!_T'[vsPe## d2rJZ bLڠj-R l9Gl~m{aޡt Z9fP 5\#8n,.uZbͮ$[ orB$j%`oc,(,ҢRc(6CVaXLa^.~'݃w-ZV_9ϖ(_<BPSR&L DF"8flO56 |>FMC^+ ^읩IoT3kZ1Gc A&DY=-hQھܴ Kgnw %p/Fѩ⎗U|E^K/(|⑑k>C.&ՙL*"ow}3q%7V4R%> ?Xw2dg3t1@3;|_Wj#?현ء+o"}w\`ϗ\}O;>j|pd> 0Gpf`-xeP| -TL;XX5BtUzgfE9L.DҶ#d\f V}"$In7{r8M շmnnE4͠F}<~دx-ʊZ-'ɀ;;< eEP Ec *p=s͉ BujT8C-RaѶ*%bhPca1)אc˜K5R[#2޲bTyZDv1:(_~Yן//M.0qkԣfxU[5'Kͷ*w`IX8Ll ~B93xG(#6fLE9,ؕL o>MMRcpSnȣ92gdAʌmC ah Uv6 i9Y38ԋ?P}aڊ]MԖc5+M [Vs їLiY4ՖژK4/*_5cLoWLQ&ǔ'u`ʅ}p(BZm߆y ss5_DE-kF0ysؘDK*zJT5 }fLY 6I ң3&"bR@ *8eq0r[*$G 'eu5fW}px;*36?=7?Sl|,&Lଽ(M08c/k8b|^ϧ|zQ!^k^/yz RpX}:'"ɑġzh%/ØkI7;ok II9&F\20:[/CPDm)9` rk B6VX*RuYSkbQRrœX s9˝lۛo ;3B׻#6~׷=w;7y>"r/iw=:uu~r?io%CXKlw[ WOwWnTz>viGow~r??|gk7|ϫ}1w6Y'n[9g.=xwEɱÉ\R\‰Pʲ=s[B %Z0/^+'78Go$\uQ[њYjSn#/w'f/RR5;uUT^$z A[hPdLγ;&ؽVGwpZOݡxVӦ1-[[?\,b9S.9H{UTb0aaϨV6x[͖TZR:BKֆ ]|PME2mT&fEa?ig8JYkyOwnb,Y&O>$%8z3I*\ҩunI*`$($\#_OJj 9brj+d#qSf'v}@Z_K͋+޴>EOM_^Yr7~m}ܰo,e6;n,;ohwʹ"bΡ4 HdCF" G ,N8+M"ԆB՗sCcr͙PlPPHPV[\Q(} !FanTbsh}:Ot6~ݞ6ʇO+yaלbør1 E.Zd\xb`8X}cld#C? iS@f砑l*ik5N&gÔZ6;n2m}':{8peVS?vU8{]Ց~~`l–LNjMmN@#MEsr-;uTיJrp Î* BDZ G61pn}֤[Po׈D!jR<$*C b("At%u %9Q{mY'vqt]bs(䊍Z,Ź\皛ar4jYN7t%}(r}2k! \gZ{H_egPuꎔdvfMFe`،퐐chcLmhlNUsu diG ̺`U YJ=+ m:es,\op ;a/Am坾'r:58Kz?.s( gQޅ;yu}eIPCMoi:KP_'NITT;PB3rp$c;S78LQ ā-גDr'o>IBCf)|:L1F,  }ޢ|L>ru~CkWw9^lJ櫄+zj~ZNOWߪ8'm.wibRArb'9"cJ|ޕͯ.یޮ_|x ήĜoaOf+ʮl! .v9!M#i87t4 4#ifUޣ8: W0bbГ1I-Gy$Fm-GS<@hLP{^?7C\blDsvpqN~~ןo>~zOxo W`=&7 YkCVPjho14|EY|qU)7{]|oҲ_wɬc"^>ɇN6~]Y]C/,dJ?'EyM?ר̺TPb7д/s>Kc>F*y6y:f`ԣ$?Y.N` `S;@@@'F5^q~l''Fg@/ԡȼ .WOApgfntL\MK$/Sػ# \բ-^2AAZ. ',0Aw ]($H4ޢ TfB2!TCI~9kKf11zћTcף׃7^M:`݀xRPR<Ƈrx_DUE*M81vUn0ۺԊ6)TlUq3+_:HM`xAP"..PcƉ w+ 3R&Qq&(p@ 6Z< YN/9$ԉQt55伡UZ:nYo{t>mR95qbX5P5Y43^URNBk2g(c$%+W.P59z{;P HV/eS2ZQllDB=nKE}&Q^3!8]uXݞipqEǮ0!?9+W\!q$aOh:4j=ߡ8#BptTPy坧 #쑖7Zɥ2ԂhT2i.H]euԂlTvE+IWQu aF]erg2 +SW/H] BÛkeeF/?]!Au8o^ohUo+ket ބ( R1}]`-p4r0UzMwf5"ѶEgv꘣Wf<-LJӫwvj{/=xIr LFloWq<:q8"ϭHy'r™*uuO o.鲡6t9K՜b2>ՀNi(6=%RX32k9~+oaa?L4!Ϲ^}| unqݒU3 ëOW| }Ȓ U5̈$J@LjHm|۳-!qLB+ c)t.@L2)ph([HA9V8<>H+$I+bDn(UNpN1/*9#ʬCT)nBpQJgo]m*QSrBd"r¯xА@EK\+O$Vj}?i[%Kp@PǢ\j齔AˢrfILiEFIs Jpl@sΉN:FtkXa9+FΖr־zY~5z60*~ܖ s1) NA!'8tbsrgh\$kw:?+-?KZ'D ,0 RT=K 4I /-EjZ~:OZ+X2GA&*}6NSr)QA)f"1o)"]e*K3)S V6f(Js .u DXO(%DPf! .Rg-ӚRO |$R1Qq3&H SN˂+v lk0.%m^8{vH멐JQ$䊽 U/8]mXlG$1\4;MATvj<+> JXzU:䎄.^]౺cwGh I|X)uT')&f1KrU.8 A1as H©Iē\x I3Zk9@IEX*89ok618&șgd}.Bo&'\G_7ߓ{&! X{k7S{"O:šOpG| JDEh@$Zteŗ> %\.MX߃irݞ4͝hyi>d $'hHR !DK9Ɛb]h, T"LVY9OZWhD墶\p-bC:%}(JPD IIFdRD*鍒{(!ŧ>19[]?j||pv1] UG?! _Oj;tDH2$?9W٠H2A ԢJ⌄(+5 |2@%RQ&H-e ݚ-g]zH@:سSA'T:m(Fje__ּ[?4=_K4@ߜ9+bDEO+aI>uu-Z%Zpeqܨ8+ZlO|8̟lQytoк;^6Y)O&5]ab=h :'o3ʸz6pH jX4JNFYG*Q/BG)IlR4^U:޵\*NjKX"co IQHQ"Q@~rs]lGN;`X-מ/i_OR(ɜ#3:zbn*^J@VY%h6:1grwп+@<֟#Ι01 r S\z*:kJ@P*`p&OpMN DyRa$Id-L1ONE)2Z]rR \B;BZ[z:kɘ|R%D $ cLdȞc ]Cd>6)Z{s1\8 s߃\oJn^{a[sẻ7PdČ#$,}`wb Gefwspsev}dYmo{26. !Dq` !`\56k5CrrI0Аȩ3Ցn9⦏ x< $w\">u~{P={{i+Q>[ Lwtlr~q-V]%3a\&*p*S+6?/i8!GWT7Ҽ]^s{ftݯ͕W/'^,:g710 ONwղ Eh.'ux~0o3<-+ʥ%Iu{IˊKъ--fQUސ& k,XF㬡WeV--Uե*ԥuuüqY\?X}9&ʽ2#ްY}bʭx&yX\)~?zY~_{u7\q4Z`D}Z&V[vl7狟/ڈfEڬ)״Y%4#[~tCݟx֫?ej$V/NWAWHl2&r܋GR3,'KS !<܀e/j}Vm9JM,5}LUÜ{ީZp8sF@ v-}e?[KH,ETbւƉoN֏$IA#DBW!2'%7%a(9t3DK H:Ց/=;26O8PgT"A2( fDB5EOf@ZYک_ĆkӤw,E}u*=Os[=~+˙_g~Z}3*+ Ce$ |9^!9].ґ !0cp8ӊa@3׈@2p/H2hgX]򄉣n8N'`f|u4nϿxI{m:c͞E&(|v[rt쉱J{o+;޾WPKN,=i0;5/Z[ێ<1Cs~3 r\q- 1BpJ"tRVH+?~s)Ox>}J`.^~!_p;q~IJ;;gpBvQKd^HXXY=]gN3G =:nʧ}PWH#\@_е,tzuJfOfnE׮&HPcuzj*GH .ɔ̍pk32EHۥʗqL,+JJY ZJ?c0v_  E_dn7$n.kSϧ5?uϧN?];S _y଒beJǔgLw{w{7ʁ[#rIfD򒨪)sf631̀'$zC?8N9, +s4Zi蹜N}+ydGo}.+2%QQBDE,iLV-8:B`IsAzN mLٌ#*sdT92߲&|qyّ9bsJ.|$fiSW}}j?J1z!ٲTqE&m+I#V>*݄#^B毴L#muV^'k4REAPRDfY6,Adl9BD{,[Z(qi嚌UVJEe !X+*c-ĹDF 5=ٕtp; ;wOh@;)1p֛%4ҪLcyVQ[S!0<ٞMLmʤ>&KxIs s%}&j[nd\< bޑԘ׶N.Lh5wB?tonY<+ %bRRQHFR#PsQHbo<*5b> !6Ae2J]̺,CtJnAZ ڭiǮP`W6 -ؘS^%K]ȊR^!(R6-IeU\C&ϐKb$dQ 6jra>D[n(4[iv4y<.Cd!TY$T9d^ҫ:Z:N8罬2r8B.KvC檠NWeWEN{DH`Zܽ!rEڇ^4ٞ)uD 9V}&/M^m L \i;)&p%UWͪ^<\:zpE\PBo8^Ǘh"28=Ɍ@f]rUȈT FT+cf9ڢ35:)Ԝ'0 gM<)0)e%B$6hr,A`%y&tnmB%ŵ&(L=R64 ts5לʼDX^G F)_$_Gڏ)&]xWieeW)bEri%l=^2(V 3s %9F,W.b91T*4Ρ%j)U,i "ChEڽztuNiěq2>wwWi'/<~P (D@ ?woy&apM-2\ =fqnm"JhkOƬ jRV@zkK:7qk+&ޱO}-Hv@muh}b9zR_\$9a0@Ѕ窐.2tK"#XYƁqd8C/S"YEKʐ9!]\,Vz6 ۧE@ ^W wfD;$,2zm}\rmp3%`][oY$Y\.x8NR-+dI1̢v<(A'A̠p XZsKf}!)6i  BkfͱV1+cC} dKk^bLy+AщLJNUMI/o\Hh9d0#@_8Ga()둌s}"-ku ܁d? Bl(hix5ΌWYJl N5![-*b8}ȡ({CQԇE <wxAA M &իh6::yQ{8cWQUt5Ӗ3LjO|*"[eih}t8gBG$gw`O$%/i:ix&2;oo$-D2ya[ R)SkK)Rue@Jb(ˀ-}<DKqwA=n!ѢXK)U6eF &eKr >K`RmrTȬd-3W7RH i-F)T(3!e)4rRMLGfG*T#g8yFէkd|Lnx`yqln{x٧1mÌ㎂-Φ/n5]&UBt2pù$"3*'mm" DJ-FxPx([Q 1XJVOFNVG'\*p&2\hm2(PU1p#tLI,I+wV^ܑ" \ɜ"Xy5r y?R%&< dl峏{kMnf5IM?wg)!qʋWy"/-d&0RLiaDR*L&6hJg6??IT=y}9 똕"f債Q9IQ#: x[&Z=Ĵ橢y ^qR3rKq?xRO yDT),q3D%F{<ΨP)'W^޿зzZtp(`hQE-qM 5,"+Jw)r5@C%=f¤% Ry\H 0R 6Zj!ioW8L-׋nxuQ|V;j.44 wj܌.Bjl_;5.vytg畨ݍ V 86\ j/vtFm~»걄-lbmL-j{N]0G 5׀/ŗ[o8x}Y6|c*JJYZДkc} HpH}k3;*8qѷM6+y*P h=S$Ƃɠ yI ɠgDL`>%+5S+cӫJxr5ʅ갭 {zmyYT޾3\J}KƪG嵖"=ZV*5OIޤm#9wJ7d$e,B5xH)3Z7Q 'iaJw1E()3wu҄Osx/|4Yw:˵mK_.7BJAȌ: /%dVR9K& mtb쪨;=BۑE0: c` Ӧ\zI-wH"'AT#nWceN-{ HLϓ@ lB2E4 ON3R`<Y(ѷb>;zҨ0J%*G dGL 9>{3R]A: Z^s vnɻi{fgϥ-3$--ˆd_|ٗ6'Q<ϤKFĕ,4*8YN; ISKng2wL0'lQ#n()&8;h4x|&_O5ֿfF.Гd,:CMs9uInJ 2V)?>B><)MWw)VVV)ZVFg%W(>huwMLkr͵.h8-$?\Aoy?6UѲyq3ҿw~jd»Mpv_sOq2x\YpW+("FMO/vKƺFs$k9VsMV0QeN>Nzq5t|erڪQ>z]vi.C^KHXu6i/1?l+*SM}M=?Usi*E3ZCp=Y҄ģ [ExW]ͦu@ult޴2]8V|;7cX&D}ذR~;VbZnld<ƇVyZA終ϵd=- @k#FHX LI+8- &0#S1r0@rڔCBfPr*awZ29C0-6 6FFnHRΖO=5r[/'B//vg:j|&[F}Uzx1]|/-p,6%ZF17ZR^K xvDg(cDɃ͠V¡Qr'Qh|nY+@0$q)3Rr9`whЕe5r Z,m`G2OݩWp2¹,ڦY@jj޿y*K3 "]k|Bӻ}#"#=:?ii,[WN_ռunvmyo|ոFu-|6lW|Ϧ_\U)Gwt|>]^L/O7GRyqڿ)YϢs7EgP_yw2[ (3鶍hݪ_ŐC+S^d='fsX˖jGe{f%y= m-߯AƬrhHQ 4z:ZʤB E+2oF6#|PHt,5ChneKT(rY `K6'`cX3rrZnFwߏB9f|2fq kn{G V~}E]){,ARTL>A z"#M}cG8jĝ߂*5q_ɨE2^OcXM1ǫs}c}G̺1xf[Pq*_O{#{& KJ[5NY5؅ e hv $<8Vx7\?Z/~^(dCx) (`)RtPUD ҥ\)lPmUj\e$LYZ (G2)ms )hlfE]J4"qoZ/6ā!7.6[sݻAXBW0qk`¥&Z8*nG]sT`bF\x5xGfG]krtM5xVkrtM59&G]krtMax>WݪSfs2΄:,;S1QnDŽoZ5o|kJ  DiL(Ȃ 6Kgޓ2e(PJ)J?J\Mm(Ox,]8J֥KAW)}&1ZěG8֡ZVd^_Ψlz_ =Îה:Ck+9nk߷Kh:Xw헜SJ'*P ^Ht.)- |@k̤| WBJ= Gm&6M$x7ド*G:R69 )-i/IƁCt.k*gQP^i傐E2iI9/|m˹@PR"XΚr+Qgw YGJR6eb`⤌ R0 ~(IAfi,Ũi"pZQΪ LR6oP(@M^,-(٬nZ͠.4I;%bIVS$mlv>6Y"HVɠCQ,DVXPʩ4˩hѶ|O#}iet%e9TS6$8"Ǡ,$ @d>C 0m¿:~ЃތE>XZ`-i(E_ۮaـ @`I XՆE'\Ũx\Auyi!Jj,{t^.ŠqH!ݦV?{ؓHԈM@lL&ho9O.\G,c:-kuhDNu!f!RFeå_U]F>'5WdQH¤ 䋐Q2)()d 6Ypmධ+HkNKINK{`3!-5!-2ŲrآS2~lm4JhX"u)&Nw:VًHX$/EAR}-*6GIF5/QF)BHHeT9ƒ)%\"6ʑThTUEBlUB$Bc$܌ǑpE~/OCBv,mi¿q߲^|:{:G^ĪBЗΗGOMɖhIKQ2 osS4(Ԙš#)QSFpPyyxYi)S :RV0)iRI (@BpKbސvD5#PAVS+gbcE5%Vi֌=j1mߑ c9Z0Ǐ{-NԝʃG>Mn}7u~~ίM+<(c#5Nֱ( :࣐իTURB&Q4c \)*/;. ji oJD 5Cz6K)ٽN jhGr1%勳9e򵶏C^C&X>@!#Nl|6#^{Fν/(3JvQX"M Ufw{RMT1RL-vڵ_Z'UCsr򷂳yO\Ub׉CNVw'yqK8M-_F~bZ[io9~ӨunrlFW7erSfU&v_s}U" ~.S&}Iz cW'jMﻩ) TNޯ6Gdu=~֛ lfdY^+/0/rMބ/w{ㄹqev9&b>b0)P{dn uт#qn\{tHZ#䆚3&CĄt|rHj֢)T r&9>yA H5K4'C^=v7/T'AmLui.v,ӊT 803ӥ9h'gR4(5.9bJhnܺz{NJqAk/-KΪRbƚ-$F"u)5u/2+]mY{>ߦ@Y/u喘otv-zEYtz0gUI|.=R)ּ֡.) HD70B+# -R- ސxd *! y%sV) T%ԮAqU Jɰ1#,r #F^ٱDIJ(Has2%n. lFvs 1zHREB2ƊFe:sCrlz6#iPC>M‰b$ΒHIPlH*4/"+Mg@tήSL' ˧9&jVV1PDN^;Ng-9ӡ&!%꬧X`)i H1VR]kĹAwgX3]ȸ.4[B›S/ɨli&&ikr} 4hp25v2)k|VaS\0ixb˔1H2I!E)j D Y QjMl4E'6K"e;k٬hϙۢ;ۢM{fPqlŦcbEb)c cxgHقIWԢy-9eCNKd De]+"a:/"QM6t'ܙ8ak/讫b m5X#^#qmW41p CxJLZdM9E6ƊnwRÒHCAblQg"GdH RbK+QLꚊ3q6kďį:^,lQr[;֋׋^\kS,:툭#!LDYTNA Sʕ lC/n}xKb{Pa2kFoNl1Cpu~xyjis_; 9 MN5(ǂצd HDճw&߽Ƙ_];O>dtEԽ !=3ul(J[$*+1II)jBRB'н=n̂\1׸ AQ 1K V ہbD(ł5d᭿3BL 5PZl$,ΣNKO2 `;ysƭѐfYuX=dϑ%22##bNŊC:@"*d٣RT2L.Qgm|9uKdMjBܓ́J B흈wHH{a2xio2e5xGG[ r%#I!(T]׋Kq'ô?gݯ㈆*w~LyCꞿOЪy/qϾ\vŋ Ĵ JUtB\>ؘ8⠞Avܠ};D҃Yf_[wݯ;ߞN^gW0a9:|w0Yg={Qm{uˮ=KbUKWW5V6cXaQ܃J>}Zt隇3NlU.rU+ V'u36yc4N k2rOFM+ZVJ *;nǿ_}ͷ~f_*߼WqL)UU"}g~w+M׫n,CMs{M]>oҮW{uX<{֠=L˕@o?0w8%Y쵶?uldy_ayrNi^E+?.I$T8*1 Xb6Z.ĩĩ1[SͨVey\ї6ƾ*g4Y$9zJg8L~NΒO)Oۨ $0XЩ%kV)qɹXHD{# ܴS6LQYgEQ@YaP2#)4`MN6&4==ٖg[<3mTzg̠w缕2J{K s~#t{Lҙl+6jv1%@-J lY&''o'w޾\iP+heC徼b/dl#7Zv2'R=GՆE|2adLԦqJɅIRMIފ'l\Lupr1VPgG=X$rZ3l4iQV:Ŝxa#EBvAС6QEX2,Q'6,e9&g)k5FL5tsAC!٘Κ蓦{hdd.]D1zg:Լ >zHu%Ĭ_9~:TqU8WJqo>Ta>W],"yg:LޜCƟ$0({dy784:/Eznz.[ϩ/e|VXt%4ނoЉԸ0|LzkLg&ęgd^Q ^%@F+_LL(=e5gHDR_9_#|Xrjϗ UWW+ڙkE{j5'm뫮w=RMҕ>K7\}Z^_=wpe q}orɆK.To^RY͕ύ|P oqwygɘQ s^q/!饧^_iĊij.yHwÉ;QH[˭w%>+Qj+6wƕXbW\UZ*]OѕjwCF;ꪊ*-nR^]=Eu+'vH]bgU>SWU wWWUJ{uՕ 9 n-hvp2͜95|.p / C?{`<ΊfM;f y8tڳo(3#5[؝4UQUW,qjim]RWU`'wF]UqXZJ)WWOP]Ν;'TMLJ i8l;ޝLA }ʼnزj4LYnUg _>r,qaxn#snO.[YS~+_E+JMK3+:ҳq;Oq}wqq}w𥕕eF}`dF}`dF}`dhؽmZ#{~\']hV@hHѠtԸs([Jo{ooBD*%f,!MUr#\&F6譌 U\OQʢ.HBl=2E*oT@EII…PvI!_v 8d3 &A.$[/mXx$nEvUyb5q69>_G<%ÁE|NiMӮbɁ` % H# P<irL Et$AϹ>R^ ilL[+.1@2 ,Tjmkq3BsYk߬}mq=Qn^gO,e=oH3Bmq6:i?|qeL!5\Ө)UFetRcO(8)BY$;B(+JNRecWIE^9$&d(xeTLY R׏/ow^"-?/+ BRzYGt@VH‘ufBE̎KADR(%'b M,EZ*aU$#!tNq59 =PPsq]֌ zFnZg)?eM"p&A\WPlGuz6yOR%d|}j.lW;JYLXqNh N lΊ {r:eUh)Ӂlϥ,MT)H`1nd\IĹ3*|a5͸/4}Ww=ajbU? Ӡ?>s4A#YOB) e 2BT:"Q1npsZ,d/$dB6`T4*YY(D3`fPjop%;ib׮kmmhX)WŒF>fM0>IVMƓ- :1dB\|MBFp`87BLY'DцkĹ~~w#VӏzD[#ڽG{4^@"c1p!=!$OV$Z#2#KXV!z3$ 7>eJg"٣Shp+&<"P 8w{O'H:_,R/~/ʍɂg~;$v>OMYDIH浌1͍Q^ۘŇ;ÜkiǮC?.obYm- nDُ{V4ͬIok ^ƣM<(@o){{ʔ5z9A,G?=}+/{}]N>48L- ߻(5(m9G͎GwVGZ5r ƽDς'r9`BzDc6Q[R\ 3Aȅ`I*KdtT .(FPhoi/k:<'Rm_M 5zC^V'I<CWQ{oRfz0:>x,W1M(7YRYK`@:YμJ4FgFTLs>[i5y{W5V$ '%C]Y2,Ma k̠VqTv^irn"0!Xkp_kBg%Kt< vXSF"2RzDICR eه5 8 ]ثb鱦BIg3SH:2EP(L(E9p( 1 0ؽ!5{Xk=+V!VY{(.ϿZ\O(nl4?4 ++q9io=3|Iy>9LΨyY0޿g^ ,ѧLZ~.ڮO2wNؤo#:PSLp%g)0i\LOp0$(!$HS'Pg׋VtA҇٧GEps¹yB !~jcuwZW2%W/go5JrF8Mn߾cx=;R?d׌.T7ʾ]f_pݯ+o ϧ'7ެX0sJ9+"ROőߜrh8|jivږ^iW3Yg30p(2e& \?OoqUnouɮV; V'e R֤>|W!?l+_-_jH'9pvR?wG?GGxwą;z7G޾YO4SD0zk~NmF͛Vޢi&M뚽&ښ]hXuߖa9$O_ ŋZkS.Lb^Z|&3?] h5'[Ju!`@W\5Rr[9U,']. J|O?”cQ=i82?ZsVQ~vAW2Ĝ ֏Wrc:2P9Jlr"FNo mw L$AVIcGGiM.w,N&+"f!*&B)PaB2  SE*PJY5gdu3%NWߞxu%k [=M|ZVkw/WY+<6(M"43a[aqeozT4UA{VMIsf4XpU^*! / Ⴄ"KdUa7,fx{b*d$F2#V@Pj 'G}|8fY[ݰsv97n>}t{ @ Dp6jv02z.et*ǃo㝣`S8 @uYTg֢vox/.iO;*y4X0cvo|EՋvr˫?;oy[?h|V=?H?RVq;qy﷐2xpKNSV2$IVZE򅭝qT|=zݜO zLofеu:*wn rX[,vѵ%LϳOjCٗ_ݸh`W{}}x89ߗ@?fS,!צJѽח.<6.F㥨HKR'NjKcy|ctT*Ҿ9NqΨJʍ曨oelQ[n(}VNź.j]nݑ2QV{ K 8fLJϑmR ɦs& er3n$$`eNkSn,pi-,]e:vR hMO'8M龂?[OخzW?M>G?{6~> ;R:5Q(]zCwkl`͕<1CjAb͕u*+,*`7S;|V3>3u‹'N%/g? `.DUMH7.iV'stlw`';e%dt#`x8<:%kj$EYJ`%00_.Ieds:Uzc!`GVgZ,R\Q[.ʔYŜ`HT ƹLlͮٱĻP\M_EI0wB>x9yNW}x Ov_!i[{}#ڡf29aXh=GeLRg@0"rxF0IC r= wٛL%&4>,KO9 RdcZZe+ϻT;،̖<)&>퐷0mn-MUt Ep>}ꔶxųD]t ikҁC8JJ*(5H: xƓ٧AVB%FiXQ 3h39W !됧Au݅!^y3#Gǀ3#)egn':H* HZBI}mRͱ a]U-G*CORTe7O q &}GmNÝb3qcYIkT1eM`=$z@=bNnB2C-NWwz7/wi&6h W?M 8ջZ Z0ՔF'޹t=]0dpSfjxbh-Tz!c˗AVoؿM̴|*Cwȇ G{N:Dԡue5r4./ġpX@6 uqc$V':y @5."1\,j)#[5ʃdbz7 ͒1e FJZ A/冸>$ JKro5RN$cp}6m 7-quٽZrZ{gRҚ݋\Ӗ ^ K;BEw/S_ ?l݋`CνٺK罟~Tr,{Щܷ(Z5pZCWPBx \.Ջ+f8m+$;k]pZjeM+^ \qQSp[gR[HX걽+t8`1"Ho7M#&zFCk˜1T-?}pe\4,i L#5Z%R|#䶃ӂIi=׳eRN{oFяo&7jWy|=a,%=ƣTcEA As[SN(t}(uUoypuC^m U LB~|vz0]Bs1/Ѳp *w{]{6XyfRp}!Uu1}r L} *"wbSx.+/)B -.guyRֱl-,{ݳIhȅ^z +y+\ȢAmEY4/ТQTAk+E5k -pԊ;`JՋ+9e+)߳n)bratyQS+EVA? ?\aڂHj<~"s`D4)Ѧ6$ض' դ5AH#+roˁ+$cmBe{S$W:Ejm:\!Wbǩ 3炫A{\~Tȋ=JtpuSN_}pE9#gOޏ\&WHM+R^ \1. -+$Xkt[ D5J;zpŹT`%[WHVm+ֲPiHg Hg o DtՂ[()p^Kb[H.mVJtF*`´T+͇+ ؊\RWH\@Rh& 6\ ;H%'\H2|)h8igenMȖp$}z8N*~|XJpEsJIX6N2r#PteF(__)ƋeN-҃RI3+\VҚH#3XdL. 1=$z2}U-=bJ: s.$/w~`t\q8fq{uCؿxf.=Gȋ'O ENu _7`5l)籫ՍFvu#]Ȯ^W7IJW7iՍFvu#uu#);PЕݚ܁IY@gB!G&3\7k'Θ8Yd\ᕐȤ9e۽2Mo&u;8hb߹J9qyȇݳM#~ n䎁22'X15QZ`XW Q7@u4s`6*b"AlTS 9o|^yZZO֡>mVeP)aNonò4ܤow@ C&ii~^.?X5ۢYAtZL\πU=zv=8vљ#)1ukGTK8ZIڥB$$teT\MoԻ>GUÛy)x66"~ kyXёݧ06{|B)-+ESoa毟R,XO?n?Qopyz^oiu/ȶoϣhy.UWlNXnV [5ց,k:X[_Ix;MIZsf3Ԓj>QORK:RکTgR4gMDŽ!J&(խAj-izP|Ie5%VXKo;M=Jq}0?&|[5 !䄵̸$81*s&Yʨqq`[o$&"eNe傩LU PDbּ3R9;ƜL ,_EA#tm /c8E)lPσ>h \}cs'u9Y2fSMTG2zKʘ<((xXi'A劍  eo2,dw``Tz/=%fLlL^KZj 1 24b-ha6@prc wdJb,;A]`5CmD`33o қ)hӢo+Z<ݢ#lh 3@E^$E@'@n)H?3R㈈g<}ly-DY Xb IZoL:OyP86Sz"ʺyjAT]"_#t?3r^~q x:S<R|~y8$eYJ+ &LkЗºZl]:T"j/nn&AdMR \5z QLs=s?=B]aob`.;^}I;#_7<Fy<&~ 6Ő~vH]y Ћֆw(Onj t9`3KCyV52g'Õ"tw.ߙ=U[}=_{L &ٻ6r$W|]`K\7Lfn/X[%s~,nIi[t8r7EV!U٫v!OCM:Q9R?־V8MbFA1!qW|A7ZW[u&poѿ*|ׇT;iʥ)Ri6^a ur/{}v8pǴNr`ufl?".R[gD˔KLݳ-[U-Z?v,6s=B.'-|Y0V~8ZD蠶n^M o[c/i]EjKC ^6XWźSvaf!J~e- =>4 L:PALǤ!{Qˋς[J.46rD鑹Ϙ}R}o`h8{yd)6Q0#[xq<0,Dd]F !0_[}-kd׍X7\\w, Vq0Xg N1P-KХMzLAODn!I$ ,qbT'Ws'4梧],+0aw.?t'/˼O=KUpd WJ&D16iVXmڠnYHc`ZZ6LqQ N$c 8"@eZ(ؤW@s9 cֽ?z PidEBĐ4IE+Ct@` 6Zmz [v!tR6`rQԆK,Q< Ixr w>2~ H,{kLʽAWn:/ϿGZ<T<cE>;  w`hKM*w}.o*tϓ.;S׽*mq"6uj)(KrN+)'n: ~u28s[E\ )Xo8Yu!_qmYCᴍF&ߵz읟.B>kuOզ%Vs<{tvq&|r4O!lpQm.lq8)&PɃ֞-=heF^b>;k/\g7\ka.ܧ(ߝ5=uϮ,~/hٵ0z.jɈΖfXg3JfVU qm]ǣEGO^ g[N*#[]EyE)=q7qÊJRBgSp_~/{}~ߜRfN_W/p B%w׻ p>y5+4hOӲSO#w{shq5Z ~z1 M~hS/Ld A\|eU7!eAhaC*CQRX0Po;s 1rwc2) (o4>6?k!~6N"<eЖs~g?Ǘ@;oeW>%xn]qc|RٹPy-{RЫ# \C1Ģ9K')矙SHhP0N:6&oBل:*b:qw'eJ(sVב1`o(Vjͽ%[{)OoZ}|×JRp2DJ 8zHbY}zd"j,J*Gxɉ5*X/<"x)^FPEι-s$ K,pA.&僠V$tAl4P21C7D/(18-(K$F)RYQ{ b>Xfe &y jw&t: Q M`:*N? ivG,Y]g\.xtCȭ5 tե@hwxfzb࿐2Gv4/7oEٷn5_.yQJ0Է;m{x]G!MP[|}xE}}#omҍųϗw'&USR j2AA3uH=w{ݏ;)-<:<5HγRUb2%EL%dpH5E;ȧ2 H%"3gV1'8ÁHT ƹDla^L=ĻP]NƟ!ިbzhz泳^9#3:zh gw6j#9G4 = 2pDYBSǸ"Y6ah-A9XJ:2K}'$pTkd^AqGp1qv[qW03}O]g\gۘ휢ݼg,󮯋 qp4K=J5aZ+] U'*NP^_||By,@()J2V /dW4H1Pʽ BErH|},z`t,\}6R(9gu 6t $xa`d(gH$P/;>YE*-a ϰd8~4XAInK4Pxg@mWu.ɦG1('ǒNhR9d|k=nlU޽&uA jkg$3N$( 8 -L8q{^2*S{bñgb&&cyAaR)\B#U2ge,Ubq-TGZ|_zwnn4vo T||8}I9@YdBȤ$!LA3@R1Fp3[ ɞpglr^i(-'&&, IN ~8- <]L;jjvk \9S+,6WA"O'0>5>ȌZ$*ea[o$ȐR,:phk"0G1HjЮp1qvÚ/"dx4)KǡQE:Z!S ""``ײhḾA2G=JiR!ZC Q( U 6)MyԂlggg[ux6Nz=.|=-lzGX:@(sqHyY 琋c:%*EK[5v/M (׷!{ ӹy8qxu4aM |984IsLzd8'KQs*aN ߂xz#4ři&O䪉\e?|dgαdyQJŭ\2^qM2Q$>.B# $^GM]9Z!JIN[ > U>؜֖=gc6RĝT謏 Q*pJF *JogNk)ānl]3aanɃ}̘ԇ(﫬 ǡזp'AȏUî P2 +E2&*d 8=nƏ/i="V*BQ=L:ARjbJhjk!p|LdS4{2#(q6($렌)TB% pX4a)^ Nsln8#!pe9aLRYapN/J+_{q.Ml oΓ#IqY^G=V?{W[G\1OY =&Y,~ /` /`b%${=sxaQR{-c`[VWUa󰍵ƉKc7,g-1+xwRW(F5!ݬr}J I:{cB,d\$^}w\SB)PSشtȅl,YV\uK-Fp}\ۡ{zj-J##6¹govzDŷthMZ(]I5ԘFR;O@`^Պ/V빋P%e V_9uh!Yzo-ԻԣKBVQTk3S'Mt]O@&JT,KL ͖S4\ؕzwjlTld7c8koOkx̶,2}ڴ+`W1H75c{5,ji@@@ʲ&$34mwO洷Rg7.>{Y|)Vu>EFZә<>'Ϛnk:цqMƞԹeCm\{\}{6ậYv~3|ޓ;N6$&_i6|9n=WG={ymy9/+zQs7 KKgr^yvoʪq"]g𾞃i"B)YƗ@䛱bAbZҒs &U'm.ccPiƴv߀sw '砽hohoii}%嘽bL'#9ĸ-i+K@LY x_q jP,!7!zvG]UBN`"Ez&x=B!"1BSA2D۞ElRVMg.ǜ;TW=x"RJ uY[SzJ4ҝ7!3OKnZ0ޯQG/ 'J'3:0nr`Ȉ6~`&CCS?t_>'`o}epW>Ei\O}&9NU&tyW?kG__#wBUN~خc1o>l:y;Rf C_7U[y_>,諜ruw+'ںG%wQ-\kˬݭ]y!W3usK>jZ>1UCwòRO3q˚x(rnş<4f3زi=#cufTrSO*lz!QJLiP$6Vxу?ڿ~~\O[}-cş.^_j{X5&j-dar5&].lŌќrN܍zS-1"rD!DF`;RJuZu=Dz0w6 eGI~Xj;Mq9w#%zw`qMgr`[F2%JYR:Fͱ[J/I20LbhLsY߼y{JK3k-Bl5W׹!dTEfČDj]jG. F1cs"1;k*z[^MKk1!žHfQxd-Д3fKq=d!*8qT}nyO1TaQULR̠G18H1d!8>"=n#M7r[;?9m߲I1ZLӮFu.E8R|'1B*Ѵ}pUTls؊6r2=qSE"?"}Ҵ%j1C"wJ#/eVIG[GwhvO jd- i.f2XR"P;SG7J.)/8"nqؖuP~5Xԓ 6JQUwL(ۥSYAP #k-ġ;d4 cm.qF4o~Y]0VglbRnŀ^XTn`tq 4sPR Gb糡i$0Q/`yU%1X^C^j Uڨn(-kwuS4ek8)mݰ^oC Uh'ѽ:NŔ*d,B BIYc-=ARM9# cJo%PG؋gh|!9Lg2[Ƣ xd-)n;`mRgSQ3JK, P@+Z'hwțE תwҰF cA:fgśnB vrBs% K ڱBb84b"T'g +vZ}<UӀ{%CEċ#}o 2{Mn%E n~۶)]Wfƃ!hm~z@!{:nA_cڳva(qH% :G quΎ$X[Hy7}|(b0,fw *ƳVIٺq.-KGhW6A(? F|}69nJ3x+Jё6-Jt9(wRt*@zLw 4&{`0wP}v!yqP*TO⋅XWH6+9n=trsp- 9N"o,V2 -¹!##Hq#Y}$@^z1@U h5H?xAmLqZ=1@csR)"EיxXv=ܬkg&-Zd2k!J!T$ eA^Kٍ%ćPe*9h8F1>u'Ј2@@h:h:h:h:h:h:h:h:h:h:h:h:h:h:h:h:h:h:h:h:h:h:h:_'# l@6'Qw}N ߪ6@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ :7aN DK;w =8>@ٯ9xƞ㣑3b}A=zsQ÷oU0t}w.ɾ.cВM%˯GIӋM[=; ϻс~|6s"obzqr\q!GOG^f6Pم_~LG=wt8BVK}p􍜼bf=HX奻A@bcK|5 88;>lE],]AWji_qs^3WqoP ?eZ FœK/crOfpM[΀@ږcZ8BX|:{{#vpn cdHʯW=3$Gcr$R6,qUOWOEV&All^17n!vMZ ~{$-n}{ (n2_tjMG=֩)q?'5>兀ifp՟%>Ko\ AH_Upr_MK׶"?{XL_Vf-XGsߧm~/z;/V+L+ì]K A¢=?郇KroۤKLw 2}FWI=ΫN{IZ֔\R*W׬Ne\aƽ*TO gN&?fqrvr'5OzjDO>(o:8Ehı壘OL ̔h:Vj ׁU8+Fyۋ`aѪ-{[A&ƫ[”~Sc{+NZ6]tc)VomXjL[<}fWlXѹm涷x}JX\Eu%RkW2κ,]^bfR FcL'dwPTl3V7{θi{p;5jn^3_3ݲɛd>x!ߓሺ͍sv\ݞ>c;:iAs;,˹r;,˹r;,˹r;,˹r;,˹r;,˹r;,˹r;,˹r;,˹r;,˹r;,˹r;sL19;a@8a@T;,AHs=a#4i.GH #r$Pʑ@9(GH #r$Pʑ@9(GH #r$Pʑ@9(GH #r$Pʑ@9(GH #r$Pʑ@9(GH #r$Pz:@wNoZ/ɻot4ͺ^ w]~ZݻTie 1.ipo.MNHTRCҭ>Unwө-l mH7r P˜,֌r;yfgPQt@;ko $VzUDTN^JSU&>ѪiU]- P颋i1FnPp}V.'ܶK %=0f'TM6.BPqDnL"XUG$j W(/M ]R!eo?o9ԯjn"'};sy:1[Ƣc\؋eE#]f=Xv}[/"` .,$a`ʠeyULb;kaW)EВ)E :]Gq>`}<O@DX|x@ybqX,~-_t]^֖1^(+I )3Lt*Ϳ$=ç0>Uwl>YdVK}W4W{z-J~,;(GK;"6r9!As$!h9 HRM71s*i(d/ހWh4^,By;ͽ>s0Jaߏ}X'?I컒kX(=k,K{u5TbDf& GDQDkۤ\J#"!JH{/g@"X-~&LOx?콚eUHخ(FB|iry q@}F^^~\*ZάJ Oy*D:OT6OKnSռd6ämg+?->Ni[6 38/~"1B!R(-,% ?bIy=%4uN~Bfj(+.G z_qm./ҪsM @\o𲻙ZjVOޔިGˢ~1:?;c<2K K]A t"hQπꏦ;(9G&{t9Z,K?3!"K&k]^墟0ō9ƒ"q)+H'@7/seVUQqݼvY_[7/Oɇ2Fw=?_9)6V5ݛWč)aJk%xs%aVwh(]Dg|2?^ u8x1Tժl}4r#}ѷ[F+|[WP. &Ѳyu#yu=gczxIX,ZR]0\"/)Jc*O<{%;8x^򃀗> acLBYJY [(u k _LE ;]^vFvxY"J7?`9MGux $l Pu4ښ#x+*kI>얮k+ܾxV vIgHt>Y6wz@s1aNE h]?.֛Nth[,;,jnm>-zn)WO7 rWZ+YBi +(|O* aL-CLJG z+2xkT`ӥ`36[~8_[FAuSJSH !$I@:ӄI„MHw:ia H1l<* ?s~]`T)4$RcA팜.Rb4LnhC>$t*, ͼx93Tך3rkJg3Uu!Ϻ^u{넌SW[^m~ůf0}' ǎ_ =aGG8w i[p bw(Br'ZK(Ij yI|;t$Lp B:+rkl;y(Z38-vG)ԊToEox y1h9`zO,)'ٰWCKg O*n%|n?J ۏsHb&g7-{_z{pSðbFsQ =6d7d+vNki @,Vls=,{+殲mO߭~ -JYA܎v=ӽZs-xۚһ 5  Wl, ˈksKbŘv{Pu %0!S&aJ;f2Rʃp-q%5^#55|~@QC~)Q;&Qw2A`ܽ.~.!6HTTqͶܬp we$:%Sp&xǽHk(B(,!0ĂR:J*i-+@AtBG,gug—ؔ^%Xlv5>k~T0LԵi{_51!4Gx1Ile]3yRU lwcFu^<3¯qݯՓ)p=?xgw%1Wsٻ6rdUc^#@d^ &xbO[,pYwbw˒,KXq:@GME6Wd4_T+bep eA><yaK;hI.aH ,oȆ!G̫5h8/d1f?09qTF6:dӨs5*^[:wq!y`h\VMPz^>?hoT]R|&r`6t*qO?|~W/|*|ޟQ^Q W`=*gM$H؛_#`5lyC+8|h6CC|qUSnsQb1Ҡ[Z2~}~~O|^ߵ?i3*gЕ0+H6eUR?VT Je(EF h:. ?侘ectKJt?c)b QJ=J);p~8/6O|Ϛ.c$xb 9xe_vrS:dPFi2r˙fIY(~+%-W9k Su_td.8#DebVqs ;AxaPKI'xlPq.ZX!ո۟-/?eS4h0ʸsh8@ݝKA5 i:O)4嚼S4 )TBF,8S$99l춠G4hopf ߸vq0@;<mEP`t<8Rx3A<x*O/>'Ωjݹح^{$m6w7e-z/7vY;~Dbyd׻m\RwW|ӫ?Æ3L[ ;Dž]Ӵ^y˝ ^uҝT|35wzz o8J[ZwPﱠzu!۳gxO1xG@® ''Ugn-~N.E?9^7\y#r8E 7?GvFy%ep,Zң,Swȹ/]❓;FA ^GT_NFxlM涕nvY|PcTY6n1罷G_ʵWu~/jWIdt6N*|W,U<ۮzx,jG?n݌f[Rƣ~GANgq045BONoY탤i)F;wnM"healHqm)x&'eHM5إ:k֔z"\ІH߼3&HV'-uuvi;viö{iKvP\l; CCZs4Ia.q\.p_Q`G`Ŵ+uxd8ѹk??]˵cG#,#IĴL(V\CIrAd#0h ;e<!1!W4prLWHqnk[ s#`ͦlϒ%J ewK4w1pDb qE)G{)%r8Pȑ"iIEX02 չ :EN-Yklgd/ ݰQ񱤥&OW s:Ƞ 9n' #^=djSSBN?I(Q:cP:zI1McR॥hVqA֠:;V\mn룋vm6)r)QA n"C&Rm֞x"m!2FM3ges&:AN\R$2MDmqyw"hqx] Eve>2om7.K*3FNcD: vPhwʨEc1@nme`!lZؤ(O _\ $!s6y$g ٮ~PveW ij#IYBX ! l!g>.Kܒ/rK-nUq!)Tɜ$bV;hr*"U;J$-s*;X8#EZUG9 (C[rYw*/>^GW9vБ!ZGMȦ‹b:LPREVNBak>&,t i~P8bxMwU枰qEs7R=BDFb[FqQ#Bv`$He:"yE8.Ϣa1סbɾ5hPG\.Ƃs/ isP:~[9bPv}!`)Eǔ$bTmA.Lɬ`YFhk1-5FΖejOG_h7?bW*?_FiLKe?g~txC+(%=wzԏé#\#<8|zMO&a/xrQLciʛ((g qYY9ARnzvg 5eմ0lQi"]K"_I‘u7J}۬ /];8y^e{8΍Y ^G.RuZ7m<߫E]wlt T)ji^}+W~E 4zUuy8`O{Wct6N*|WT6ϨPjӺg=܍l|KxtryܫNGEzVn⋺ӆ[VNxFBD +-G^r)몕Vm>j;MT+&kHR:"`KR в  !Z"q^+9!9S4jY{L@xbNzcpe0JRQkr:2X`Ѽ-[֭se r6=Θ]R}?:z$X!XCuhT ;ٻFndW6ݶIV5@>dg7 $X``? >cglɱy-$[~amKxYdH)$,.IҞQϡ8i SZ*y[C'4L5'y4so˙~G7䁰QF^Fշ7n7*#y(L tW]*uUJ߯ٽD vX`bgઊw :pUWCpUWU\wXZ/ŶURՋ+~ʺ:ki44ZVW5+Q/7>k|Nr{{tb[DiW.pZC9Y!76b@>j0d[nA=6[?QKXx+PWæG}5Ƀ,N򚟎2b<ώ[͊vmzHYGaխltl[רbsճRYɓc? yeJи1A$(ZH`k6KO+? [:cڿ3~kD. [3o͠G غЪbB? tSYRƇE_)}1#OY!(dg$AJ$%DX/1Q9Z}J!\R2('\t׃\oN!F/I"rH(Yu46*qDYgy*)doa}ݓ|pl Ė%ԓ> M‰b$PtT ̗I(>.SL ˧9&j +\xv&:\Ѵyx*mnhCfN?o Ygz-ddMyttJo96yO,r%r[RK6$zQyLɂ"fr$)Y'%Q;.8\TEt;v8[r.1 C NBJ.9msQ@ٱ0R6@Rc( R]#cgA љvbb OpIu~Q-l-iq:?/M8>}<>LJ.F!LZ'3%t LRfQdy D5QuBVbT*l`4Ճp(NrITDIg`LՈG}9Α+Wwb|sLQpkięG d1Ľ׎rD @\{,xmO6ˀDT}{&˞xTW^vRz0pV,^`QObgE]u/HHEl:Cd-`*9II)jBR&ihW]Z.x_&HsTEC֤@9l> h.&jM9_,XcM_Mp_䤬!e:e1bĀy;[IR@t'͟@I3zGCm)zH+̪ ƈSN9H 9| > ?2-[F o_x0k.$*2 PJB"% ĬT`@hwxmUi|Y9>\oڈS#8O[UoٟuMFۘx*tF%otևpr }]Waƃ'^^tiW ܊ϛ)M@ wk}??ORxZݟt5gW|T <ȳfXj=15oL&9G/e~TXt%4ނoЉԸ%|Lz&Ԁz\Roy¼d<&Cb*^*t4ZbbB(t _LBeHjnXvNlx:§%kˬFt]Qt^XѮ\*xS{̭9 Z^t-/j-t*~קˋWz ։7Ɇ1_Uë Y'hڼίzd187Z$4R5g͟=nĂ\)?co{YH[ϭ͟7wNB:d|]Q77O.~9BSB@%@9GPTz߽{}K}r%JgU)1c 53 jT.RbKք} S]L,T@)*SFTtY$\E}Yw?!5gO!Oώ>L춱4?jkn'z.'Oi`4F&e .u|e".) HD7VFt:g0408J}C2 Zt]AU:TgwY_8{YyogW9.)&b,dztBx)~1B9piZXj߼ͻL]k$ֳ7QGk-|@.6_BN/ݛC~bmBY >&=؅`=L *78a2d%Ataًr4B\$$ػ$.C_OǹEHQ! \4?򜐩 ;>ѫ߆UخjDWo̞܈^ckrV7׾x;9C^5Y7wҊT9q*+BG64 ߛzdIkA0}NU\!⡡A YEVXouIknnٺEv+v *VKŕ`.9DMvLZ;+ W -A,/{Wȍ Kԍhj+\8].hsIdSZI!) jw3n<~Akt!9yRGBT^HRK\i-q7vyLE[9ksǻAX, gk8Sxޤ\ZIJFX0R « lPq\H)Y{tE^rOR7b# BggrtwIeu#.`L&+y Ǭ\B5XLJR!gbe#,1Mfú6X4`d5*(SVIӠV5'5?VѶB&p- l,A`Er&:GkI)DϭpT1j/WkNr:>dkBW\T*=8KF.!PyՋl1iaħU"U>^YI~Yc2*4 0 rdtcqƮFfCCT S)]+h MHhq/vԟOqu?$9tλ"Qd0ʺ:)ews>4]9PʶWU{s"]}u+1&#S!ĸbU~\4 0: :ET͖֭hbې !-c,'CZWӳ2Vs)PfsF^!+ bƾ` hOZ{ >)O*gcE6`#KE3B~Z<$S]a]3d+|4:i9@#g$q<&%xmZt8Xl1kvLI~" Q_8$͑fLe$eQY2I;F@H_& ܧą`Q*+ O*x#I&R"iF8i8v@-q4MXCSA%&J>AII'D,yc{4&] ք!ԭJ浅зEm " ]d: f_zd$Z-u*и|^[}tq#Z$ѫJ2/@r<"IKZ08.aJ5-..ruڛ&^ۗM6J$pG0_(QwBo4e;Os3uR*wt.$]0J+ЩpMleϽ)ʞmS 7J/5PgS,3l"W{aethIQƋUO@{{j9n u$RdzŔVƯp#-M5Z-ّ, RNC}Hwk?ww/,ۥ{ 9%NIwe^eԠu#ZG#9Xˏ52X%&Zs*F+>W*2,Jm!|0'\$D%JgbYF&sJS8Dfʡj O4 pV?fp7zܞ؞{o;B<﶑XaYS0*X'(N^9E^#w410}4Ɉ%*&D#ViExP˯3 k! ` )ҞkU6Xs&`jUYQٔY>);FkФRJ^qK5sdY-}g۶gWz"$]zAghDi_+}iLWnxjp=2:iCz#o,2/dJmcq̸\4ofK%ckL37ƯnF_2N䃟:;4%_.W/$]Y',ť/IP?oJXRa=G7.&+LqhQ(A#~΢'05&AO;2T3ָg,{5!0keݼ{|Clh\ЏW=̷x$Ꚇfr ]|qv'aUg@ڹ@%n}|hmzx^<-IYi,>1M~\9HGmqP|EoC׆JSIis:ONsVy}FoqD'hv7KwNW{Ƽyv-kp9Et/Y .'rO.w R_/ez 3>qjnKY'F.I/z $Yʄn2ѽK qJ__,{TZ:Зҁ {KHݥSo^XB To6}}=ly~z𯷛|{~:9?_,竵.0wN֑/w4d&tm۫hݻ&4'+QϖKzWOY i勛{ K+˚I˘I̕\=se83Gd \c1WEZn5s͕P[uD*⊣1WEZn7s͕L @%8>ͱ$MoN&k'l~M"\w6Y5F)p?s>ۋH.Y?g'+?aU@<~Eoy}uC #I)WLyJkc~ㇴ").f=ckC{d ɻnO,*˲ss1>vt;7:#n@sE`GcYIZ"%o\ rm÷q68|kEl÷q68|o÷q68|i÷q68|om÷q68|om÷q68|om-8|om÷q68|omOk{iUD68ޭ[Ide3RN% 7~Ť%i ÕOYg,댅[HDe%p*Fc* O DЋgƸ`-JbVay Ue 1dVZ :NZ91%9 l'+Cj쇪8XН̀HZDH2EH/aM5~ IvDY/] }P,K3I NqH:9#Hʢkȳdv,bus (5d/E,5%8T N&&%hu4EUYWs`iʈ4ǓJ(yd2Ax'v&ipi3W MJ Btҟ/ d^[!{ܦ [" )2@NN$Z-u*и|^[}tq#Z$ѫJ2/@r<"IKZ08.aJ5-XֵLqoVp kMj/ek:>i mu |Z`ޘoKG̪Y,DMp"v;bn;tD>4bIij4GZF0Qy(!S|pK y'FȒҒpܼWV)BͻSp0.2.1,Dˑ'sho!Diڨ%L X)8R2n W{[y w{:[Õy[A>]6rp 7numC*)t!ˠ7~QL\9|wgs{]wer\y- u4%ۇ;-[ 5q`WO9Nȅ[H>FPvzn,Zd@UߏǦ/ÜJ*ZM.WURfxg8Og]A?Y^>AaiI!(i)E6`*#9mpJ*޾Ĭh\Zxs9CdRrI9̞QgoAeYob+O–GivL? D2.pQu`.uf&PStj4AA3uHӆgKl]/=kRͣSYJ>0KgXswMߪz$AĢh&//W q#mWh#80zdZpc\aLX  ,ZGp X)ıRf85c2(pQ@\~*m|߉~ui1w29:Eu@>+_5˫>>+]שbՉFhр׶GM$Jk>.HN| efqU enB.-PR")e@ _"hPa@)7-DAu6QP21C7ԢJR!Db:Q"u "wVz=S-YQ\(,-򹣾=P\Eι $!I343:SLxHr$*Ae*Odd8~4XEIn(/QhhbBm|Bػ#,~ *_MπZ lKXunh\wrFV{GQriD᪦*dؠa9 o?^2;*WHx4WC (m(T[Ü8#q:%AY^hadcr8  Fl>h!*T PH-[#g/adak,T[p' WO0>xN\auܷ{uF&w~]tNs6I9nNdBȤ$!f; *0fb͆PDrñYs=,MtP+59ĄTSDl`%vk,ؾ׍*Hָc[I]JYGʠRWA"O'0>5>ȌZ$2 H*PCFdHjсFYs(,e~|L2J5hpk,R6f5u%nY"DIE/CT+TƔ!*(;HUJUNEEk"hAezҤ]h %D!%Tڤ4Q j&Ԥp@ڎP9%)T{&a6u%E۲\4; u#IP;낍 yiOqb%!Y)\|5s[ܱ-pxC –d3[V{{6(jFp;'~dFWóN|*F29DEwPB-+p]FdhYGg^7HqnUJr ^U):UyG#ҚivÉ(5yjMzWU6'WLa)FQBE&U;!6P{[FWP Y!k"`x|wz}5oVsӹʯ|m"%ܗYUZT ʜͥz#6El:/&+cc kw_0`|@3h-B -O:ل{ wyL aI TRbK/EnYb^و}夃D@}9^1s`l (sF#T\bQ3be;xr w>2mcpD6a!AnkKYP5ZL)(BAg%#u"l5ymo3#YdsQD %RBLyD|B`]voyF1,)~J=y.{a>>tiTGiEY .8u2_ اՁ|F%IO!}r(RHzIBfm&Z+Ư8 gx28tI4_{.Bg)rp_KW(Ⱦh~̺sxx0JS N vl?RljcJYIc} l6?Ru8VѨtQku\GMt$8MsᾜuSYnoFvհBt@ d^KFm)yͰ(I63*qF*#G1nk3Ѓ6.Ά :*#[m:9չ}ql\ KdW*UݧE%r.T!O|+;֏pvy?~}ǔ㣿w XHPrc،MVuMWm45Ms#ViZрҮnsڽ|oОZ- ?]}yٍuRƈ2k47{&jgteqݍEl^Aq_ZQUPd1U"V" X;n&cPG$e e,f`TPg4 Ƥ/w=.~J3I$be$pe?=鯽ߑIy+u@!uaI(w:'9[xp\"PIAy-wJc4򹬖m<K2*4w,1 ^fu:* J p TgdwP\B `xu-&Ě&kR7ݶU[3آ|ٲ 3$Cr!vjRo*IeJRQ4~h$t-Su(=k?d@n DI>;R YDiЈUD@b4fT,QkxEHΔt%ZR@ %' e9a~ot]Hb:J^uM>p*r7`1[p 2"RJ&An(r8>&)9XC-VkwF˛,|k' Zr]y Nj'o$`2+9'&1|Ď+&/U?_"ȋw7xp(\=8zpW}،\!q|j3jkSqV، _TNEnp\u҃Z*q>zKsم~Pg`aTG(#7!]v[8tAE(/Ui4>ƕ+R Qvw@)Ė[WlQN\)WGu8EIS@=l dB5`8`NK` &2T"-U(b8ڶԜx"Զ`oL3ge3 `W@zBIt5-,Ie6<󠫶l=X@EMmV>x@=4n_U[V;M a$DCL.F ]ՙ=(5CVX;s ^pYio[ s 5끲8Е(Ù}7.>+bx>_C-Z] ;)I᜻FX#cCjrM3!T= P^HWzU( c8kM'@>j)߼3&H^[ETKQ)E90@=ʲ?ϧWreRb[!-!`oHL_] 5%} !q;/O'=qbKUn6)w7C 4U&⊋@IJ"Фc\rn ZTeN׭!A24J7x3Ùy /3D\1t+Z,-"XAPD+J<֟~~2 y,lhCp<91[N^'?˄]q%~L?$0w_Op_ހMܕJ7?~zϝ$%'9L姠z֮kŃ'PG ?h#2A^Pt9u3k_kΥ:[[02 [uQ֧uW޴,jquw3qUΥ2Y]wYRl}\NDN>]. ѿ9NE/gy9i}4&]-Fގs&^`Gu6cdQO8>RĽ<&fј,.Wgi n4?-1JRi1ZF}jYU8d"h_-6ݏU7Ns\:48]-7ޤ1QG']d QB+"w @}0Skf{m4FoO8&:jt6nq9]v} n\#\m>ݵX>ݽ: ;?|kÀބ^1_wm6|׌.,WxG9d˘Gy%~^3NpL霊-x h+^ qhIg/ |/-uݾl~&7o~z08./_xa8¥GĮJeCWsWPdL*.-F"bґh%0%'%7*$SD}OtFkw|{4j=i#6Q.?9ڢ3w\hVnܷ:6)|Ղ285TֺCSYX+ kea,V'TQPYX+ kea,Vzx֤f8d5SfL)L,-伦)>Lq`CE["WZJ\i+-rEȕ"WZ2HZYZ+Kkei,VZ5V=,VZYZ+Kkei~Rei,V֪mVZYZ+Kkei&$hdTz0VZYZ+KkeiHkH**V@v^4y*LP5ENBck,DÄ:Zx(TXUaխjmUf,|:d8j`) |fE21JSS|q)FNkI>uiO5kp;Ӧ珱c5- uĩ) J Sw iU.tls.s&1e*VJŹsvQ:.B?5 Jn޳[ivyA\d!A)U}GջLisȮ?l/cÆհS0Ԯ!w߬57d%O..ڄ67<[NrC`LPN>jFx4j\-7Z񀯼ga|lG6/ݟe~".m>p _VJy67>VHmza!$9m&)('Z$.Vړ/n3"{<ʑfto3v?QI%܀7P߱m>C@\ֲt#{DcJqkxxvKߛ.&co }7Y$?>U)u.:}ŮxKcxOLy3:uD&}n?c?msleBs8ζqgDt)d[":T\H*T*ms]p0!1)]1EO/&ZDg`xlҖ@J<9tLf  2k mei%jm/Z: f\Ɠ:7-M/>/CIxyFD+-G^r)<8 zXJgCM "`KR e!DK4k<1$cƢPՓd<9 *ÕqP(I!F P<`AED CbUafmK8(:ocӃ5TGFGjGeނ՞&N4`3A I9h(?# & gՎs2LYe(j8 t!RlxHKR54Eih$^9^p8 cpK 8 HRRD 6+3?h&tS;L3VAXI,(ˠ;C385 Dձұi90i71QA5@@J2t\W יݼDIj,D!Rc .pиm!*NYCIjX$*p̓vf!x` QCR!&NQEz2Jz$ptb9JH([gBk0% y?OpŬڝJn-*m"{[C*{kiZhzpqZ%CZ;[2DRL D]i{;hgAQ̷O~ ?K0~M`. N_7Ӹ{7?~WEݤ+qzJ& GSF.z45qp-::w=A^hh[}MC[{L#箓Q'w [?~u/N_;/]io9+ }l` UTU]my1}Y:$Ec #bfOa3|E}{=*^e/ / `:: ".F1n4C|_G_[`kLxVZm) IsbJ=(bݹr#Y'EѨs΃ז9%[^IfAP+SL:`Šxq4P21C7D/(18-9w"1Z(O٣ҞlklyDPc.ylrdtVeMZa ]u1Bʔt̏3.-znXr4/nj 讀dռTr| k=-5gݟtsĶnپV?GѠC/.3Δ@DK-sqsP?okqzяnBBԄqJJA &X:sfwrw'G'F|y@XLT Zk!N#_ASoMO+ RT'x;fj|g:9j`ml}&//PWipyg6BJv&0c'AZ10Dv,7(KQ\8Vf qlj͘6@\T)׶#V[l^WYO`gVݶ?{c՜yE,YZ%9u&42=֧R}|4 GM$Jk"gՐ.HMG(!S!7%ER$d/j^ a@)@>z-/FowE7],k߿#\psnyփĐ\)X&<$C9C"Ʉ|']Ae*Odd8,i)c '1*/441mټP\_P- wy+qvBJ{2Q.o4WC (m(T[Ü8#q:%AY^hadcr8  Fl6C4(UQ*KRm#cklڨ5]5,lmf Վp>i֌W.nyi|6ͩՃ_(Io {$J`}ए<ϚSdZ$N6MmiT CF͐ dсF9fB1 ?p>&I@R ڵ}&ܚ8" w&|[㦈[FD!bfVHƔ!8UP>Drs,ZU$*I5PRpjҔG-X8q= LȤQNSZLh;kklFoG#.g>1dlmm˸h:\pq-7r*1#.Ry'VrB˕>]0fǮ j8(ݺKpH&nҨ3ڿ8ߖ#!RP1g*-H@<&ߝ-iBH-6Hzbگf8 Џ*>z/;^jPh]hYdqrQDY%΢k;O@S(`r6')yf<OWs5)nq9C;m=bjGNzYr1$Mu XCV0)rn?8Z'rhQK"JABd9p #ĝ[cR:\]H72F E P.D]LS.ē\Fʤĺ"?\w3K}A夨O{'60>!vl b FXRWj`ޅ\(/{׃l]6̞T֧+0V⽿(R_%'SqNf 8o )(!甇}7é?9OOĖͳٿF@U.,.@x,*u:dSZk\x MG灋 W[e{q gUPҌ/%d(;X߱) jcJYIc#s{Q]3:\RռSqGM.͌\[]tr|zY9,.9u{t|T,]2D2@C{U-%!ntU3le3J,O #FQ;r6iosȕ2Y'WVQ>ou `Br?R-/>-!%˕JRs⏺^~\_~o޽|u;|!e_~| wL ưJ%o-o6cpyMkVެin6M6{ 66GE+B Z@*~~Տoz:쬵:ɇN=Yq/+]ظO|q**qA*Cqͥ ԀUv4:Ŝ/s.qOZ_"Jn3'J8܄aI$$K5n]q}Y1>-ZvD-3+cml;*(rNsIL1 R%4(E'`e/6&r3t+Z7I-†xie^tz^az;L*, O˯64%)8P_L)P9(=gT0? `~Ba&6sJIgrAlHrzA:C&Q-P1 D6jQ_3%s kT u&&f5 1q4ح/7ν"Jk92fkڛ|!1yKOזJXm&ȈlӕV2i ro)cC*4lx[YD7 P+B,`b%SBı?~Qly>V~ ,5CT<"W t>H%xFَQ8j7;HD8jJSZ"f "F%9DULS="[s"cAh:"2Υ˧MI뤭M$B$SkA?L+Mxj[GlM:""tyH|&kW[vgP]6 \]OJv 5tpuWO5cT<"+h*kc+ߵz{R#\1-"P`F飁J\X*Kwup k}DpGWY\ \eiuRJs+K/,_R}*j5Sl^KG ձot˜(<U럊`R8'rߡ%cOH߾qD_#x?՛7{E r {͕]ή>;9ɨet<{n#jLwRVf1vg[q֞8#j4}X ~mK=PrްJ1Bʔt %Kup*ِ׏fU!,d]7j^\yCVy0 7{>#ŢyԶKtиnw Vm{&|jκ?+mœP/s/՝9 sXoNl![[gJ| Sԅ#yqяn QJ))5P c Fϙ:wܽ;ݝjLRQ$Ytb1Q"&h2D8L $V;˜2 H%傩DU p&q.2woM-ȻPGgOwQk#nW-vU#t^L'Gu%=ٻ6r$ewid 82 ]`o,{$yb$ˎVl*3 Gn5>X<>xJuQBDEP4 LZrU E$Fǹvs E7C&lFύB1@2lĹ݂⬯ }}N.=Z3ǛzNQ~^g,.׆ #=\NZMF= ҃~>Q,[-|Sx&U#eOv䧛,zmz\WoZ^JEM*725ˉ/wㇷ_o+y9Nr O]ɴѿM޾Z:$[&dIe,F GGg3ENU-ph2km :$1dhL:sd>SdҘ2b )u}>=.{WޥaiDH|[GAQGRDfY63G'dl9 NtcIpj(q9& (U@/zyDHH>J"a 8*LfB'Ӄva;jzrWQIw'^vm1Kdǽ*Ўdb9Fc9G 8+|0z&gE${ *SP3jkrJ9f'۳)V]dIo0isj{jajfʾP}V}ᕤyA=T N?L| r}w4~4:}7Ng_GAYKBJ531@2*Qj$fF9($, ^wqj5b!{> !m6Qˠe2%1벐$爙C8{p:N+sWv5^{`ol A[b1J k3R>!Z)O?*.!2Xt4k 5GQtUbH5_;8Ö/Ǹ+~#=#Xߖ{FӖ bxtkEr6% ikTpHF\L {5jnmĤ 7\G&%2^u 8],0ۥjZrW*E{x#7: U(m:]B"(+ '%$Ę,S\k ɚ_܆_\K;vޱ\XUbZzS>RRV~\F1#["=|MSIӘb%cm 0$OMR 'bȢ-.9).{/Cӗ MK\'g<)zN&%Y̘- >SHa$X$Ep&%ސҝXQ#YL'"s(/1D9FQ0!~{XRJڢa =b̥Ap>nE\,c"G8k˝c /9GoEgyn=wɧ.3Su+ų\uodZ[ex 80qpt'a`:#7xr#-!Astd|1 :Pzqsj)n@2@sf;? |vap[W2%k}ѳgOg'gOW589\D .vMs%frx ͯtaF;=|v>zyˮd%+{[*׵t׌mF+̪&3L8>m}`Oǣ^9<nx8mUo!Z}V9,[9\XHiXtn~>%eD)?n/Vn 7դv7Ӄ8<;&u/}o_}W~~qyuׇ\×uD=0kHսE~ܧ>i^yFj[4 VnҴߤ]S{||ޖ2fpK\k;S&.c&a]] -Bbt 8E+_Ie9)RxZb A  Wf>g`cȑb5^T}x)#\K)K.(k[}TͪԶjHMP5YK 즠G=ߐI/}v~yb%VxzG_^Qb7֖IJ8BpI&$Ι+!YuY92*[c,{>RF% cft](&$lߦNn}}z{ݻuv|n3Wd_7֫q7 |(X/; 0 Ǥ+Fcj}Tos *;Usע{6{@PdunkP/ јx/w!y YSnDY|SR=UEZ~c=Xyd2$C=HAbp.if Eɩ݁@QFC]+vy {@HKs۬fEæQऑ:1lb&AG~>>Kob. kOI$vVlMVgڀx6kʹT<[,x8w$9n(%JsYreR(X$ќ6fh ʓĹ}ڥ;}fn/?L]IC5t+' >} ST>feڪ|as6clthlxJ]/K\5)hFI&V.Wlaj>n澥&lr`id&s#~|5̾yQ;!ED~,F,1!ϐ)ti@oMX>3٣v}._T]*goem{Ja_y,X1ODoi972)I"~*I TV^NnWS>ZXYlyطՋn?_ hLn|]]CBmզ~`2`Ǔq:qo+kb{\6RxhI ~OD=sDVH>jJkq/{uѼ.&P=C)m=#WנӥMHx#< V(Lg3^xxO'(F??6)<#y姯3uŐ@|^l_OD>jz>+ߟ~Z]iP^Rʮ4kGqFvJdAH6\g_ FYJbX#=zWJ^UqIRͱ4$mh4.ě"&3DꕻT{P^W)2Ƨ,F52HI,Cʉr9gSQVYhEn=ZU53`JKeC*ǘs k\ҿk%4]| UR|yuea.KޣzdjP3YgUUTdkP `\TAEgEMd-6vv3v=3V'=D"L$,+ bN2 T90DQduU;,{0v"#'тІ,q mA\2-ZTjlhgW"Zޯ, Jw:ϑC4^9l8T< ,QC{xReﯗۛLkO?C6J'ょFq,F)ll嘍sUPԊػFn$r" ${^6>%xǶɞ~V˖eI%ʖ6⯊NѾ\H> I 88l6h|\GkiI+*c:3ҙa9El; tsX92pq^Y$K˲wF'BCG|)KJ'T4gǔCUk0քHG-T:De z9.kp'~.뭉hHi!/Q $J4 {8Mb&dHztx~`Rxxo)\^ GoZVRq/˞`>`ُgǿ^o&%pfn;Z|Irߵhҍzv&FG=k{)K<_W/ؤvWxu Z] H'w#h2`.v㳾H䙵X )s}(Mo- ᪻N%1b򹈴co'_| jHzJ^&hрCf9`n; |#v@sE@P0i/X*5N4BKd+עJMDƄ;2h|@ރL8TFZ^VU ;kqT5b @YX+8&$ޯ8<*-4b4]Ì%զskȊ' T%"%NbIn ,ʡqWRApzSI Gcz^<[ZJ#ܑZ\(qXl&=Z9Ĵu橢y \wܡk8Xs<{' yAj%7/L3!8*cC/դƱ a -p(`>j[v~i`YY /PEAvCxj\ͬQ,ŌA]`",K=fلd,:`Jjj|8X~36U{z4[Ru^n[coIޅ0gAwQӝ]ۈȻq,lpϗFSrA${pWRRzkJzAytL/5AC$,Zjf]KIybA[ AI%LF g%$Vb܄Cw&|CfB!ie)"0́A8f87 f $e+q½_Tn,W-c4㫳_2_bR*n &@-k/A^HJieq ehxW8zwܸp_b21 #?#3jcg']a w^ne87GW[l9 <^v _nGR;lj} h)Le.P,9מw޲O_:}Vأ&Dz'FHByNLzY2 k7UG_Uyy.q047lz鋚^Kx| _|`t9Ejo@c7פI/-z_\I;Ϡml*ߓQ‰}ϯs>kG5',5FhV8>e?H$<FعQE$TEE;-[ag1e5ʗiB8 ",pO6)ab)poFgI2MK{/d![YR2yP9fYм~p!F=bi!*Vz —P<Yh {W)FS40Ŭ,iLc9KF)|̑ls,,Ei@H@WOnYmBq&ʗ83f\[xKCi_.l/gQs-!EcK@)1`76ꘅI pL%`ӦjnzGAer2Dt G!krTY@,xaҨC:bݳsDzFH-DdGLƘde-h;W 3dPk찲U#gM9k f|Z ˡxr9p+z%A*至cʡߌ*5kjol*t5î݉hHlD $Oi& :VO"ĸbQ ڨF _[#]>,BױC/>.uy[ӠltlpAfvZf0mZV w7e~n~EP2mpkbMy9hu|03$4d ^}`k5wP{ {I{ }^}`k`k5 @i@BI EQ8HJ:Ć68h|Bx@!$#IfB;=Z 5AnNjR&,Yf.DV/8a`IQ U;+~yغ?IiE3 lgZ8Bv~? 㜃MrACGŘ"#!%J-ivTUׯ gv>y4=yꭋco=u4SpFB lTHQKԢJ X%8RTeǫH*`&C\=}nO'ۚI3?|KioR=I%$䡲(?%jrpEN S_I'm `7O'a݉V0;A򾪝MsPH'^蚨!A㮤`BQ._ s6CSزH}&O(RFy7wB>FC9sGqnW/]o2& NXR;^}Tƀb>񀢐jH.ckhˇW9]^1|+!4J  Y不DtTr)1Q{-V%}">^ay?3?O3u'rw{uM"Z|mXs @[$WCD*qJpF;B+ T0Qj􊧐L'W=+Q3F9O^&Y0NHtի۫hyN!ig;7&Z~uFD~6pj)N>x%EII(Y+-{W̺1OB꽣GF{]=2Suk}iwݟQQ\ks{1.c4'CNp s @|]oʮ~-˵nS"Hhra>^ԩ!擳:௎ k/wu.ErGڶz׼ՃT;f6\_nI~> c_n>p15_.qkݑߺ_W y6;{ׯYU`l$y>VHA-R+˛D7b$K8:#nvdrS: G "r#p[g[8lMu>=KTCmՀ Z=ڸwl6mstgVGm=Nj|Զ [:6q7m{a3@{.6 u杲N+?Qq=?XA u*o%ICJ.46?[}v`CD+ (w<_#2%& Q<|Q'`@d!r-% FEdXFe C5=_'to8nknV:=[@ǾhauS/IG:7-"l?u;P" UEc$V1S 3TbH@Hc.zJyY7^a}ITXVztޮCN!c,IV!QSNpp^)@ؤVZEbwKz=qED ֒ aδhp"+d&F,Z <6>Ь책u'dzYr1$Mu XCV0)rz%VD"+GhR `\% xDpcLFSt'$e*De.x'@u3La'G!p#Ww2)wv p_?ԟ<T[^G$(mG<'眰r dzrLWn4^CE\t,ާW*:8dpq#jdaPN9甇#7]O`dQДK!KENg`=Z$7^ q8u'UE7W#\ \JY™}j/OVWS7XN'rvi|lnL)+ ]8ǵ_R5~rGmA<]ksm}}U&1KzI̩s8Nᇓff+"dO:7^N/!]#9G:u$*C10: gvq2],&zv9t8psQ<&FW(_:[:|4g {ğ#:4{7XT *i:˃0<;Av˿~|wwo^=~1e?^{ W !]$(yo~Cv.14в]˸-w{S|re|5 gM~hSG/LDlteu=b~Z.V[ E[RNcB Zw3g;7o ޘƁ2=t9{0,`;K<8bcTS^(=aZ|#@%eTiXb>I < DAXBRq6841~cAE\;F\qR6Uﳽ4ʏ^]0=N -K9|է?|$+CUXZ}ǭ5*$K`oQR$Z%rAlHrvA6CEQ'N'rϜ,]Ƨ->cȇ;Rg:.3AF4g V2i ro)eC*4l/EޞH}njTǙM7 ^bWܭo{|H MϞ${G l\Šeh2J*CUO(ۣ=U^юrE{!Ǣ bِFiV>ד`tdԼi|&# Mcw'Y@;Wi3Iƿ֏:Yon[7{wp*Op:^܏`KN쏿䴒S_~~r_jgLG(5ep,2PrV+,[[T^4vݳQu7jPypk$mjumvt?I^Nqxmi pɫuC.Š|\e=h:ȉ<:0/X{@L@MrVP$8aN,2h. B pe42Gp+lR>jEbI ?/[K8%3xC-@r>/@Y"1Z(O"3U #ghF7?3fo-]]Oum~S55"7TUg.??!eH9"Cqn]5 -erlȌunvizo|ٸDEWZnd[߁|tK!__+wh’5h$&.N}@~V;dUֻǕhz܅[ݵKF"p`,~{dhe MCd**P`)U: gL! P2QBq%9%x8u$y9nB1E%qE, Y2j󮯋 q*<}4ygM$*kVW.%]Io Pf'(3E@]ihV,0\  & h r/:p 赼~b fp>Rrm科mTQ .r-h= @ IAYd(g$Pvc *-˙LB Gne2_K_Ep,£ }#&&blTkWS4Kaz=Cv0)2nڲǍ=j2h,Q4PP ੶9qF2tJ^had…QbL#ʞ `fg=D8 QQRJf,FvXV) q]u:꒽.M ϳeܑchb]4h4<O\c;Io)F&LyL"HPA(@d I(1Z<8j \d{8j+JDlDk9161aAr H^qK]c8^qǡhmkނX[G )Uȓb*ILyךSdZ$2e 0Xo$2QtQD A1`&(4! $# A{ٮkPHȔTI/wՈFԽF5Ɣ!4 29\C$Y44Xĵ,ZsNkQOR(*Sl咴>_p(\HMi+KaOMwU%U._2zzhy oYz *.YeGv+Pȧ7š^teOf ֲdxZhhGPSk'"(CSJzT+A"% ƃ6bs-I' "sVzIYJ"&5eS[LsԥՎFoQ% JW UwA Ε  6SF#ϊο~Xj59G\<[Ywk¶)k[VJZA{PVbOɞۺhALg2F TCjʾCL%8!F3RWPG]er9?u,3kTWk && BJIZ%eI*Q$RbBhr8PD3@ALe(#*GA9KVPKz-ķk>-{œc SQWK͛v'7U Y_gkufilv}&UMqu<[{,էT*~NcLC㌷D22NtfYc̺.j>Lmܖ͊|,kŖ`s\HKU6&aR s ff#yfkʎ)\P*9ĠTH42Аi%ܣ%WH5Hn1ɭH("N(-Mi Gulz&H>]]f(K-8.jbDN,I^:.U{O:U9A:T%)fhNsP:6![9bPaCضN)t=\crM`u!&Dlg]V>nKclu3`gH3~̩]9lA}]5ƅq yaS"$4Y^jfA#٬ kb];i]iʡw˿wrY| Wwa y޳P DzʆVzuGm ({[5emخrޕ52}uZav##~$gzREɼjji&"˸ߺb$'Έk;t')Q@^}iq34_wG=?01s}W'&5Dv5L[owytM OLrUY j޷Uv 1M"c 杲N5?Q1q|Es+/)0%C)J}D% pHuLJ"nyˋ*tV +P<%ON+S(B1*Y4fI '4p2J/sOufK'(Mzha\{]z Ge(T/IC3MKE)7D_wP\“"*][iW{ɥnh[IO܃h*`5D$)r\R0%eU޲%RRS1EcQI2aDqŜi8atu"^j1rvCU?~?ǽ5;?Y˗#Y)$hܡ*ש=N&g+Nvv].\bTHʎ4fHbi$ָl L==y:q@p9蕷`P9FeLd/%$&E.୷A:aKaHI8K]au-ccdQ;@!bB' Dw"0Z$~M|CB=e*==hƱV1R+WƆ6`0`)  喌rGZ-USkh\5Qqd (1 U)Y +ia>FfOOEiXeldl8Ue7v2mxW8zw# A糫je&:S+nEJjaB)s5-ph}T*4xW)B[9.oѝ|wYHy%.dT$̂e Ys FGgda֞`,uLяL*Ou>8XvK̪Y[rkiه'-ܵlhzkG98Nc6?tpkvFZZk}YZ^yi֥V۳{+Z jG>`0{cL8`QwY(5W9}1jm"!ʜ`! A}$%6 rw“:McͰ3r6ؼMG-dtޔ!Yjt|K<=_՞ڒs՗o6Yk,t]t޼isiq6 +IkqS_>Ab.:Y`Tyn0,nHy- k.6jz|p\pE{|y3EMkNO{] 75gß=vǛvo6;;a;|y`_zqlsʹs`xH.60\"\:tj8Ȭ0&|ZFj)11{o^)F$rkx!O^yʠrFˁU9l Q ʗ`ZꕬVYk2foj|lr8-+I5CzAk,@R)Zv%K6rP o`QEAZ4LZr "XC MǃG!E 47!e6c8WHB#]G6tEz Q丘yr\e9lSr,ɲ[S2o[K2/85$D| /5?{?ѕW~,Tѥ靏Fp^ iHB%D.ٺ8Q ~U{|!'$ e R 4r,șK{ɶԮ MYPWH7Hn飗m϶%_IO˷toU60EqTӬͷ78YwGY5[6;`܂uP釪3pQ *z ~EXFsRI@<e2?nvJ.JcHlyנ-ͲvQia{g[*nK=`NMVHE2!DL2b )u5kTvgW@٦tk.$p hl",.ᖤ3GelAKrɌ$ *[AO =QO k%4 Y/dGdEm;lK޹9??S;E *E#c!Ah,b,֛%HAiUښRYKV][do0i(3T96LW ;bޒ=n^c~0qrV4%Kۛtox#W9 %1TRA,RFLp<\9bW"kbę ^`Sh Z&].s鐌1x5s3r#v҂-wl j{ vep"hKVl)D^%K"k31IDl֢ u[<2>@L!H@CXJ7FM#a]!eX$KgƨM(5"M="x LSӖI!}e ٵ<9,sXڰn5Dg1MF% d$gɣQhc$K Hbb)X;#g="~ǣau6h)T]8H*>~|-;;ǼlRs$8$mfe/wZ}`hd 4Rc pQ+O0 זS^Z.h7B(R <+c*`y)ZeBL.aщ9\PpYȍQ"Gi,7MZ lTYS\ 'A@-Tח >WB6#|lc}la݈S@> hB2lXv>9G%h!$瘓p QؿmuXH0B{IKv2)bl{FLȊD8R!UnVVtI. J!Pg%rKEGLxyK\]ngMD0՟94ܗ`hT8nD~lbu1 K ||_*0>~sP0퟽NٝISCWɳ_%O0bNظ}d@0-c< >zzo I1GG@K#J\眚iBip0)7'L={թod:K4g-V_88x699T nh?nu8kiFt*4oK#>]7#ڸ~njn|w69^x]'Ĝ?a;]삣mY~0c<-ʥ=IusO'˺Kъ-fU^& ,XV|6>N&+Z+g{ rYK4y3 )+b G4}6&U?}&5+*QM)ڐIVQzC=3̫>-GyRI{o癇T 򳘘L2K j͸`'1 7v:t[Nsu<%X҆v{(|1y5zUq0)A aRx0{LSVR 2mkIA }BRȭ5Bi \$?a(7 \ MzA!!J:UHҸcJ`f gΑ%c+A+}fdPS<{ >դY&&uֵT|g~4kѺrᓇ>6>?=`~4i*:[teS\YK:f9{X zDxR7}:$U#k[=.6NX1ZڇYZz+Q-/&ܢrlr+QRz6`q@])i%f0Nsfdp̔A:P ʗ&L,+L6e :ˬT Kଵdr`70;#gH]_B/ [YQdx\#ty}<=꾞Sb3غU5+8*]k6O;Qkͮl"jmTlz5,(%U!ȝBWZJptʁQjC1rݮQ;f*T ՗W6{@v/gcg,g+գgRgRV}ۼzkՏ*A''A _l274) 4fgA:,#D-SuuFk8$'1JL37oĽo&]F%^gko©ELGFbxv_M^u^ea 㤺6g{7~co* 0; 1v+Q@nN Ic٥Q-JyM{|MO"ף#J,Ĥ0}lxRH?1Sˋ(ciF7L͟W|~E9_Oga,kWD+I J9FRF >${^>.%UagT\+Z¶jB=Us#!*ݱ, F \jxwjt=\=A3;W`vrUVmB"\9gB7ekz4AbZs`mh¸:P  h8iM4rQ5} ݁T {[[r.W&Gm†ݫ0c~1 wnq+A=<} r<)Oq&L]Bʎ =Mi65lիV Wyo^pH|.,|y?K7& 9ɯ7l}{.N$W ,ȕF;\)mk+PVe.Q('+>* q+=(M.PL+Jq"WJWﮔS(W.WˀùeLW-\\re7z[I,[ceR\dzȕvRJJ\]\9&̧J3vRJ%ʕLr!M#WqW^\)&W)WK䅑_B0%Ƿ7,%o[ho]_oe%awh@*ݾ ~.vɗta{%5a[K ݎi$LܓOc]~榿vC\F%N/^٢q{ _"gtvKB4+\(|=|sK~QF,0%av.6v_:"0 LNNjYx.iN6-WpϽ}"ZkϴԾүl/+siW}"`wbp . ar1mrurɕNe)"W&]nrurm01L$Wp]Y J)6DJR\x٠Se#ۃknrL2L# giM~2 Jqf ir.&3^zڕ쥷mw@i&$p '6Ph{- Ymi;>L#ehFqQZZi=d* hc3]9qur&˕RʶN~r%H##W+Ӭ+n[Dv&W_Ei"RoWν2Zk+ES(Ws]8yJqik<6˕RnHȡ+cJ_Z>=}U nbwKk_zytW%աo79ͫ٠t{y,] Z$EDA~"\vߠa+s.70[kf+9uvs^JJ)H8γiEdЊY)ztDh"R_7#ȕҺ/"+%Udqs^4JYRZ˕RM.QRDr4Q13"WJKvrk }/+>rV`! Wμv6]-vekW@x "!L$W+vRZ˕R&W_E^wˀ9M#W{ ZFvSM.Ph"R+=Ѧ՞2mrur>9sLD^:Ҥ$v-!N y 7 0u2Kǯ/$zޢwmx/6hY Z٠eA '&Hy*9soв[JuK+ qRڸ2m!Wfvs~?m`G1/_hDNzd4m_"/>~r!v_!Mc@4??^PG=C`@|sw_Qk1hmݽ{s`y<~&@~o!j\?+/>ѻc>c)'>6Y?=Lǃ)w5/O#l~|`|$>I?{){qa>5 ɯfgtˁy S ԛ~K9m5&Oq7))wMTBa {6?=g}~lE+ݛXt{8v,jnx, 1-Pg[vdk> +4|a$K65;pvybX*ڜ R8SJwD&hS KduhFK2Qr9q R\8H>%OKyװTbi ڊAkPs;-'](q1#%IG֏!\fx5$/Ս YTMiԒ tI |&-~MM>g-  <ċ0G40.;[PAa f1M Kh*e4$(p5~gb1$D!X3n #Lg|C&&(1W*{t`b @%Sb)j_%:H%=tnD:ˠZѧS TRmm3|fɖ쌭Q >lÚ,nN'k;)Ǟ1Ǒ֑~F_;3dT&$)aM("P;< %!BRXZ L@KAX=u?i^V4V!93LFrr>;R37g,R$ L uAg!8dGhݢ:K+v o'(L'L˂HJld %)-ϰ`Rz  iao|?8";JI ‰R!V9д+a"xtyql5OC{zR5PDž%̆mc݄hJC.+QG@PZ$l@6lmbB̡*"Y@n78(Z3 (j= QLQN ;O-F7#\"+w}*ȥeC2#jBđd)a!VУ-*(*-"wci:|W \f07"j,V2AX# du'ە<[Pl#fX57]Ґ? e v0r0'1zPBdEsidà:SmAWP zs:N:M;&s a SԚ4B`1EjA2Dsi3@fi#\+tbE,l %RI9 ,4,x'D)J6wCTP󼯮BA*1tj B1^76 2+=dSuDR[[ZDFࠄYֻ&1b]r LJ#.1ɇ g  4tlPBr|Z1fHTTvN(\Tє0gq`vǻwv˚Z/ƏkGW0|h3_؏Co4bja&a=Ì .# .tU2G,]͡=6*l`VPĀd0, /:),J肸PV24ITiy],C bq;F3]: K1:zOG3Ht_q5}U;do(/} UNu~d}g:]OU;6@dPJið.v~o;-~7"d&Rm+`+|mi,#zܥ\n{q:EjC$*#5]% _J7`gv3Iv,y%ىw0VVdfT[CxXEi&cF'T`w-,YuЬ! k sy0e5fq8@/n 큾yno;\Á %^ i:_'!7<М]v!A \ Ypf*z*@%a@z;Sn ng= 5X! r0D W)}j89 pᲹ5S0)E^g#Xyne C,TqB "@JUg!`GXC>Y\j`-q;7NWh3.n3]װnhun(]]=97B˂ {S ]\L)thE QJCtuteK3IP_m `9 b:; nJ}:8k(^.?,f8pOwh7ѿ._zyb'Md*縪q U_3R곂88&^mj/8~g}>XubL0'w[;ho4]^ m5`<)U:d(/$9<#GZ:2ж|ߏV ǃNQl@G*oZ@(lP8o~vp͟>V̠ hio톿xݒ9y|H5>k Fj9A |I9S/UCڕ!Z+!ʭrUI +|+'DBBWVپۺ=ЕGw[yĸσOӕ)WJ nӁ57.PZٖj0.&h1t;Pc ioR7~t]{=eav3bE^G?$L'e#QU_)%ϯ&yy5ÿVIkZmW:NAyXr W|)~(7ߠ$BU+x1at(#:@2q + .@1tp*s=DWHW#]At.-&= К^:]!JTMM9tp-=]J׷]]yi_ ;Y ]!\_LF-czFNڷx7fY N(Y0hw+KtخJh +,M1tp.sGO':@J?{Mt%r rU ]!ZNWRi+\[ժ 1Ί++\)th;]!JˈW^CvC)_K^o%zKֲt2hGV*kޥѓ*Kn2/FZOs|_`c4Gj1;7OvށL^*QP=EQ &+<\5W}a勂"9Yx•J^ʭD4'Zs]RCWWR jwBhYIxo BKDWHW0/K  HR+I3m=D)֕]9}++)ֽOrY"Е7s]]!tEWh}׮`++wrj'R,=/IJ߳w]=aDWظb eUR(-ҕVS]`L1tp-U}+D4ҕt9W]`.t1tpe1ս+Di !ҕvkg?Qq6fPc*yVB*&w{Ҥv$yR;$&  ױbDx@$,K.w +y)th;]!JωJ*),"gO:]ϸ=C+% \b *Y ]!ZNW7— -ǻBޗBWVޯ]!JN `ˡ+ ݓw']}EtLxn7f߯F Rnh;]JzQw+OtخL)) +T1tp_'+D,ҕdDWXCWw/ p;T]$])ܖwB ?X+rukLJE14 pe45ۡD/CB<*KdN)ܥƢP:jvg9)lZRrZϦu_]w"36e6C&-D(5 f&%c)hZi" =jK #kM_ `5+m7E<]跷o߂g؀O.Fur]UyytFtx g`''}y?ۻ[/@a '6O4`~?[4țk;_{+D>}b7c[V޶-۔gj:<9 4U=vIM\Y&gY]UmWSc1qK3QwT}2O_4NGoۗǟGPoz5Ny17 sQ˻;:;G@n>rFpJc2Z`t6:7K$8dnTtf<8F@c֍[ADa>-SH2;w>\ͦ<$̮ :?zl(< $DcMy pJquv>Ϋz`yqd>|:JZ?Ca8&m$ߺյ{Jl:_ͅpIPV~uN6h3:s]ֶvimEke LܵMJM]3\3ΥF3EQekBNFε T;_3.2˅c.ąʅDXz'7CZ^uJίq10O?GWA(o #'n QQ0+!2V:lPn6B1;el$<5f BLֵiSJVdP>C0 383v=/ݙu 6ywY|cFATG^K14׍7>[k`luvf%E1/:E;c`]} Q!y#jlkYǘw['g{9we}C8>| ۤ\jgyj`+6^\g ,N{MnVqOzG& p'S'02r0TUj\ 1f~t~x" "wmHwMj Xef/ &Xnok#KIv_[%-+rVy߾;} Jtnx`dƶZ܂L"+gA;ȎyGe r"x1eOsv˽IsBIE6SZ˙Xw&Άޚ =̛7VE$XJ#v^?zۛ(xl%-*7dI! PAdFGST$x uP*gd֡N$Ɨt>rR oMTh짣"ѹܒ hX.D뤊 YDdY*@$QoPj-ViY; ?{.7$/_.l#+eOCFyi.vKb ῟]IͯWw7]Oݧ:@-ޏų[znLݢelRDq`}0CRHңY^Sl3?_'o]NfYgw10sfd݌F)2?9d&y-IuKm͈fb,;2Ôz,XVpp=MtMU y'ZmND[HiXu4.\Nh^?G7j)bRTI'T6{qF7߿|?~w?;ž7~GΟh9p&V 1i *hڛ7 VnҴ׳ߤ]e[ڽk>F&XYkxw}>)N7k&~]e^&3?_-*jzK*IRTO!v0m_N5c-GZC;DI6)<M `VD"K#DIJFkVQDws^Uϗ}cQ$@P$PLFh Z3,  p `I^^7Wg:7niC={g&;3tqjW|LJ0)U>L{?_~]4y,ӶW3ao٩tHvږfӭܯiQIݳI%spRir(\H&\dfX0I;$MFx*K81@xE*Zyk \ɜ"(:I0L9qtJ+KP̿&gWK!1rr\%-E /9JlqNPj xv$g(cDɃ͠V¡Qr'Qh|Ũw /[gYO2@ճUxKs#Z&1~#Յ/@h}ե֏vrDbП㵒JWT)]5WO97ˇ ɚwk|}9>-/XyJ~ę?]5s;pYE3糹k~| ξ\;S9Y%͕*U>@8c"rwW51$\_%j+Dl4#1zdPS.%:IPfnC!RhiB61똻w&Ά}|.ƣ/`Vqj&nםf8=fUi鹜5epS"3 <>xJ.#(Xʐ ,^f2Zpt4[Kr8y3HoSEBd F5'ߋ^m1g߬}:x?hd;80gb9ߝ ]re8^z_ /lblzslIۯ^S/uy5kwᤙ|Oޛ5SeW5OlH{ףB/f/oW/vgqh{eG *rՇQ`?LaN*sY\ Be("laפ1d:/{bmQTvqYꂯQ%߮2NgG:M#r%q Z\&Ak}$ogwzQ 7Кz7ri=:Qmz*my7~> [rW ժq_< 6O+乭[oV5q_j;;xR[GYW+ϻiAz*DqUy m6l1nfL@:9ԫ'ݪɾ7nߢՅ%@UNJV҂|⮲Aaqo:4{~pًM`-zE:% >*uHY)k]ɂDz.qk#Dx!q]6;<Pɻ*TV{#^Ϯyhd6/a0d&c(;Uu .zgo\\vϙ.7&aE<1lű# C A*A_ǒMW6T{VI4і+QѤ'"o'+܁];а@v_mP5*ke2)i%2$L̡d1g&e-Vu)L0Q[CJKA ]68ƜFd87A 6%] t&ø̡ttr?/}̭zĵX{tփʢUWxPak(a}V$\J1PȐ (W}E6rdx,X3cq#L$,+ "&B%Þ9DQduvz0v"gN (6Pۂs,ڜ:d`Av֙8}'wEr[d#hr"ؤq>!cRqX s k^ <@&pR6Qo0O?Z\"ょ#fJ6G)l4FE9*Ԋ'xҭ'YsB#WHYTP"Dsv*,Ȉ zn8LLulW:BmxeZ]qy%Yr4Ԙq#f5h.P(sydYOz=X[w;"$[\E/z/יlk|c70K-2 hq54RL:tPr~qi9Z݁7 -CV@wr>aU\_TUEZ|ã`-ɐ,̸bsI3 EV8OȣD-jyt/ {Uci]q콻3!-$]ҴŵiѰI4.>PuT(/#+a-b >AG!c Z[&oУ}ݳdHRkgCk jd[Ԛi8ڌ4g޵,"i7},/IEN⠻ZRLHّϟO(Y#TK"1 =U_UEH6tkxtLĞ̙;zԜR.kÌVtÛ}^Xa@s35Tu %\d:G9;z 8I@[;iDm2 #bЬ;I۠FfAZH#|-`'P~9Pk V'DQR+W}["Q^yWb,A)V:,*1K1EG/FDD5O Z0S`pI1]ֳCE~?G7V=x4^f&O:F>.JO)]wA.Y9cSK?/AuT6e*Z[-NXg18 #*6$Mϖ[&*p -#|cOpYxR;yx}dSyd㈎^ }+sj"c nqy1iBvZ4uE L+OA8& u0D|1;x _4"cJl,mq$U29[^ RYxTgҢ]MfJt1t"I}B" zv<^=o>u/`ŞU1A{WI$WkP-h(MZġoQ0jD"19XMʝ~Ƶ|Lio'?~{?Xᄊ+=@.@Vv>[ɪT4(ԢĚ#)Q{@Y4Bk;vT oE1X/TMl$ )A„ D L> l&=&7IF  H?mfuGu0jR_X񠬅 1*Xr\:/L. Y]IdуGaO#3\Q QxTBG2ZIR-۬ 0L&bsHGӪ>31ʚ襞헨dPTgnp҈,3 pfR+#ޅ2+h*Fw֖yN X;lZ(*:GZj:+3.fTd\`WJj]6ţ"GƎj>](=!iu`,Ndu"o4qgY^ZuK4k}U-.+J?O~#*QN#+kLmfiu""WzYo9`9iNk_җCvwD>Óƅq wX!%bvy׽S,f}_<4AjAXOﹺpR7|+.Nbu?'{[X?86a1Po5`s> 뇪#'ߒRďݬ)'Zw_%6C1?`eu6g?K9 G [ݿyq3/S^~=^y{ }aMm||tm:2{J=_˾d}P.;r,&c15_uv%1+(+(2kƳ6Jr nN'jA{x[Mҧ>zgEƒU\ޭnIcM;@#j3[돦P%7ci3:ym*ʎm`!g ʶLԗvX*%_ ~Wk2zkVڟY.G=: tOsDF_wb=,9o?ħXSa f^ ЦS3_׸|%VJ(dW'iC,֔ ȃF䡅8 IR),E, Sdqy~zZ,O+",x '9x}Zu߾a2a=+#VV'يŒjD*[( cF~L3m&y.1y-dFҰL"2pdoĔ dlTτ]F:_*ƃ!jjDX#Q#њ2i,c p!QB_+NK lת|&gKl$NEK!,S"-2ΨRLby,Y4yDaFlFvQ'֋u%Cu6}bhGmmQ@퉭# C&I+]"N )_sŃ[qǡð>|ղdWGn11Aa˨Apc+E?',4~\Ʋ釻v'=Z0ˈ3W,] e4H" sYsTxBguFi(;ez >|l / 4c^6pLtEHO$C*1S1')u[G 5CJpN\\3:&E}fRLGDn c PJQhRro%PJv;} c?kKFS'O$:8'sX}"8}˘QKb]rSYyjozt_W̷NTX1kyZ?O65{)ZZK|N|j{num[A  ,|(%:Օ0!?{Fp^8@ǼT>ΜY``1y Ė=\~[-Yvܲ,Svq*b}d+Ư4!~(zf-jvyb7^urZWߖ䵾=89;X=Mئux{Ń)e;ӟh8*5B~?D]ɢyfu.}OG o׃%/Y>-fVګ]("E8#7[Mq iAUwkOt׍F!f]Uޑ4;m}ࢪˉ_9=^lx8eW6|Ⱦ^{*.WDZZH[PDʓ#~;/V TN{!l:[= ӳ#Rǿw~|,?~.̇ûU/4)o <_6w]_zQSF߲ki`SwM#hzNjem H_>}a>uyڞVk&i I\~Mf~ 'q[7*IV 'u!"€: Ɛ6oM߯Лjʁ)i~?,"F圖Nd3bdY &#B, (lǃ*;RM'v撬n6|mȮ;Ofvs2Ӎ.IMUcӓ05snRn}<Ӕd1L(8hQ M%q$r%kkk5oS.=OV 01AK`N #r%_vpRvfCV* "x4 $ U)F>tX/`w:Sv>9_ZWj|is#2_[fGLHlڣEPN멷\J'm# G wv4*Eo/v/A4TUz+^蒠cqř^B`Y,K,Ap![]ij6? S֜m]~474 rzAECj>i,ZoO9!t&<@jP%+b7Mw-BZP [9_ꙷY\3Nz7@Cy7\MXMsZ|m1[='!lۈ͟`Z-|8-p?奎P1 7mGuXcNP*w5C9!bmZu+?$-Y6m3KfU]eC>øF5@K 4Z,H2LEIr4ȦsYFCx+֖3 br[?oGir]@qlOӲ 4\P(~5!BЛi[G}߳ϳFeո]_װ{c7-wL`S7>x{F}}Y|Sak!^E+^ rG>b7m뙈[cDbx:fpu\c//D`.s"[٤Ѐظmci1>:?Q;:у8 de$ 49&;m1Gw* U h1!c]u+V_NM1M*s/$Fg}iTF9pJNQ\J}!Xsܠ$X{{s˱"Lݏ-t^B/dIyGzs< bxU lau嬣h$f`V 4aNU)bx`7 kȝIc@c @ЃWeC dw}~ KQ+^V2'T,콗NE38+8z̶TR>yMݼuH0 !e .Sm% UHMdBH2j2ܻМ~:$ge֐i7ۖeXR 4j;w)H,'ۅ J""ɠ"2`)rW5˧:.{ƼNc9L_p8 5 Qו^,-D!ue,.Q<+ ՜WWc84Llʲh݈T>0)أA]mnx~UI_¼<]FuU! -`Nzr@١oG^(Gk9D+1[NOiy@K *z5vOy0 1[Ў @uC A,Al4f+WkzE%& >❃͞#KʞJttZ2Cz&%x @+|̑ls{6fMRpBOK[L{=O!* T$)x˘Bs<8%\+sw\U|萊*h}V=Tzy(ޫHRq$uv E)`8 ,t\NAbbAh, `k=e(tAi a":f7BW2Y`,5{|eTJEڙcI-Prih^(DaIZ[&oA90MJ(dvVm8[*zmPTz{((Sr9pC+*j$$R oyٹ <._wz ?/1Yh\p<A*.礑c4)0Af- OYcPj ZxI#NQrEm]H> 8:ma6k4Ck9s&AF$88a/moL +.0Α+3jFIEV\fGzЌVoGVʡVC1j*Ǘݷ0CKe;艃|%Qsca\**[1:jiȮ!*[xea @4ms (%e0&zv7÷;rlԞ<iƥB%4>ԀL :9dIcJ(஝vop{|_]ԏ딸ZA9< K YuMjSspRJZwٺ#ky.JCqxiS r.mNcl?kX;; 0fL,x9ׁb`gIJ,BIJ9`.쮥LFCa2&F+9#(1Q{VYˑA G?z=}栿zJ/ɇ-Vq=)t=#*#;n<$EJ59PmS:@31^ 1lt<Q]O=B\ eDHh1-wVKCL?<90)3XfD چYh=z=mC Խn 5,AVj8b]ۙ w9 b=,sU#PXQPGY'˳u4]nR+r] B%DQd[!#-L9L6A22Pߤݤ.t<9rx:ᖽK6/.#k5my:9t{zj 7fiJh~Oj&]0~kd}{$ .v($ eR4|!zdn5Gj~B!o1RV}xX.n4=OK}eMX\}rŶXŔf.NϿW.PQggm_ֽ'H\NAhm'-^jYܬ}𺃰vx(מ{cR|.&.şOmyk[[׊iMl_͏%ask2k}%+Ȇkb;&Em )7[ȵE:ݿ=6ͧt%݁d #@z;SI pv>Lvd5lZd4=Ե[WFmvI,|&fފ( }4w";u74 uD=qd=WT":Rԁo1AV0u\i6H4Xј:lYc^^nN 1:0g&dQ;!PFEQjDp<| 9nHϹ~٫=gsr"lʽ﨎wN2Zx;7vNAJ2]9[kM㔣2(@tl2{ f8cw˫(β™Rػ6W MOd+|^_k9os~,YrdDzeu -KEVSkT6@J &>&v..F8o/BRgy*)q1mpuZ\ "vF%9Lf tP7s{\Ώ]F(7"BŻ;h_ @X iU b6& zʀN>5>a 8P4Y1;%Uƾ2g!))ꢒ`3 ; Xo<$qUڥQqxjD2CK^2)*4אv%Ƶ %+6,EyY-)k'hXVnWsf_ӰbL!2%*j~(ˍ ))Y|!hIk^\E )cW%"k?L :v P?;:∩!G],h8f,܅Y=QznflDLc`5DRAU+RR0:eMY)JDTuW`o96gÁOO7EsEOVOZX"z ' 1+5\3i1aw}uyHّdUSLra'o0^/|yq*x {IpR0|RM@g#,\NGHefȥZ &}X4A T/A+JШ o;nK~7t%;FV%]5k$V+LWMiYfjKjB =]5++LWi7t Q1̵f,Y;bYՔ^aB+ng.\[=uenOws۽8GUpoJttuq},wL>kӺHٷ*ZƮz6m?^/ThsrGy۰ 7|tq^;L\t Aﮀ[""p5(-2';j-s~:?.^NO] m>ob-;fZ]Z=V?" /mG82Yrw Vy=ЈU3؏']x^;fqCOWJ)]teQ1 ]p4骙p,ʢVjg^J_cr`t%IWb.t5ka`t*ӕ|+[Ag'sِɢ2\͚[Aw|cȳ7mw[Cb-C(5˛Aޘvsj% bkoeI?Mrh7yuxZ??mۧ\ʘ/ǻI[tw8xyMɼ&{|Y/ˢvomor:-\yx;?X鎞% U*?\9!w[ ռkзߝZ-=ihYcguPh;欻&j IOZoz@MOZk /Vb)iE(K'׊1׬|5ss`Ǜ(k0S@.d뒲 'FX8z1Z(΁F`z˜eˡCM^Dd}9b$8i/kTcۢC* \ RQɧ|"R:W I $SrIY7Eno?R =y8A9F1@/rL!1-TitZ)Y_张jg{őӢN.yFsT\*x9z3g8/ ̖턏8x0x%Ԥ>y9|`* .^ R5묑jw->_$ *s41 Z#MKB 3KulRVVoP+]MuT'=[J,V MqͳSDg*X{Y;h "HA;2S%qߞ'tةx~弉nM[ѫi2mb AWڐ@[]V%*)6jBU9%)>^S'l(כl|[n,[j ݨYQ%TUչ4-F=-DG_mtüMЏp&tEnnoU{qlHMlqL$`4s=c>`4xf%L:I!Ur}RZ!9-:_UmCQއAttNjսHn5iNCqIzsOv3#?PJPz-9yxρ"Y}`.ydԥYG@XZ5N)N XK\z36GQśx5綶z*%bI|3B\3"U@ J*q֞1A!KNhRRAs%TȒ:ت&iEsVRoqU1bo<*]^]|:-ꝸ|F7KYqmX=F-\W>(u-3&l ĉ^ ,KW6l Ikkl] ڬG6^ p`H `$FvQԀ: y:oQY TE 'oʓh}:/i͢:t_/GЪV,Zue]TzsGϪu# ߝ+MJ+&Y=[uP:Wd췋Qޗu쭺}B ո)_\j;]ǻ}V}q# 2r&JZ*<`SFs S܇?IFS5xQsf=h0%Wփ뽄'yyXRd: FwXKITHXqbIqs,aIEXRhO9UkjX@S-یCV%O+k}L\]rqўT9VI9pЇlx#znYNr;G&3y7sF_ ˋuUp@@oDG eq0-h.TYC K_ @X MUdY6& 6hcTtaD cͥ 8RkpoٰL2Hq.*٦)! :+ $wxHkKU␞+*ܞQC¥2)*4אv%ƵTJ[aWCbjIY==@òr븚;F״ycǘ ) PGU FYnTHI1 |@K^C.!N16* X{պGOqϬC[m?͠i;@ہD@G1y\>h\X;;Y਋C-2L $IT=m&Qz7C0 `r z)[&,G`V*+7p>##:Ӧ佶 ~yP((\'tt t 5UDL]_]|v$Y4v!yG%f# Ӽ ƺr//Ne/V2_MWO_HldtWk RRlFDIգDnYK0tJfR"{b/)SCgKb-E*B]k,(L~}{m%ܡ>75}o=52tTѵf,TKqCAP% e.O3AB'?{O_Y>s6AZØV#K$Nrak%^;jYg\Z!;z0U5e+{`p"@(!R+#C>0A8pẠiI Fs=g/y>[H1V)56 V,ND'cT %O[)NzϮ##zAps*fet3d7G55ɣv;. <ц␔MHH)j۲BǴ7@%} !t4{&_5̀&01/SA+X$INGg@U(Q]؋uIHd(:񎈀\fLsn FaE蠢r!o8bWgW'os2kt~wWaQU'p1xgR_xRh :QN7@$e[@M bT[5ё+иp.bicmx1ş>.o?Ѵ`Mp\W,̋WW--ph4n2]q\pX'aUs 2gJK)\^oV 9zՋ=NY5diXf`_Gpj> o_nIoMϯئg)ı,}-MF͑5n{r$y4ʲ\~i&HSWdYI*Tv\G/Iӗ7ikQD[a:O2WNӮ?>q8^[3jrbyM~ONt:*hiHGr64BԻmX#kc5k2B\q?aGy6F(U})l]4bmW!k0rС[riMQ]mV%n TL޳IAVa\ #g. T; 7v nD~{۾'Z>?D0pSK0lXMd[6o B*2dSR!JvcOm)H@o@9 1hT ,20Py*9'aO *D.0)Aw~[%05|Ӱ"6nZ w^Vp&kn]{aF L"~?e#7_[*Z1{T0\`xl Ȓ}Ѻ׬.Y RC&9|u&TD{A#1B*ji, ˈ|%6a X(%o1z%d3:my- e {f)E ޒȩCR8vh  J(Rxs٥@ZHE")9IEdx\afXTqUʼ 0B'Ll6wcYl;Zd|if݊ (8 zI4y%x͝79FbPi!!z}PfO_v|XxP9$a!k9C.1DR+v1]JP.# jA<fAH"60.2D"g\ IzrptUȵ:3ϥaipfݞ"%9{.O,L"|/FdI]`ָS0N>E@U0УdW)́B#L93)(8\xb t^@T|`aKӠIPTX֚ ]wZ9MtTdb,Nf"-Sf _p Mw4_%XŢ;bT'g#!)]Ձۦ :#<(30:R0y-ͨꝕݓetr^]8ZTή##)sd.̇n?gR=_GZli{!4b#1zaH0aIЇQie݋d⍣~zM6U LgӫPsHq<%1H%!>٠|7qM%\*&~/uess yy|g?^y~)&ǣO`$TѢ !p56Gl?ty֧azh6C6g]-660ul> Z0)Ó5۟l$>-6E;\bT@|3܅/DH:wp;5dcCW]w=՚NoP'*u۔)Yc%YG*Ac#N)E VZ|EێG^^#!ȹ ;L0Kv_Bzb;;qM2⊢v#*xv;;D+N1L+} ӗaRZQ|oeʫl1M3*g OAs5t @[ǩܥ0l/0x@v{( E"h$Ӕ!"T^NEVA&nyd3'upL8O5wxh1y?d =*@fFBkNR+LƔ (F+c{my+k݉;fS8d,vyhߴ08XyI]*B7m7xK Dbܖ$y{Ɇe߂\Is$7qhdJsOSَWL\1Nڗ¥}ςcdAQ Q{) K=*idޑ@F+ꈔsL(&m EέRȰ, Ƹ[R`rZ+?B"[Cg:_л^[L/v֤Q)58I0zg:uj^W>*R1}Ŵ͇\=>ׂzIXtXEkIurzIV ^'zvN;^P/}PKYƨAi&2lU Nt1Y$pm x$ )[X1c2b=6M%e5t *eנX'"!L=iF8ܜʹ13Zp} [:L I ]x{ЃᤀY%\y^Ar-je ]>4Iڜ2-e4࿋I[Nzenq5ݸ`-5m$=!/\O` us؊A?nW; ބ`K053Ҭ_w69ֶo۷D_:ߟmN}| %&7arMRtqom\cIb%Iֳh,񆁩*(rR}D{Vmwˆ^N'xd!R&R/5e)A8тFQ4`pHnvo -m5..CsgQL6niVhYwL'eٞh0 e^*ƹL2r@!.E; Z*SV9iV"h@QA!M6f*$' 0UtȴWY_'=6qkko9d],e9"3˲ߝ<-Ud^J|`TT-VY 2H7<*~?~TvN]?<7iHg>”{G0W|tBOmY3 ˪s T߳(`Flp!|VlKLǟ cQ4PUpp 6<\Vqhuٖk[<[lb62 PI:YUOqY<ީ+g~=@lX4=?.oȎ~}zZf̊fj j. tB2PU=|GO=_oJm^W9@LEkp@8.{7J)gNFxG__`PMe`+'PJDP ;JJG z+kؕǮ4]1}o3Ri"GmtH( u(ΔJpلB*&))-g' qfI?NbXXBC"Zf̨,Q0 }Z]3sŏ<pv\*o~B#NT`Ԋ+2ed ,LJFd>Z7{G crKuzd ^`A ƜBm:%c{Jk5ʸ,;R]r/ Ka͓iҌ?[Fb/Ǯzɻz{+؆cGǦ|0ƣ@Y#s;`D4H­A8܌L1=(B2'ZK(Ib yItD)R:х՚w?:%+R5-R{-=%ٻF$+D23*=~X`gXe14I݋,Jl^JTR"2`[UDEFYOydHKa *}?+΋Z* e0ⳔN ʑ!#2YׄHdi/8tX#ʖNa)&~XEAOKz)8#Qu;KrRTI$cRp%C$y- F\h$sԑ "eєPjT,e?52J jI#(GKW./&N~ڡ_.RMa;ŝh.뷎Ȏބ(9Rr?+10Rq*GpakuwW\dmR%kdmd̬n:z*.i4U J0;%BܛF[㗅0 MOn}%"WIH*[5qj~x?9 7+zzn]d2:Nao盫 JRj%* 5Q҂E\mc7v޸ݷsߝqf, ]yι"h81TL6SQWi8O S}*^m揺l-(mAQ/S{Ovr7~]kՀenoлnx|i;3YŁ(+5"ƭBUo[7`'׀!g.gw׿}hnWl6YY+| ċ/wf>/5.Qq HA{[$PB 2L,gGhk1[?{j7_~L=;lּm)W>ns]\~zh9rzv]q+V/hLSZR0 B21)#4a"lr1^gաo$g'A:jiSR}j?O-u8 T3X]%%7N"m J tz<Ѡ<3S!,McCe+ Ia:n;`+eC&zVT縓.@L2)U!JpKԢh R+|<>H:DvØbJq+U6 ԗn,YL*GZ^&,Zՙ{A#BK?[XBhYs85u:,Gܧ1[Ŭ3m * ϛFm} FJ%J6 E$ 2^BCb* 4dk,LX{dB3*푳p9ŒV.%*1w F Ј[G17ǧ҂N؂H= Gmi@tE@f'T@5-AR$2MDmQϼНl'w[BOOȡG _ /K*# g#ĻN:PhwQAXʏ_O:Jiȡ [{AӼz;t%U}I5dvxD~]_T:etk J $r"!K6yd.K%AJ};ShFV9< I xHwM:& BAAC+TJqD+?BQj䪁VDhRqXVb@_:;k'-g|ꔫ4<ٮCe))M] #M92G0N$C  J Rax,NB WV|%ߗͺ S޿oVs.x/\OX*[ag!19$h1hɋ73C8rƷ *hpE*@ AN>6\3C}(` 8<,YN/M􊉳߈k9sL[qў\ Y|O5l[}Rsv]|ZمۼyJJWVZG^u5;8jOf]͎(,THa,I-Oe8@cHNcƢ(Փd"<EN@s+jM8,hac`0J-&~@Ѡ6)P#/8eAq` QDR@HMh,՞& Fa rwN ݔ΃H1RHp88m[@J].+m, άŖT@#TrԳ^tsJ4ˇʥ:EQĢJXaO$y;t_!N; ۲?p9رZ5S90(ɽO1A3m#R\DyJǺ FX1֛);1fNN%yz{?3Z}+W,n!h)U ޕ5q+2 % -:r*JUp(;60EGMt%56z0C{h{/ HB*9,A "<"$Y޾c D"$5XD[-FLGϸrJhtYCV#|a}ZL} ;[Xi"Zhq;zGpSwmp*" !R+#C>0A8pẠ<zv0 (#.pmr\cb1u""%w e*ǛW33*kt\axV'07) rj #r &uuVQ'Ho7o@>pf"kQmf AhWTظp.^c)ꃯӇk4m7 RՎ|MDW{Mt5yYƐ% MG˼ѕ b:s_ o7" թmDk]Jw& \5Aq`^q"M@L }O#G[Wv5L%M 7Rڧ'زYT%<έY9Ξ5ϞƝuMF'o>̄_<.Tud4Q~ *kӲ;0w%q\fGӃϮiUFo> ο07 3~*~Onmh/? nZųxz"e-Sţ2v1g8:^=z;k6 NJ.d%~ڏ? {(+>}4tKq&=*H9HE \e2/]1]ǹ|>}{$=2AZx1-냤 "Us%)A8 Ф\Ïuu_?yl&U?dQ|9_.ZXv? 6gO6& 0$y<& ЛLuRh]}; bzg肔GC9 S͆2%H&Rхg0z> +X:ЯF9߆{{+.{HORh?z1^{mޟ!@hgCYW2#)}k+;Q qͼL ~xz1ɲ'AEorne?ytܠm A]-[b U{#ʺpF>n .pL?eGorF` VUrU98B]HZF[w~<ޢVo0B(͂AQEO*@cjbT?Z=-e x$ )~3jXLhe4zl5ͭ YsMPoxfYVvSJ5MmPA)%N&̆0%>P„&`CR+,QCRV$J1,,АHq݊r>Xgs4Һ'1/nMNfz2l8ةZ[~}LN\ R3n޼i'$Y/w28VXiQ(#cd8Ї`dG[."Gk&#> N9ťoW ,3FY\lְ5 ŖDwX~Fpaq$+B'iNs^!_,Wr[9%ۂ|csWa-ϑwFeo8yJIk$8R'DY,GOx{ 9<|a,K,8m=asUif[{p;: K6c\@4i>d#}\F=h#٢,FVcd!wJTDiaMFFy]5eૂQksKbŘE: tA2 ǨU1kQRk sZ~k۶[cg9nцW6ET-"qsR`IayvndZ4-cױ}~φrIgNrL@%Sp&xǽHk(lB(,!0ĂR:J*i9%Y <еY|^BhF*#R"%1H̰4NthD0AP4\ڻ$nsY@pSYjkB=Ӡ1DQ& H ;5#$DOy*/35HcT=  aJA $'ITR)SviXhI: -Ra"8KX ¸4q j'6B kGCj.I/Fvn9ckgXjK ?\؋^%j%9..)dbaZz$0 %XYSᬺ:e{Y ÇV;8&Q=UXLKF1 $id\=#<I抳39YRUXMWn]TSDa7՝ƣY}ɲqvI9^030;=F.H@ІU33m? BM=){4uC5v,*G> G`Ə@>{7JgyeZ.3 A|~M\"}Qz">.Fz KD"V 8ˣgO>~=:'`N<~rr:~ ߛ7c>cp_~Uq6f]S隷ӯl>-7;=7YQ hJ /gˢy&̟ϙѕz ? j>rZq5ks>CH*Nc*יjK|Dv$yq"P'M3E"+Y#-j3T]N)E*(;?Kf6ض|~x1;M?C;!(hg@ %Tw:lUdbü+ͻ(֧}{ѵNG+̪]6R`}SwVW;'eH剜рmG!: tFD§ߕ\[v]%4(hs'kcځ9F`. ,BF2M.MY$ Q+JA`h_tK{5g\8ݠK'x%,`FӌTQ+J5'\)VcJ (F+c;mEykߠSwP?M@5UYn `pϬb[s}o9[MіA~!)O>,c-x1`*a(Ic9e: 1qhdJ@vot\l:ba#^w,=FHE`9 iϼxm%-;(H`%C]r ܤզQ( Ȣ`;p,Z).'RHDc6bkm#Bw 6y6-VA-g9픛=`:uej4>v&W31Ua$ I5%;SvF@>"_N_?ͭ+ܔG"Ik&bgu* ~1 c/8_<<,Fl܏KnQ^dYꥰg=OZbgn~^~/sZG i~l0* ^5V+pAOr77 26hxn񪺔}X[_u#Y`.v_,6NMN"}Q$$ˆTIsHjJ`Օ 0aaM%%Sȕ*2 0EZ锗7XFi_{0?"k55lI&E8ʰUr+8QXǰfM d2J5IJrk]OJL# lqk5?ujz`#DԔbT2B [J.{ks%G$1ZϢIuy)'`GD'A/jF,BtH# be}0z)#"b1h#2&"v֝m5o_>6ܯu݉] mGfh;Xz&6 81㕭!Qb˔+G4qGe1 1x*N !Xhm 4O= F虊`) B]?)[wKZ8H޳rd9)d]ZOdYREy—zav³.EJ-c13!2<7*tie$<-!V8ɚӅWi f\t'UZkv72iYFsCg'&&VєnK!zƌƁ#@#ؕ (4J%N!|#oz̥lu$w}w"&=_iy 9bҢykUX~&}}?7H}_@`ѻ _Q{ӝ,Ѧa{טǨ˥gӇw}} QC7r~Es9g8e-'i[< B'z-*[Nnߜ)dsCeLQ^uI-oϽj@n8yD(v "#Kl"CQ8h@:j0Ε) &y+pFCJ90[EĖP=J ÌМ+pv_cɻ(.0dw 70Φ01'9Dȏ2RlX.St3'«)s\hpde\K:5UrÜ%Q [Zca^YB%#x\?xAT#m a]h&`ǀ@{0^hmN81>oxG5{¼xjsN]bs&.>Rv`.,ڮOdJ=Tٷ-u R_dTc XZ*YɄ1(I,XZn>0)G3U3^b8l#r^.`uV;wC2:(dּpu?;+E^+6{L,c05%a𨢚+&`JV;|>^I9;9ޘŨQ;h[L(s7l2wCwɡsa {n%ä f&q4&_N`3eK *T2Hn }΅_}'|D\p="z?jmϥ4>2\IJx#u,}ԺS<V:$ႉh m!M$i2#$ i~ugtsR^K7ƢS]Wš[ JFW#הvv!ؐD7Nz%Z/J.{%XД-SF8I%YHd1DT9'l[@Y nU4H*&53I7HQVP={aKuL(^rfP4TJF8`%B"D)5n9C*,[wvJ>1upQc."S' */K\0,j!8:lcƣ14"S_ ׊z=ORPM1hLu,MAr̽ hD%w`0hV6&Nj~/N^]  lZI-'V 5`Mu D,9iG1F,KѺDu渤am!9&z.4„3!aGܤ "NSIʴC&">s\Җ#}2EÇec}g>\ .vh0bggJlDjr[9e/"!v1(s062yґCY:ȷՙ^sIhRVM#O?IW]1W/7 Hhc&6 +;`V.5g}⮻;΃(¤aB`RGTRJmɌ6>৉d{E XfSϾR@q=*[cR;ĬW Y@S mnS` 1ry4 xo7: 9i=b#H6sx2cѰ3.%lH)EԒ)E 0&{!kK.h狲2 U- VQ+8jLMᨆTO#XyUo'&驔jRV@Ke%):SNď?ak%WЮCK{Ns?g,Y&^ \ .8#0R(*T1J)`Pqm:N(Sm.L>z]9䴚=egVO9[oeh !m(muB4d&C'?;0yj`&01 FE,$Kt(dA㨍.jj|#B %C1$wDl`4hB9ٵ z:ԟE)˜. +0!E7Ҕ) WҵWÅKN|wUL}#i e=T$>i~= ?b#sa* #)]pcoä +!vCo:mw.ut+~9M*T:ՏL?{ϻix%5WQL;9{FMpUg[ʼnʒT%5Pk$t}uF/eEo]4ܪvww/ R-X#k_Mi ;Jγ-ҥmp9mS^qi&f3wvt.УXz+uhop#1pԽzn]e(5LPj;{K8m!mn[SLއMԸ"=sE3qŠĽy'S:ÏOP̜j3mklJ9`rXMd[6o B*vwA2)\jYmG^^|j f߀9Qc*<#$Y.A !gj{ȑ_emd5@p^3XL_v`b-y%;3[/~e)2m+I#[jU|")W".A23Fe!"VaJWG`.o9o66vԒs -]4ƅ{X@do*NàD9rXY|FuXt49ELm7:Ol&g%յ±.Rz Q!Rֆs$aIr~,J@y" N%CD]2Ff$8DK1TmffXVNmUzׂ6_DՀrYFc,_q: =  Jfg9, $ڔٔxKN*viH>d>H$2CÄ.HpuVl))}Ei-sx~ d<% !.z=)嚘Wu/&|գ֡#4(1aҁEa|݁Fi@L1yRW3'ۥJ'`i if=\p># -D2"MIcC@RRJQʈ'ɐCj]jcM?lˤ}(`]*",'CR*xVl+:d26l(D˳y!v.N֟;'|睼?/e]o ذvb?r5dۣ<::gΧ<F~q4[8p {33y~08OWN'8.OG `o^jts2wr~̇ow5K Bȝ/@y`kjcw<6-, lZM\PQ2h^NѶW O/6k͟-5,׎oq 2bsi^k2^eK/f/=9=믹GK7J0N#uim@Me3_.ޭXgwm7?vطW/{W~r˓[HL.|̻t2zˣ~1h{eK= cd=W\m0V_},/Gn GYD_܆Tֵۉe/"xZB' vl@r=$*%.1)UX) "@2UHԔ+|iS4>!*ǪUQ+Hz6 c1wofΖ}bN?.gݼȻDuy9?{7?g+](ym OZJ )&MFYɨ )mGKAJB@GOT oH–BQ:9@vElʜ|+ n_s96Sb`O,qNhqz: %%u {MIvS71 ~ ž?;R2Z4,|+O3a"AJb*vL09sS}X b𚐽/ұ?ǖ|8"9"1T‹d<{gADKPL$oG{>~ Je(!)RTN$lﵴpp֏2gR_eDoh؇G+})7Rm-ǞE5 ّ켔DIUL.ѕ4ֻXRĪMҔ}a 5YO٘4:R#c3soUf#cXh8 /Zsv^B[nm]bΟGKi]_' Gq4t1:}(YoWxZ!N|JU݊Iݘ)wuWJF_Ghqh~zM}-xu}j#/9wpٳW$~]/ZzmKɺffk:Tް"'}Xgbѳ6N9VMku+ai!a/i>q* ~.h2OEM㢖/v4w<9?|wJ_7YOgyb8Xg56p>zM;ժino4xMӦ]/:|v][^(fށ~twch6D\cW=bɬ8Glvy(/.[VyɃoeU>>xFuk?U*?W9Ɲ:,bNII` jm*b*0WJbY)CrWާ¸dxc'!>vEAJŘ2L Y B~,aXi3}lL[m2v]:_ts <+6V{ڗwӏJ-d꼰 *vV ~PziKkCBX(̀JiJt:!2.@dv&IqgDDT 0^5Q"$TZƬäflIs5UF#[JN>yoc'':@SBǹr;ϛLPH^HS )eTdJL3eb2W`NǮ]|bNc 2-NhXM ;?2֧k.ߎfk%:@H7=m<@5&[w"vZ].{x4w#ww37̽w/Y.ep,ߥp8!BgY1/TtW xd󯈊E[)(< /هN:+hsꗕ;5gU /j~='xh2֓NoPzَ%V7.lȍh 5svm_=<hC^L0058P\U$ Z 0&}?-E5, O V¬ֻ. SN(>JY|a?1f mXm^P_nȨ J*Jf*)t3?1\?&u x~&Ǣ2lTJکg4ly4*+cQW݆+M!w9+U&15Dj)՛2t9+0 Z%:oFq8GUT/bt zuI7:+O5\-Oٱ?"2x:dpt%[J-C"(Je"h'ErTrQ@jINg7[/M"@td>Tk-]u6Yr^0_)U~~?Y[J=l3| FrЊ_^z:.AE qQc)x%8Jiq}"C0{[ܞ p[ktxK%OV2K-EPZe񧏒{ŕP+Zy{ \-Yt&WRp.DJ0 @h%JAki iDO VP9 jõqP(I!Fs=g.$ m 9'ΓG?ЂQ䙧p,c9.B;rQX҂Q"< 4?A:I Iī{Q4lǬ;ccG6 *EOTI/A$+И:: ". t6qjl!rtLCEW /@tC8O&Jod  I l({ڲ(%hk rgPmadh F=?$WGOJI6}T'݊߳Y3\ ksbڸI';Nju +h[ TN% 42'[h[y3V-nK˳XPhj1ZAsRr!!9"9)E-:QFf'=OBvULE I@.ΨeVJmB`j|B!# (%mU@  ^9-!y3p 8i:H&.-\3jY@Z#[nPRLhϘ<(bbRL4y*9Z) 92>e]s>X@|"20ZK,1@w !S(p t"T)EAxv_NcZ ^f@;6׶ V',mp{[Gxuɍ%؝J *o|),J/4D#K4l锐dԜ.jW]aJ.g[Fud'CL$b8iKTDi\I5bVAMDsrp %&j⻨o4x#EuZ5sLT4',a5sBLdJ8~Lpr^k/n(ɱCG\]Wo#/l|R*^xH2AKԢL,RpJ(=CdIOlPKqƣ>XP JJݬD9>{.NWge=J;H@ uHu}wN7A{*#7H .fܸ_oj$`JxSB+)؈ĵRC4Q#&ȱsĴ~s<(%8xI)mA.JŬ`YF (ODLˮȹaY_'ϧդI܍sq`8]0> ٙ{ _M6>ͻ#jLBׇ@ V$qmG<28+ *( "9#i{6~:)I p AyÌsQ"L BZ* hQƨdrU@H\PPLpre(w"YZZV sUI\.xJ:֔κ[\Mٲ5<^<2N!NTDD,qT:Ȝ?J{񋰠-ٕ,t%3Cz֪N<=[( V^%*0!X6U:"Ny2\~~^8vXsqGlԧ˼LQYl&{;a_e]9:.]\RFaoٴר/ǿ"_r8關DlDONoG3΋0o!9i~I9?apzaߏq 5ЗW`Ks_^[tI6z[lS+/s4 !tg؏? vJuF_~]2tCqH5'ՏT3E~QΗv?'ʡ9 j_LHV~dF=JY sV9O: fSaCasėƽ&TCV7#v` 'džwa>su<..rּky׻hUT0Uּy}۟)*oI۲HA^Ow/rx4 0[SUe<1xqcxx3$~ԣ\sBWާ,UwX}hVCXI!#οq{W]GGP|Y|$:8] ˣGK怠 I8IK0Gӭ9aКSҮ}bI7؍3[. ;fw;>?#n^EuWIm &;ηUl[MOQKr6{)?3E̹$4vj5vqyvgսc>϶'3x?nIBM\pf5gEk:6fC5"{zō~U;|V6,nf3'Bʝ'Qw}!dԞⷀcܣJMH.ICjū?ݰ%xzK- fK`L'`%[8Qs:thӒ\K<@("v^@iUȌo23~62vYe\j36n>g(U/mmn7ܶͷ ߶}x4;Gڧi ]8A 6%8j( DZ$рPp68!VͩPr+Zl)A퓔#0*76hG R1YyyG#4*p@Mo9C! F'S69k5)X [#熻TٛX $|UN|9N"zA{*e8SlUfa|)Vjk|)V:MNck6aVY,] lUO=euZ^GJctT`i$Sy N-VS9?{WǑdJC_f23 ,aẊ]A#BE6i6)YZꃇ؇dK.CŮʣ""ߋ`Dgo\TX\|;IxNQZt6&2)JYi)sXj6lu8w?$2ȷmugNrbn&b`rgPb3jQQ$ ]+fSQaam bNH ̢3c[SPdPl8b*dL6z{ؓyE5{J>EtE,D"225uɘ%f` jY90U%RLdZ'"{"- 2uӊI^E+ϙ4NjF1uf٠Le=q<~||ͤd[E?.F&|{dv!P$I%PQ sCRn{Ü[IǾ-95=կk_6W8 7WE?N6~FU#<zZEУ-.+2"(ː@]e4$PQHjL- .JhOg@'I&~p$ l8K:HMJd`(OC)A[Px{J5m֝ tN%=;?=9/)<"wK6mףMuzٞ0w O,ZJI dlX)N?D-4zȁ\Xssd?f n$A G`KV΢rer}yg/u69Z5#+Z0U?R}wE ~ʺ3 ,bRҺdrB,QddRHp8c͍6,M4mHH[r8Qb%]z{8YW3{A‚6&}qNj=s-IJ^ڒ6wtkf+VÂF Y+|'Wm]fiJnu>e.kNR6,>? [/UOSy GSgV>Ԅ,>h<m%UL1:T"cr03f%2[+ 90U a_ɾ^Yk&[1hȖ:֠k,jy%=E,:o;_7W|wK_x|dSoy'e$Z2xrz^F?2ysM¨>_}tzGܫz=z+If,⶞o4_ga1u &%kwi֛8.G|?}|m9aflϧUu6s#9o`WGJx25ϸы6XAz<{F-"jXG-zv~Qg =_=HBH;אw~ڽ?{jyHFya}t T02[#r㜊VA4t{}NAmP(F)ds1RX#0X"Rt:! `dvd bH$6an /UEU n>yy* 6E^f)̾k6=r06+$@,ZTd2Ay(l~ ˰o6_$[HƟ^~^vp<;F +us~SƓ eqdlPH{zOثb؎bm՗/1i )o ɡR@.XwI8]B \䴐m1"#us c#iiId&2gʙY'ZcfYBw7,>Չۣź뤫.ŇF^I5aj~>HӘv]69|^*MYz]p:Ep=ј h(IQ*b-2ZUH"+|v63z 2 )wWZ%(zR9mH#`$(.V VE0z@鋰YIܽYw6)OΏ:vRk]3tg^z./\l L,Cq^jr&,dT`PS ΁EAJBAGܠL=Q:>@V*ʢu1fYwo@+}eG׾Z*x7Rb`,|yS'8?25+Z?eIZS&to;4BP{(BL )Sk'ti 3}88]Uq&FT^/??cX'qَ'v&xPg@:j!NJ, 5B..KǼMB+Ă~\.c[Ӻ чC.h~5`j`4mwo/^cn|a$wE] JX*g`6O!]@ˆ/*䌹??*$ZzqM\585GGfv+ y69n7Mwqz$.bbTbt=.:,83e8Hvv - 80^;)StPIR9eQk0l1އ-,)P Oe*䐥`I&ĐQGX,Pr8]y 3^ϭdw.aMaH{z1x,[x,jMsu,ebHx5 :27>Ϗ"7s5JZ^2ltt)YغZL]'wG*)Titk}; i")͝YLf\n۶10-h#Ȥe- *)<"ꩂlK 9B$_83.W&q@aH Z?S AAiT޲*ڧ(:eRdG16z ˡ<䁖{/qb!0[2I/trZJt*TD@SJ%]Ak/4u \qsXm)YJ9:HIG@j[6z\Um>&x}9"\_Zm{ڣ[k/2RZE4f҄\YH941DYR_eaݠ4M@7\Ap^#Si2"TCJd!SѤL1NJ!`zۦzEp B %ePUP m˹<&4ֳfP_ʗZ$b,!hJB,N"NJHLs^:Y6#Р-U"3C~d%@TJR,&3Y+əldZvZ͠V8xnے1cBe@|f&"%^;*V"Q:pWV2񍔜]#Miat%a%TS&f!&D[ReP$|,?h[;+UrEOEȶ'+^@=lvupE6'11|Đ%Քzۀ˛fa0Cgs@VNV9%4kEǣI"X)Kok)>z'+lwRTgӌ*UcaMՕQ*Ae+l'ܸcr?;&BTfMbçװ-Rw[I|~7Ap>$}6hva-:\]̦0/S>[fKtA㐫f ĔuS$4,j.\\6 oZU6|i LǖdoձXR6u-;N2F=8tF|a j TIAgd*rf[*A[+IOPVI GR #l$%'%k` D78;Zun^wlhzUoԧ=}!hM]\u馱ud,\ue]'wz%j8\jh |8s)Efc*[CW .UCk骡b#+T٘~! 7(m:4YtߞN~P %l!T݀I}թP 7~x9YћeuM{"*άxFG9-*>/~Yg)l.@zb^Zۂ/FӛY}v)"Fc[^ [抧H B _/ҲhǼ\tvywNw>JpVB~?/kJ1N,rJ䩷9t~ -\^'Y-g;f'2+7O܍6kN@FYer drJ&Js\C4frGie+VR W x}tHWGHWIz; `eXuj(tLXT#]#]YVFU`$ Bh(HWHWP!M‡Uz(tqtPq1ҕǃ(lp`&M5'Mk}Gt[=M?;ퟙvk9 ^j7t` @W<վCO =jp%ZvNW #]!]ImiP %\#BW NW 塭Еf@`PWKf`Y _Rɢ . "AYFdQk]͋o~|[ ;:,lh^"Oٸ< i-JʞK ~uwn8]uD{3}og0W_NH8\uۺc1 ۡh%~3y[nVM E{I7{ yŃgX9۳V4oz^FrksZb/w-?vL 9uobUdWgՊ4oT2(+AuN;krY:aOl!7y(hw/ ߞL^/cNxyW a>u|/W[y϶r3෴ORKeHVkpVkh9n(ZVcp],D+?KuvޡR_y]WvhfYqںiZn` 膆ѡFu6K]i7X]f(t (w#]!](41$u ]5z0ꪡ5tc+k]'YC+UZ:]<ҕ̐ 峯 .UCkʑ#%v<\NʃP*?wCWzˡ'ns3*z&uJy`tGwsw`5֙CۑDW 0Ui(tК6Еd]%3jp ]5JC#]!]`Q萖du܉Ղ|+h޵dfK7[9Yn*ל g˹oa< wk؁亮+*R<id6y< $r P:;c"= `CWKb0OZuJ=NB8F?$u~@\5u>T<~#]}r+Vn/;d( EWVjá+}^vC;n/#]}!s^GףhRn/3J11/'8?*&\L.Mz:p?߼y2t`sgF+4ڠfY%J_/I:];d1Ћ/ E 7!F DNjq||] Ʋ>aG/>D6UD<[[\+\tvr+;%r- ;g~h`'rvn}j')k,[3 aw#8W@T]a,@E~"iޣ[:0 =ay>##ytA|_A.\޿H|">w4_TRRZ8,#( Ee(] d9ۮw}kÇ$X>>C,w_Z.s.#knV![}Jfh b^ Fn ^w9Z/cL(}.! eRc)+YՊlB;r]MZ*4Rx3 \I9-Udsgka,Fd#i.9́!-{GBB֖JJ'\tK%s`xb.$ьm%U3Whl-)jm9$U-ժuQ*cDrU@&bwX"D> R0l4X%6X34fxI֚#'N8_S Ai=#M0f,w;L R,n@Ze\NX*X̰!+ #48S@ë Vgkm6: f X!>#.@# {y\B*6Y8B"BKk2dA>#XǠ}2EojJm(y5JdE!R$mQwb u]|< O )GXG}-%<_ HM"tH-^ Vj) ARda!>q`"\ibً̝`OdHJI9G }RTaK P>熠Fk`E„ V@zYT%BDtP)x Hqۃa "^F| EcW]`fL'XE@ M%T/PtT-ZHT|T*݊i^(۫8(QKPyUmɭ+q1A[<*J!si÷Pޤֶ;PsAS\Z7͊|5lY֌9&*.ph 8՝`JڬA4̹(\oP \rCPPlh RX.H;'؊~oөꬄ*0W K%c' c l҄;#|`AV#ꋠ@UE(MW)Beʐ|! g,)T^ ߁02#6CA!r%Ōd\ !̝V;P C,Lh-W&SfHKUF1J ƒ ^ Ks`)I0ږ؀,Qt)oLW2/ɴddܛKHSѝQ E#(#uEVSkϲ o+Ex@_Q{  v J/+RjWQ@}*)ՈyE!!ѿ9RD0PJ 6@ *iXduW% 9U(ϓVwZ<-&sJ$Nw1.f3_wq2u1!52#B80`=l^w5AJ ./룫Y"DhHʠv!%%tMކ!JA3q!2N6!\hs^u@XK4h9>p4i7e#s40q+ Ftz% DFi4Hdy2Z;Howm=nhƾm`vd?,Ƙ$/YJ+I3ÒZٲU8T:~CB#4\N?CaεcA>hs|t\2g-J 1CH[4q1Ok/Y#9þr][n^2 |VTqZCid,C@aA;XJ9 ZV2AZ / ~EL p؂* jOR S:p8˲ MEPJnuKd `vNِB%T m6btlJDeHZp S$5Zgf#C3Iy϶u )..ܤ1/WHb\Ϲg #h>/i,+_lYRZMɽɋIw/xV3?_y.>y5zl50~h d't(nk۱cWW״- q58~[W^q@9@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN &紪 Tz0+p}5zNRXrH@s;r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9N $4&''qa8*>x'P\DNr!`@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DNu!Ѽ&'f\z@kL5N gwf:E'ir@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9N zG[ϳԴ_^woԿsNQ_V`RZVqJU=%~nWK@5yb՛3QĵscLW}yj/ĵP ?,2{Е!تJVDWUX-tUz5tJ$ ҕpK_]VWCWSUAJ8vb@ӡ+P"`x5tUj+UR ʡmCtEJ1eVt?7kK3:W6M) fǣvfBYu'LP#Z'bZ_PSTpT17+3;w|;j5e}s-,}߻i:~34P_c4o L^Ϋ$.@ٳwڿڣ^m<)}׳Xx|3Z\^\MVmWO>=??g]djq M(ц`W5D.V)p%n(hziN0n(kxMoe `[)p]5t=g|2Vxi \kvs^@ii$zuMtUpPtUzS ]YoWy|IC:]=zNW1f-WCWUo^~kW2nKWGPEWvDW[h++VWCWc S+"sIJ-NW@63HtEJrXYY ]Zk (DW'HWL^[ j,S9V{kTE4]zDpithiZHKnl r>764u5NJuQi^h?1N>u 5.rQoZeC%ZݳOiV/|$6/Wxĉ9joY&/'ev,Hbnt9*r}RW;"QW6~e3p;H)FoRהnCP~J*,:䶒}Ƕf(RVMtN}9iIrHEGբh mtESPjzO~Ɣ#umEtUz \_ ]ntUP zO~tUN-{BTCWݮ 8x[+'V5Ulw a*(!:A+_fUkT-tU:3tJhICWUΘLٱj?kt*(y"]ymrK.WZ誠Ճ JKb+ĪgL;k^{/t\qd{~h~(߃ͅ:DCNӺSt``;b'w9@N#)˺|&Yi`Co]TxfhoM+*cko~Wwkoӫ˳wئZۼrUSdބ9`/sDp& 4[:}n4D޽|/AT=PaG2_4mBۋըM'(QlPmTSRIbMڽE WyU5B']R}g߱+1޽ _nW?}!zXoءH]FH\.`< j"|at=vg:'dhQESjƎ;iwAkϭLAw-wVt]TZT5"u\wa.Z:(IVBV&ctl;;]rd^*a:.A%f"s!tX  \^%C/7gl˞~Cb}N%# ߵP7O#ky .'vK. BČ[uAH=8zEdϹ> .{]l2o; s]Z!tyNE܃8+Ew(_Rk)y ey$ei4aڿz\/\6ū? (s\CjF)שƻ J4q%( ">AX>+Ug AVvR!¿\FSfY9b(U̩MQd'Ħ{7Sg=GvwmmF%=u8Te^Iar-qMQק(h ($pp z\^S򬝔 KPL$_Xc**j&iY(d47V0,3J3བ66NXQM~Eœ]0,ŋ Y&axNNSւk&=u v.^\@Jomyă{Lv`HJK ^䒏(*>`N{Cq'hjwT $jB1+DJ QJ#jmM~4$c_[hBG7/*2><Cyqjyoh48N/s̊K( YrB:Y Q[dfN%ڬ4MRLŔlGX^,J@J T5`]:Y] s;)I2e$u[od\(Xfq(VVGwI<BYl2W- U0$2 cX d݁LTbXdM!bNH ̢3c[SPdPl8b*dLgͦs=P,b3"wILƬ,6[;J́y*tg2'(:!/<ihPVTLb}-Z}̤@tRZ?El6-SG.N91l&%..vNn-)I&UC̡IJ+@)⅑ȿTcŃaέPa> k8"~ *vryϣ||zK@T ߃z]˝xq|\me)Bˮ;-Y$b0 芌:Hwzf5YPjLm/hO b褣Ȏ D!9bIRBaSf~9RvpYOmAecP^^% Jkl:;U%-^Ik_0wx2w³y;*Kxc7WcM $Ep6G,'kx l2Q "h2o< ]]Mkm{xrLؒ3@Y;Ύ:jAұF`dE@ j1oǙ@(FS]^LFRVʺ3 ,bRҺdrB,Qd̡RP^Ek/~kziEEÁB9bȄQE[`!I̞x'J!et zEjH{`AbIk6"sR"(1njJ@LEDI,dXy˘_K_GZ049HWQ;SoWꮯL>μ4HxlKM gu}}SU9އ?ayeLMooN=8.iuTO GItzBPGY (NxFlώpŵP2ʗ[Z`Pr=<YM|wr֥Iڷ$,&qՊσ83%U_٢~:<;?^ ߄RQM<7ݘ wJFwx`z1« דq`dFÓB^5 ׫j'Ox#I mƑ<ٶaa[Y7ǓXwxy1g;nzuT%uyFzJX:GԐ:abֶj֋Xd~_:oK-?_6/ͻ_嗟7ooI߽wo~yŻx.1pm \|=!e꼰 *vV 7 7+\Sq͚u@ y𝵯) [ a 1RX#0 ,i):@I N2; Yh ӌs)!"K L%P+ڀ&HYKFcɠ(jsͦs?Lbx19Mp)Cޱx7Q73ӾzS'wL/OX>HÈ"o3dQ{^B  _Dve 6Qzt2/fRPi-zc(k+وb]÷۞<ate}4~1plUKDΛ5emBwRۓDZCY.:R䰯+gA]V`c:8<\g{B~VzݱenQ{@{K?\ jg 58k1HdXtAi"D 5w@e+Pְ^T nuWrv7ޖGefF S.}u{Bh0i^ven{gx϶}Vۀ }_AHQ1/o`m/,߹[y ᮻycQ[Yw7I_W"=olnzPgF\|>=ùj抉QQѝ`Xd| ΔЃGj=κhDZAq2a"vR股Phsʢl1އz6 U(l3D$T!KL L!)<+o6%XUZM~.R=K3ϭ w2q 4`'/tܧeOffALW 5YdK}>!61!9%3Fɴ lA|9yRϓn5:y1cBe@|f&"%;*V"Q:zpVV2.۴J00FxkB&k:X2( YD>rﵭGV|~CQf"'=dwP-2J?]msI+|ڹ6b]b WŖ= G/%2 e!lUz+L(7$H'K.:?dSdԢz\cuO>vF;ڒm3X;}W !\k.K%mcH`2 SlBg@-oJPKyrȁLB:cj!a6pWB #?wg[{V1"ȬaJ$trKKNMVΓ6)+<%$j,<2mcv}>-ل{ǛSXgmFef>!s|xw} lۭ5W|Ev5lra1D[t ikҁx!юPt%%dGD$=Ds <:yѳ'1;4vfvdq4<|?6_o}rat+7 >\69 T1;z {w sM?>)&ƞ-ᪿgsΫWW/zyt69ȦGb$Ά4v WM,hlӳMRrĨ>GOoqO O4HuOE[<4M iyB_`k1_LV-A//wg;]fD5QJYcu mf5 N_X9dzT>x>vwF!MFBq619]}%˃e e[7JpKڙ&\xwn8 G3 dw$rTv0( Oyh|2Eh<T@GFB_&Ko~  q]nNù?k@W彑oqPh^χ߼(6ZsA שwi|?Uh/.z$LXGu-fK" h|x:\LY7=a4ݚ Zs! oh]Eoڣ׶t~5cv/gDzsȬZ|.aA7_] M)Zp5uw/;䧁Ȣwv}ynJjX,.IZ6yy߹139b/Duq}}g[Fyt)|{pǐ125ƚ}Y5fʻhCgs11 \矺_aMN>hһu#PE;v%: R "ؔ2>f9(U8t nVr1)kӻZ:d_RKZC8vl96Q0 -Ԡ1ɣִ$E"R@h.F#vH%0䓺Sk[-k#=ඏe[ ;U (s~P_m{ڶͱ\۶p"֋9RJ³ZqFkr6$ɃJ(ߗl N9 ӝҕ[a9ҕ!,Y ,Z)rLHRp,Ru45Lv56G\3uQ9m^X"%9d 48Oh{lǦe=nm[D7eI2hGgI2Y#HO"914I  r#NJZS\h5Axgˤb*`- 'Q-87ܣ kJ4Ei"AY8Ap9GgrĕeRTL1G06W$K ]g.cDk$8d(3oFpa.[`ju zF=B]1z:J#%K C (FtC\pl^[ ױ4twߺ Ygm;Π'w0lYU~u'ZrFhWI\S+*Xin 2:AX3IQss[ne̞H]NJ>dan޵ўkM{ ;-Oz̕v9)%U xXL!b>l7'`I A<&MF3h":lT MwMtSY(9%&t`JU)+2)9Z)m,+.> _D&xaֆef a+BN̉C@3J)jDՎkXͦ{ZlM@-jLZ.j|BwOtWk&:;1'ǜaLKb1sՃaFfz ?~Aу /yp+0Clܷ4tquk&_ޖUa3]nXUՋӏI$g^^ G/Xl5iNg/ٜ|Ϫ‹y*Up헶ڇGHcLKOi筇^a/eB\BWVʮӕ61L s+KE)th:]!Jz:@2Q]`Nx1tp-]+DdOW8srD6,/]mWfh%{ Q؀]+NnNWk{f5҂ kE+kD)th e]+Dٱ`OWwCW`I Si+kH)th]+D[WIW-eAt%$CWWR њ/ä+m,+K7p!u?pt2ę51fgqdLTD$^'?WFGn,>#kj .VTiu8W86%W37~eDap#Px{1G WC5]/g|: ::QoyfoOPW=>oY(.p  \sSeQB\=Փm?ٓGrbT%,b2v*:'V$` HAvV8.nt+(Yo7ݠ(Ѳ$AԖBWw~(;Е\ +J+yhKoOW UԖDLb bVӮ4+K%%=GCWCWZB:B]=="qbNWlt\A~j3{ڼ!Jݱ"]ўz$ղ b b :\t(9=] ]1fQmµd+iRlCWwω6D:o]!@OWCWcڴf 㵙P~D56\{n3Δq2 8h<;."%J(!Dtrܪ_+n<e@ Fz!r':UJ} 4pBW,lЙ_uf5YBjEuo4x_gX-8Rw'e[ 6y e0:Qk]2:[*XrԜs_ gUm]5#ֺBb8CZC 9@W YI6m 3++l)thU QꞮ ;k(Ut(J j9-h Q*U +#M!\JgZ{H_!*6p~>F=-b)&GڕRCJ-r>qz9 3=QU5])Et*-+v$+k AޛAE)@W=4V{N}깫' ʋB9gnN~0PG!  \CFU0фzrڿ&62wR_ֶ_lrg H\_ndͿ_>G:  ͅOO/+կ}nonc}ǫ2z_벊%ؕdtufOy%oC!x| !4!y9[]tS/e8L>G-} +f^׭ WRJцt(c:3+ RCWBf1-R9U>%hbJY ])ZNWsdSHKRW%-nXܕMq*@W4vL橧wj7]t%zГO8`yw^ }+É!]Yѕ~r3\BW6ӕL@Wϐ(ud.ЕqJQtgHWL&V7?"wX;vhu6.f?~o#\i yFnlS;iw)zC.qF9ya|99Nv9_)ܧ~znh9S0O0[O +yr f1t_OW2с!]A + nnX*EPztt2̲3p=/m52Ɂ!]E!%oJ>Vf+EsWϑtDWO4-gjbԕwR7CWC_OOvznpO6n(w+=I+NRBW}+E@Wϐ b̂JRBW=])ʐtꏿzJ eE˛1G|w驶L_u}ss뛠ߎ7b-K46(͟5}$}>=.L @ow#׼6^~(V/ޠqtu>Dl %Yo{MӶ8ˆa*Z"ד\9: )i?ŸORz{%r>wpۃo 2jKG׭9Gjl~0>SU߆ʶo>,,@E y"Ee~DL~qQ>o>?O5Jy`~Au8y3y?蝨\A߿:Ҭ[_7ي?.gqĚG']B%wt*S2|+ex+ƠWwu7n]r}B{ eooZ_vcf`/W'g.goGjLu?8Jn,ζP6$Y$Ep=lɅ)vgZ*Ѕs&7㼱b%Rp`#4c8$֡BР>LC@1R % /$PR@i%#BR8'ݷn%Вi< DܬuZ]dBTчl0L#Zd9;0f %xDM5.Y{ bQ =wuTVhn'0Lg&˂whBJn lsPP[ `fRm>h-ip~m]VQ,'¬aD9sMoڕXɠ-]^-#ne蒯e ,q#YgeU\ZF6taGixu lGCVj;xX 8W}*$X_Ѽ1Q [}NrAEۊF mzZ2@p:);O-`R4g\GZD*1+w!6\m`6 MG29Gf5\ZD2 W4-~*ӛ\Kgd,yA Av%'܁6CAw^KCw(c6Yo%a iNB‚(!2iL~RTPcRlANE΂G nB:?%0;Z88)ԙpCK@4`PV !\ :7XaSѝED] QQ& ڳl;!JQ` lo\`HuqW`$usj{VQRA}<!J%j!!uG( T'% p`l,@u'@H=al! l∽}Ai!}tOKܠAł4$1P4: z0U҉FI2KDLr*_SG/UV nԍ fB&a=Cw /MGB[БMg%stKٷզ2AVa1 'z;JYBąL^ \KH VdCp0Z4Xyy@{:K`r@!Y#e"ԭ(w2xhҗՅY,TGշßh{B"tV nf!i64؀qp?xy_t:Ygto9׫3ɢk ƺD ]< duAgjOa#M5gPpljY:Z-ݚZsLdG0PFh Hfr|ѐYeF٦FQpݐ0𐗨=T9P.ϰ ڛR Ltp&%\]RA,S!e ۚ ,e(!= \߼UOͰ(*`r=?uŢx)1|rpO9F~GAɟu7(oV1 /9ZE#YԌ#x$LWus[s uM)68&ӣ+М.H.x*2Vj cZ R'FlLzb2j=ڕhg#yhdu5oD@R8*CNk (B5Z3Fa2k!yP/FWf črX48f(8qK]RZG7$[Эxx@Ap&M Yc6\WҘ" kЬ'WGR!.OS,(Z@Bׄ ;/]]~FzoW"[PTf.8][K7kD bNJ!mVٰY1:E"o2sylV*WsY% \$p+tos>kpw.uח:yug>`}oN_E7ެ)]agWqoy~?#EQџή:ƣKCݬT˛#gE1z/ד߮|<_q7WdKtM^9QwdLӫ{/OƇ{g$&}iBv/>{6njjWrԙgL?ݛlXq h^Gu#T)ual}zsv`^-.ˊew1 ˱w9 jN.?0Oox}.yIz~]psrxԣ.oq>oFwzxөWJ! T|rYZy]DW5*,R܀9s UV↪) U% G7(p  U+y+vUj1t*":BαfeP0ٌ]URbWUgeJdWGhWֶ!Bf׊]UB ݮJɮЮvCvdN2Xf쪪&bU虈_]'69aL;+^ݮl^|D=jWUuj?F]Һaٕî աM)B ˮ\[+Tk]U\]]qB KPU++|tdWhWq!BV]UUUdWGhW[w^#ܙ`'iㄳz}l ֬rlŦZdžnӒ/=4 %7}>GuV)Kuw=_?ZHG3me}oANF溠 PdnkeoNE%qy9ט;UY*r\{t7ۦdfIR猒]{]}s:^bb!9c'{읶tp2 T_RRûi}SS;1@7T\T/S=?*3n'TCvUvnUESV'`J.ȮЮ 7!e;Kjܚ[#: **ص]\ɚIZ|2XUrZ~vUfu,B ޮJNOlz{„X<`~r+/O}rhɠî,աM:i*ظfʵC ې]U7cWU~JHes fʕЊ]UTjGvuv%2 F $_r8ԃ\m6 Zn/kۃ~ARjnUCq \47T7TriNUPpiw=ϰ,esK7yͿ޾}!}lEZREI!tiE ZǛے {&_˽jؿ9Ӵw?MM`~|)vY# u+]eY{+y)n ܙb?t"=V;տ~]ERlOfٔgTx1)`ZtFcNhm",xCtS?ǚf.~O񲻖'֩s"tLI\XTbcWF*P2(x|wlԶ:?Zhy}WP-GVh˯^M&x `Z< [v?Vj걷~q)4O7SG:j=e\)j>^OU֜WW|O;;mfWu\}ؘZ@1LTR\Wx=7Ruтw4x{p>.FtuZn!FfГo٤omBoAFOXC?^%"iD)F7fQƃO-d#OV-6pq6,|7' 3g8&_ӫŷ浠x]q{lAahtr:5[zbt.pYbۮjRَ˟IQ ??#jNJ_M`W}f:ƓIrCAQg,`CQi,,4M?&?}軿^ݗAFf+D*%.ɔ"0Nsfd< 9ɠQ(\p>'V ]dKଵr`70Q{xqEzPxrSWg2t?-zsS' PO˾xeˎn[pu?r}OO֫PBW 1[Ϯ+\?p6?!DwIbg^ -W8/^rpPx^J}Hǘp< fV' \Q kxir9kUe6Xp9 A} Kə$%x'C&psxٛe!,f/{qrх9ˏ-C2 nn_|wUl=/ ?bY&OzۓE6wyQuBb1ރIE{wuntNgw2@ܳ=uo}zNg%xO(yka6(fqttRwG ^?O7<>5xkTf݅~Y>_2?m2w>s;?_CMѷ37_)7`@% VT*UbRU9ge@%TxסV>_dӥa X'F:S ;cbFjƁN5z*᥄$vx+z-j/+nOK;Cwg"^z}d"~J땃Y%cRxٕ,Y"s3XTQ`R( x.4֖`Q@;g!NUxn2l\[8JdܛGqW0lcy`v[&vMtݟSf+h"&,h#^xOƏwJVӿ;Px֤-s6C'ʥJ K &TBY@C@B Vʢul #d1a,sH@ɤ1QȘSH?kvkWĦ{شXj'@s}B8Q c[LdE4,qr 8&\\:Pdqt#"X?+?yeAV M0{y|_ naiy3.}rߗ'C)o .؁$[Q8ks% +?>ڱLLdXg&Xoޔ"G9H#*xWg֔J>L[cfVT\ɢorҠP479;cJo^[8 u^ _ ,k^\?6>q%<W;t~<>/r "LhfbȨTAΘfUu1/k5ę ^mSh Z&]2s;f]eLK̅%qIJ1ڽ&צvW/& EXRJD\BQ`O"F@F- ;$۫ⱌ*!cd(AIE3QcJ,bRporeR@jCqǡhzvDCH1551mp5;0\@tlt9,imXF%\D23ǣYg Fd#fo@79OxyƦoHsC}싖||qgnuPTf̎ѥtƒ\sJe 6!YC8̹1?</`a=>砉_k~|ՏNhǛQQ[^!%&K1N1\6 z%͘E- }- ͸;O=U޵#"䟽g^3LfU.$a[ۊe+vۖ'nEv=~U,VI=!`10$IG?F8PD6d%I HMJdDJiH>d> \ːT(]2KqMwBۅC啒anؤumH+8,Nmcx68*YokuыgG'/^j4HP֭<.YMS5dZҌOچ~yxu:;\\x}}I#ω9c + |e/H?VE>>x}ǫbHW#mFmqŰQu%?Ζ ==sx4]39Q~$7qJX:w\JH0xvh>A?{rp<ꗚp͗Z承Ϳl8!~_W?ݗo߼b'^f6 {H$v7_ _~h[ ^o3r֋f\78}>֡emO4׬cQ7[m&{ AY~f1?Ny-UYKfU*3 شW6p}R>Ƶ>e,U~C;畔!j3ib_)E$_=rWpf16&gz?K״yCl;~-bATPVS^ C,~Bd_gӓ]Ib'n$&C| M\ԝGI1<iN69R{9؄NTY;DV'd@hE' N02|_܎׾;WPkH+(d2 StPIR9eQ7l)})t[,PTH_h0 (Sfd1XI)1aǍAK u{W zyjj ! m2c)MzX>SS_s{=h `{B/ALW JvW/mx/K7R&4=B]K:5(55-bL )S2/{imc>gqkRb@+=Yǝv jg 5m8k1H.2M$&(^!*צ\IpWV^{JF Cf\&lCf{O"?n7xC ~FFȞ2@}c.AMz7O*LX`pM:>biȬK9Ÿ+笲̭Y9P{^+x0%[p|*Auq@԰MU^PJkûo ]fէI9+Ͼj0^4հ+,K=$S/~{xw߉8))@D0xxJ)A4P`"M1FTZ%(&!ҧZY*X/-#1F쌜-OMRٮ@;;y?g石 ټXecug +v `eWg-`IKUZBB6G#,S!N\vH Jw^Á2,He$IX2(Zܵ97 u-.P7 U:.`Jnjmv,[L~)mzgi}w}$՘e*'biKģq(S%TҲx/u/M wJ1S}pO =zˢ4Xs27'.]rou~5b2*५E'vχ|z]q >%6JGQ(] ɒ=<LLNhmȶ/]՚䉳s;S85~2ޅy$IP~~ds^41a>C%v7^?ټ1L}_ʛ8}{JC3ٝ _Y:jz:?p7j_jPnT^2EKQ 3+VZso[0?j ĖBw.7?;7X|狼Z?/a<;e3ꆅ;׋/K_s0;j,_`~?3;s5VFl} li,VTܥ1rM6b!0[2I/trZ3D:*"Pv+?B1kXgo:.l%bR6 } BBt褣u 5ͱ]sXu.N}Ң`Uumkޡi5LܰYW5RZ0XFiKrr!#d!PM%}º^r;f=vftEvAGE*QxH^),d*)I)dLY[osx۩EAe5F2iIY'iG[9“Gtڄ b3rݫqVWFۻR1R]4A%_lq!tF`'A$vs^:Y6z^&_(-#?(jΟRA*rJ1˛HHS]DiMj>_g,r[ ߲6wK~,8G(;^ EN%)+2 Yro:kHGWr[PKjXs=O ^"~ii| Z Q }n#^wlN{ߓntRRL^06)9jd:lاtF߽Qjw5܊"łюvRbXY'R 5-0P@H1R4It{oL=Tq=]*B) %3O,Zag()x;90&ȵbBd h&ORKlj_iZtF H8²K"U!D]V*(6X$$7!,)568k/))\q=^X2\ r- 3옅Qߴ^Y|AyE]P,3!m%1MBZkÿN_=!ڰ ޴URO׊f=bt3ٵoSV&<'0ge__hO:תع1`F]4'r0x&DF1a(6jARexR9_2*GO)SCGT 9f.Jպ t'ML:O1v+ؒZ+?U)AF!Hngu΄jgV>"ڶe0c{r5wNnOLԭQڥFvƥ|-֜OhNyݜ]S)d0kem™r_hU", HJFPyy F+E1R_] o۸+B{bwsŴvZI%:{$ʳqRpl"?|<$ɜJhD`@ye6dFyQɥȍ :*kU2\6B1 \qgbN78Pdt]Zdk@O*]Rw4lh T7 7(퉇/ũ HG<"dv:깒z:BEό%F0p?h%ܫ`8c&2)!gS7ۂ Pέu#`zc 9)*˹ih/ N+~g#je.N7/iq؄@emzdϾ"brC&Htxӑg?9PS\o, GD <;0#::O}h|}YóU+괼kUwO|:+E/?Wo&ȏGϥыgމ1MXL,GOaX/|[*5\]ZTrZX$Qz<*|~V|di54ﷅ/ 4&śe3>5֘ɬg 5dc"aмs# :"#.>~m"++ ={L8NbO$Ǐ23i`3 Lj) V\h~8zɳsz::Ao+3GiP}B+U>_!=zfKץkGC׸;3?/7" 5YI&Sc24HH1E-eeV~Wz]LEB #!{b8O6Jh< :RNc kvwi1ׅ\bm,5At5 QV8VlX pc;$` ~ItXW|>+0W7};ϡ*x2|`j?x7\cynh%hYMa#2E8 \,zBϊ l^I{TMA@q`'EWp#$ LDlPZB]SB~j 3ϒ8%TF% %J.d`7*e8L w)p! VGSV (EP!E{_K9*eC ZwW ~+Oګ|W#I?BZoBT2-5@{5@-vwOK8֐AI!/ gӇ;j6ޏ 5u;mmSVڵP$xzXr k-U;cmw5n EtM{䭙 CJ4ƖHWr[[hws>&gJ␆Q/yZ-\*ťP(J16Ip  yTz;_զPUЕ zrw`fC UL ؘ s-MݰvJ"-th%o:]!JC;A҄ҦId|?6EWW4 +C8%Vn$2t(鬫]+KѦEt-k?%m`ъ+zæ=[eZTZh5ْGz(it ]ݶ)1֒%0Bܴ햎.XU] ]1Jiua)sStpn ]!ZÛNWR3B7`ep=m+Dkd Qn0t%n6MO{[M +)0zpykdMiDY[hzhAdm!p4j+?SZ JE:ޝU;d[/,kVn-v^jlВ,KsY*8 RdL8y'ޓBOGEPZ6Ujd&DQ$NzaL,')w-nٺZ-z(UîbkMO97Z5tp}zh%i:]!ʦ:vtu/tŸMth ]\IT[ 2tB\wttn+]qEik *,Wtvv_Z (LFY>GGgezRb,~)|.omf&˄q)h哷EZ.o5['t;84wZN^{I0yeܩu.eTf24#!hk#ɤsc8'`9w_fP4!8=$ Ā;&tu,/21d!m`**HM \N΁vpP.]Zģ`I| /7'f@(i`OBм<wʴxu>m ^e]-V> '7̻.a'e:Mg/,?{׺Ƒ_eܷ `d`m1Z\S$Ȇ}Q4IQTSqitWUU]]E ;A\|Bit6̶,Qhd`-v >ַæxQ jfxe7etgy͓ydv]Iwru y&}m^]tqd8ЉюQgf'b#%}˽!Kwrr,_pBa%|YxUT6o[_߱ gZ}蚉WTum%o]weKlk>Tl\װo1ewrV]d1Nf(U+FUnr:1Q5S03mk< ctqWPO=kH^AhWKjI-hZ'H<3GG"G/-BT1Om0clHD/hE5\8%d:%B $1C〉%8\\D9Vx}6׈KNy4.jg1)T9NXrcN-B/F*c_k p \_ )`%dBh B[7{^*MSQpCf̍ۻnbкY*7 Eq._(BRF`"wxBCb*Bʻٖp^y"U+$^5@E nYDihJ ‚%;8} "1PǣBj齔AˢrfI|ZмM4Ǥ", fJ sAt1*-X Y1rvF>~b6J¨x!di&OW s:Ƞ 9xbZWOp* YS+%E3E~dI(  ey .6B =Oi/-EjZ~kNZ~swGJ>dB1`h8`NK` &2"-UZQDqbxOqjGݾ{8Z"x~,l"y|؋7aWHoG,Y@tE  H08)9B8:\R$2MDm)kq:xޟEFz'(IL"5YROR<Bkmv P{H`ŷ 5eQ:z !ߑ:k֔z"\#"J01Ax'6,B;nB[7jZN.z.VyUg_M]^ /qEAI'Uu8ȍ.ꛦ~:7ؕi%Nx=-^?ިl5T[0 fb"GBZ9%CZxed'ѳp9l\Kբׇ+(#bZsX:? B{@,X2MA OhFV=x+.uT')@I-! 0' 'qscc I$<#I3Zks gJ*S<ȹsh#4m|/^3n۹|x=y#1b@V4_~lcAxZU3@L L~2"r'r"Tmo1Ql6KJ|%eڥk#!\ L$ck)q;݋;C~Oi ʛ:T%)< i'ـWұ NrRbʆk]g9oELvdG炜vCA1YޟtX>=M@H0QETSxWM.*!CN"*KTf(tyEPPXO}l>'D`ϐYO_鬲zן;?/>L~[/_i?;vf<2vӼ{=>Aeq2{>_Nu\]|u)_Fa߆qg*'/?+;T_·ɡca)*v1z;|Ͽ7yχ9J9oۻb]5ҧտ_6үE0N |i2IB(̯}]yAO󠜜b/.p@5iJS/9Qr7}4.]=ϛ9/dʌ&fPc욽^OlC|Eڛw6GK|mؙjʪb`s۳}GHeq0|߹sW_rgo}Kؼӽ;ZV8$x,L+9evGntJ;;aZt3"/sB1Hɴ_|\_IWI:M؛rs}o w~~*TS?.&iS7RsAȴϱ87W.Ӌn~NS*8U Ek4A+.?:?yp  go7HK9GͲUbfkh˩PvS+}=Hbd\͞/n2 <.єo ً`)-#Ӡ2rdavm5B-uf7U/p]>t#83ㅷP4+;6ofý6Xo$9<{φqtmn;1lM٫-fl3,y2%,i_77TEA1B{P⧀BR9(Ri ITī?>%/+ (N%ON+S(B1*YiI '4p@YFe]}TɵƞLÕ`la\xӖ؞tPʙ&'Y&)D^q"ֳm^.cZDk+-#p/ҵuUj=1g^|u}`庺"`KR e!DK4k<1$c $0x *ÕqP(I!F P<`AED CbUg{p ۷ʓU/ G3NP!r$"=Mh7g.&^RL.ΚN/V;!s!^.(CQIS"g3_BVRHRF⥡s@xQ;f G4pl@* A_PPU&:sXPyLeNxfr ZB: qQ\5Qr1e EVBZrCDs\L2hP rhedl}v+e Of&[A;܁F(z wxw=t08`NZ-'Y;RMRQ#j݀;0#qK7sZǸBԥ3@H6E}! YMֺ6ʹ6ڴEcm$]Աt$.(T2v(Z>|8k($@Gt1$$X& .9oMACë9w5qBYH12G X '?hf)q30$2!'R(p t"TJIڝ2/C+JvGZ, ]'wݵDmqϴ9c /z| nJWXkD1Pqdk' Qqz+I&&y蚘jb,WQDž('ب".`LH%KQiR,sP#CJcsSr8c 8}snK*y~\d@fF9G}S %hk[R&L5@KI3(A8!v9Du'Y+ERX D^ѠRH>z-oZiBA 1P&Xn {KDy%ĊYѠܨ.&?_? 2"<m>CLҁ4A33˄d(ghH2"~*..-#Vyp$@&Srv3q|#&&Ֆ8'**WUNઓ b8bܻ:u-p}W3y;־rkNMih6(o_=;d<+K|! 6xako8,|/02yQbL#]fc `fg=D8 QQRJmؚ8;p/UZӅB'$.=yohuFļ48bNpYhtB/}xGbPa[׹5r7Yl?0*zE?>QԖyc~In̓:*@j&@ cNOVxQ|MϷ|~΍O6E. ǟsdGlߟsi ,4K0yQIGr)gXgȾGu|AEu0:*魫 3^ﺑd<^[hǕ SmQSU ˏW~YGͮgƓaT+#,xE>\"+ޛ$eX\&ùa1zk%<_[J7F"hC08$#u"KL@Q3];, H޲%rY4MMyD|BR3=o< \vCҽ=6t/˕9ܿ LʝwBif|??kk<Tۚr Ozv20͎I?/S奁C@]RG: My:.=އQϋΆٕɓpJogD78 z]840J('j9FpϯFcTpnF@(B '_ <.iAykZ97^cɔ$#菳y0pb>|\Ẹt*odGGRpJeMX*+6?&V JUFbFӻVn_+ &ma4t!̅Yݲ.ZvAd\QF>syӞC $#Kr}IMŰb$,8 #̭WAzУExC唑2Y%JmUKB͐\0WOFR7-^--_Z[4T~Y__.B _~o޽?߿:;y e>y:CLM"(yo~i[Ew/ZfEs#v)ZYiRn ޞ>Z,[0_Y q~| f!&m^h28{i~QFE72\p+bA 4v6_5/mv{XgrQ \)30,w:|K<8`OeP^NzlLs]l{*(rNsIg0 Z(hРa _mLie_g\6&-}u$\T:U]DjnyH+rjP$KC)sJE b1i3F ۀ?JJ$ ,rAlHr OF'$"АHkRF9!9S1Z-J1p.aRklI |sz$ىWjm+inx#G䧷T"bM #̦V2i bvCs2Iɡ*4l7y[Sdy֦r@5*-߲ſ)bWr}DToy>6^[b*P`A*0vVƋ=B. أ,db/H= "f "F%9DULS="[=x-m#r))\$)}NEn* K5ƠL+Mxj#&v\_IӰMY]*sc#Ogե1j>La]CࣃO1A%)P\i50$& c=l&glj^#lT|ϊIZا6xk ~S TzyT}AYl ˘2Gp+lR>jEbI .Kmg]Z!"OX'œlkFr>CM 2'y=\3dU p4 U?3x füߢ:\]"xQdy)<{#סwbp"罤X_!i :`M8F#D1|& 5h׼C}^AO̍kBɡXsbި;8S6G6ʦoBI9KFZv \D(q&_ǓJ3X?Kq{!\(Bnm4PE.^__H٪ltcxwwWwݛ_my-\ =5%9֛7ol7/ݹy<]r.c~1+s-/&v0l7jt|hΪ?uS P#;0zV4@J,jPk48Y(4VjZHAH[ƏL r{Ė^z[&&374Dͼl\/;]H.7V=cm8(5x5Z)CTQg))E)Q)FXB樓RU7 -s~ZT=c=烼 +6/pMP9(է+݇\ܛO8˾n2Oݓ65/9Y Ժ\yv\$ bvX(q>0a=TNm9pzOJ*[,:0^]]\α윅(E) Ԗ+Z.v 0{`"0 ƣ.I4FhJIN[ >Lv4TRcqsMs0OHe*Qϸ(2 d4Gm}t)tLksU`f@ IyL,%G2{j%F iO߽ތ4qP b>qE|T?M0C:*{?{a4gDR*AyW!O^/|)Ѳy{2z1P\TRN(֪ /r&=Kmt`I}}:P8|o߿ySD}zSa^n\;F-i{X)8R$X&W']'?H'yrs_ gUi]Les)&NJõ䷰{*^}`pYzEQ×/eprx gؑCs z](>_aT9.2Q:@Z? 4ks@QZл=@v7hJM}N}ب0\Jk)wudٻ涍$UP+ y?X[۹;uUUi1HAʫݯz%HdbƠ׍~|J1&zU"XGR?ubP &QCu%rF]%r1JTjګP])J0hR%eT/MͬHl 'K짣ՒmreժԿ_)}`g/DNW}?{/?j~}bUOtgg^38,u.0$#LJǠ!Xn&a٣ڻ`ޕ&/Qw~*=ޥRJ,_]>se4<0smer+gIq_oM9$`!XBQ$"d0 0Q"!Rp Qb1B؀1?vY4a#A2 <<Xe;N팜-R`/)'ҐmSe`_S^˚\΃.*,gCWѽ/-nZg>L֩Ҝ > i$=zx//U ^7jkmW=w]w_F;Wynen=zH„t.X.Q/@'96)\YIrn.Twr{o\hf c[][deHF)s1Q&DQ.UИ`&05 #z!Uxd!R&R/5eDDLtziK Xy$RD;v5vFΖVe5ޕ{fF>FCұP?N͸xO`0~hQ+hiF`G FQx$RP*b#u B@sX4w/ljr)jUsUP0hSi#g`N9Jyi%Bkvd =!"}ʍehpNJ7Ŝ(aHNJ3rx+i5ooIHmmn)Mc!g_;E%Zknj|r?:fCY8<,RP:qrCVFH|`FIpr%'+GHLf#߫Vkg aQ)F\*Ԋԉq>F% `!My:< 7J|1rEq?xR!ps*fat3d7G5Rߙԟ;{o֋ТEf]6Mknh7شߞȎRJ< p=pm84@M`#c^Z΢ŀ"zD9iu同2hq؅^ot1$Ao[۵T;ݡ-y_6ׯ͟&󉛌?;zb>mY{,&Oc_yre|N5Z76ZU{o^ n^|t~)WᲿo9l \}os<^Ym (+ԤC_/;':M,VtaNfFP,'ix8P}@OEggg3,]sW3-)*H79Y ~MӕC 5yQXY=^ " eHg e#ZV+V ª d8}>3)ZoA!/Ȍ]_JV5}1<~5sgWxypc=8L}xVU,I7S&V̛l:4=Ak,zӑy˧O)t)Lerj^ic߇Mp;K$aU\s'BWAzV ˏ.rRjQL|Y|~e";\,>{<ʭDJ{meΌP ,*הI |8dMr S.Mpv"-ާ,[|6/t]kք[_qIAEVS maUˀ̫am!?lMckG wOH&+^:՘|rq453[yt$vjj8Y~v6.6yqi;& 4qngk6Lcv[ϘBMTQN9]E_ron}>+^DDnލtS7Rr rS mt3hwL9.O0DQT&9A2)\Ax _RO/|wx!|['WwDeFHlf17pe4`ӢTrNž2Qz]a^nZMowh;|+R2{k-^v w^WMLVw)gz@S1;եLbVP™07ʚHY]p|gϪigo?L~a)R_dvz+wN wRbnG.^򎹡iv+*9[WTJ-Q&ҷd68P{r8Q=TO5es78v}6R$z𵏭<pE^.OM!'8w;R7PyNY5>n47_®ic7d1AZLݶ(E1/f٠v;; #ͩ ˭8]L? 6LkO]S\IÖň7!xrr[Qq}&}Np ¤(lOL5%)_3C%H:$$SO8+C4@ & *YBYCư$H-.XKPm \`y.v~ΑpyݔH$~n盼ZZ"^|v: ۜMuCSPLO7Ӭn)YunXU,Knlhe>WY[N"K%"גVXHS0kv:@s19mjNy5Êl]+.7烻̼2r=?&55g7W(~ ]йZ.0qѭ+w&ziɚ7Cj/m_9._CŨZ;~@/ؖ60+FX6UT&iY(d456~*U-tI@jm m2l4𹱢6{5e/F4lnfr-5r`[Rd|yh=.loUޣf6T\.١켔IUL.T8xNeQ)&ք4go\PXDa }љ?TbXdi[r1'o$0Cfd(Yt֔ZF2pdb*dLkͺs=Is5[S|" qWI}:m_,:]+ȕEєy ;r^BLZ(1"զjoETf\)U]7P`fEQ,=jBA!`r(2(ZP h&>(?4ܶH[r8mKP LHUH<g9CA*ҭ bh#)1'l{is̨~7817v-ߜZO9Mk?V.h1Zt2>( 'rdR_ǣ>8ΐmhI1Uu}sgLu!E6_.vK.zSIWx3IaGz-#R s$vZsgq{?|ÛNj7?T|Ջx 1fv6u;wu.cYכ7HעinoմpM˧^Mu-yC 8g^~=x(μZcW=bɬx\AW5ד2*[۬懋(/z% ߺW&3 heZ뫳Ly,EN u|sBź3}|EWWx7LO6lk/[8(q SgRQ,I|Tnx|bu+HOݝJlN}! ݬsz]"}krb{$ F[>&.{wbbT<xmXd| )asTumyGueqt[Qu@d$D,kII JB)(fM>жFB&>IHC%6%T )+o#DXi\Zuz.oSٷ[ ]g<<<Μi~ { z<'}Jɵȡb8``,#UoAfM0[~Sb>)`;7jg'gOIQ%Ϟ#/ PЯx8}[vf9ݺkoxz}]ypKf}Ϭܳ(ĺTugo̞<]']8>{@_ 7|2 ]^z8nvus*z۰vtrqY -o`EΎd6/Sn`/jb^`_'34>NOfq¶+nbBzo+)e&<#|']C E',@Zb]\dM]_JQ=ƨ2 H t@!$ЁO9PE -6biXGiA쾁Ŕ7yp0ۺE3'ہ>1°i7t;>:6 :8kD !H]gY$4*_=t$zH|*`mJw%Mi$=Y+,IkYRU5 6W#&mX% `TJ)g:Yww_Bevtwṣ듘ovws,p ?Oy$#3OM3>U2Twt'9Yl|Ivq\N/:0EuL`eg@P¶ԛCpGMB84IOՈS[=F AM6^Sf|S iSHAY*XAa/DhuZf)h6q[ ST3Tݷ"KC};{bkRA*rXM$Bg)3(59%3FɴJj5Z`IOPDkkKŌ %Vi9[N]]or+E޲eU$@ 7!/Oً ~ʒ"{wp4^ִglÒF=Ӈdթsl!Nv%u,ȕDA8f!:oNΛcʴI DX6UսIcpˆΓLID qV$m`%}IBG3'!nGKrc@ot)fkXA:]㸄?d:˱͘[1*Mv!O;svC}= kIg?{!|A8d>d9{؎v-$ &/DXd2zTނ6|x'k+13GI|+.VYy)MLw\2Y$,%G'՚T5E6+؍-U 4X|Ix_׸/{ͩ4Js0oTnA+wOY:M v*ZPfm-jZ4:9z6bQSllC*vLYp,gOY vT{mTtM' tb8̷@Nl:as 8Հ `>Z3,tu0tTuAt5r p=@fj=gv{e8''?`ʫC߼>=6-0'F^{m?9C+znm#++'įyxЁZ}mٺf'߶s%v r;y{s߂>m3Cg$.9ChOi/is.6NIV{oOz>95zYF\GK hu7D bj q)t{3P*aItskuh9;] Yh%Yٹja)t^] k^IWJ @˹5bڰsW#]}7t%zye7y 7p_ Fr/CWwb/Jv+9S(7X<]o!gJG:@$Hq1t5:^ ] pC+2Xe1tdRj{OW#] ]1y7mO¿þhd/Mޡ{O:c̯/ݺ`tKXYE̽MſӳvOz|݆|X\޶j4:O3d«˳C~C?.OgvպQq_q1ʒuS: Z|V^|poUmyV(ǐ#dwgmj B%gh+v4J6ISI5o9Wr;.{9Wԯ4վF5Z,sx꒜_\f'ɾ;9T{trK'rWbV! /\P:s+H5ObQ\t5ІݹϧrߖUp-j4]Mn0Y45:t5PQ]"]EPDWqdCW@t /t9WcڰsW+ȡ̸'ؿTNp ]2 ~Gz4;.\BW}#] ]Yħq Xx1t5bz}Rtut8j +j,-\h;] ){;-k_H˸~ꆿmlyɺ|d7h! BV\K $@P7j bj؜6(!ҕ/iVźՀRjt5P{ݑ|{Nc@+K#]]}2#>$vFevD8N~zv6Z & M[O^vsߟy1Y{+4m@_FdW?utA?tz׿K:;(& DHw-wV7W;G[7m{cQ?>WLG9:.x?q6ğ?#>1MmD[dOn[s~E(ލټ}0>SU5mܣ>U/w 28 E %#I*{ǚڜ2SlT́Ѕs&U`$usj{VQR@}j4 yŬf5Ɛ_Μ|p@eR" 9&Ȃ 01!# ʰU[n lb}FФi!}} ՘\۠bF\jt̘uV'7cELD#NPJ~ XymgU9{}ӮVcZs.OǍ nUAwd=2?e=DQZ$ >:%AuPi0E-:f%St9@ҕkɁJR!ˈc*,O(v5!LųΈ ՜ (dUFYl5̨t: h^= %&0A)Ms֕(2xhҗՅY,TGշh"Vwb@6Z ×ةiV!~5$!B2,یSе 2"Z豨wPRVny1ρ 7D2}ͱc8#lϮKD0eP_f!nhB;ڱ",P"|txEJjw ,UfՌP#{u-YyB 0HdL@Gf̱0 "tf% L1c4 !JP {؋pZX8ga1f&njpF̦(ZZ "_@ `i̐ڳ64M2%PYICxSXP)ʝd\zE'BZ*jw$a5N @y8F `f\ڪВ|0GK}Ӭ2C{Z)9x7$% <%*`S!9ԤV\`5)QJjtw `2 )cͨҳ0_+p}Vae/V+XQW,'jB5?i%I bX 0p k-* sc!oeИ]t`:ғ`g,%ى{Wٱl*ĝH,!yyx/Q%Ad`bͫVu,t.#X\Yђd|Nebs80s==RBuU(ك' e<3I:!A‰`!peR?]s4ܳ{P&̌eeTW(%6HЫP ZB =,BۃNbae#>A\Xy.D|J nX=0GD{25!&'%8% ]a$e;=@p! v3X-6W,Rc D``&'h ADCL]bC/ q18- Vi DLWt@x'Q[e! ?L/RJ"Qsqx<ֹaTA+M NIE1Oo!8|q=,2&SeUHWV.$!LN=^CףK1OC-KF/OyΖg+TFl50 uwfYGf#-8οgJpڬ&MyIEo~TngEжYiֻqnh֫sJ,׸ =4?T  Xu(m*(@%)*,3*P J T@B%*P J T@B%*P J T@B%*P J T@B%*P J T@B%*P J T@O~*U+J*jSTijըB%*P J T@B%*P J T@B%*P J T@B%*P J T@B%*P J T@B%*P J T.)`3J?WJAQ tJ @%*P J T@B%*P J T@B%*P J T@B%*P J T@B%*P J T@B%*P J Tt@tI M@vG ѭWJP tJ VnQ J T@B%*P J T@B%*P J T@B%*P J T@B%*P J T@B%*P J T@:%7djfzލJiiu^ovtw l4(vITewK-.JQt ¥{CkCtUk*pMgtNW%HW'HWVз `:CW6]ʶe?]g=DJI~`p;Fȑ*si]-J#]:TY)e +:CW.]VѶUA9  m8 ]t c[%tTu,UI|+tKj;]JNB,£Co d-L*+~FIc2o`#aGd_>$__~MƳy9Y\#+wpn(7qf.l0JjsM(z]K&z@I7u xjw q4az,\ >B'~W{5F7f FR%ԲnE5\ ~8k@*[p儮ڜގBN(SoipuZ$-"i1P*^"&6s+ᬪ*eBc,y:e0UB >O^&G}Y4+QK2Kj)` s2r^\4q&jeNR%*r% djNf2D ޡH[֝}raJyW"9QLc$w\9'ֈ6\v \ٙ}V^AY'IWW% s:CW&EWVvڠwut-t `ݝrz)] JtuteKYݝw&,hE JYHW־|Aݡ++ ]j/(%EahߗKhHW ]Z'OW%HW'HWi;DWXxWeێENtem >}N%/o(ۢ݌Mz\dUt@٣Ť4S:Qu9ĨJTPQ ebTUNF3Hr+S+Zto#[7p~y]Uy+LSrv*(A:A |!*Mwve:1Lۡզt(5]S+8Ctwp5!]ִ)Q2EA,~vpYg誠 J%~rp$!/h9Y[mwj;ev HW=XP`3tU2*hh;]"] ]1. '+=˅ ]Z~P6OA:X;DW[lg-׼t( ]] j4W}A =bLB"%heSf6TmC}f!Ӹ;W7cZJmWP)z;~mWc/bGK2S8mp=T&W뱪'uΉ(YSR:EN `:Y 4ziX/g`(t}6/t)jKΫ,\,+~߿W|~_r=?ʏHJ?5. %DFy֥Jxq'8W9 ^B7"u\bwx=\CT7O%>?]?-?ME=[~ԑ˭;\ox `<]G5]/CχR=ǫwrau9?`HKba Yo~+K͉8l6kw[5N `vTts>WՍ8#LM/n H95̣\0)93='WdeWxmف->?4˒_o!,qU%UK4wihecz ,> ד&u`<M;J aٝ] *(SuaP\"L2dqít^])Jg!a&9܂,yN72Բ3eխ0#,Yׯ_8ЙldBףtâ8@?LRnnpa>'1Ζ bfg+0R\jYVa~.fi2}F[-9 g?S2-͆fm7i2ȷͽ~$wqZO Amn_-zuЦ{en-1o3BKhGovZA[OYPJN)ϥ͉M[{+ (u֫ʘHMd潕yePVm<"KD+`Bn6amc]oqL7YTFFJx9m&{oM̜Q@Ȼ1О^e4ܞ+W&qO[RYҲcSSa`Fld̷*{/pt'[\߳'d}   TTq0y|S^͍KkGnz[uS}K@<z<(*Zr<:s=VJ)j]LD 6xmt0u A4܃] f8 yvFh+`E1ƤW<(2Qf)%PFT4̑f)K(r3 3ý,炘lRF)8qx 3Y=Ym㏏*|=s@J.K6eDuQ3Lp!PK)xHSSGQ?l<KRF}10yrNZRM fePBXZslqIznݕB ӀJMMk9SS2jM &jɐ8B4ړ66>R`徟M/0cj`Ӌ'89/j-HwL3$xWIpœ$ZtP<@(( 2X΂48#خ' wS@O2f"N>zG9@9e;(| %PdSpz*n[3;e!vNqzZ~0z"(REE'}uᦍDZh8>fׅ1~\1cv$J2!s㎔g i`.yjM_Ll<9 #?(+tEUpR)$@8 ߋ4c$Q22BҒ`Q2o3ZJfʲ2WYE.\P(-K I,6(IL`Iu=8o LPz( 'A&`,f߲~w8qA&]m;zyxAs|a_ ew iu.']z±z縁,ᣱ&B'--xk"&DFE+}Q0p rM/r ۋ8(xzen|F+gXb]^5 µMC?wO`aNS|Mys8VSVz`#vEOHO\͊M+7~IJ G.V҉3E|均"7EkK6wwڀF{iݲ܎GuQ(e #蒓yVk."dfsGX8$,t#KIn]$KufAU-05Ƅ̼i QF}j.&=j y5{|[>\]B_tI+9JOz@w^9#:XuԛWwem?PZ"ԛPG;wG#tIV)jV1C\@Vθr:̓J'C tvILjǐҞ2pk:Jb8DO%dk;ƹ6`-_\ B hSRZI 3H,xUR lT@lABd.kyMɋ) ȓfEʢm C sj3r>U@D?E*,DxP*ZX,2\1+(^asVrQ񋃰>5 f>(o}PEZA(3am \H3,EH ,+[Z!ai@+AX٭3[eAwa3{7g=9Q:}5;/k>Y/)Qm8JMvƚ`k) "Ta /V vB4OD#-H'm+^8;RRYѳ{&g?\Clj^I o'쎩tT4ʉ/UԨ}Gev\-ۑir!bD)4M6(@:Mp;;Mt3ו] w5 Euq3jwg U=hEd.֌ j윈ݮ63>!71 *oλ9@uqe,KDVvt[Br.r* , 8JDrw*/ٸO [Q :gU LNld9,MjkJ賷J.,\'e=o+}`$nvwh3nf]nE|Y_ŠdbHY!aˍ?΃0)$du%PBŠx:R<+TN3h@(Q@6ebWDpeB2;p1 J/ =k/l$<]$#Z.,zz<>ۜNٴ bMd|p1֣{Ci-쬶o֮ k>}=I}P?!Y޲m謰 szmX?ZA i&-:TYUOf _0Sf cf+JxbJx~̺QAhJ㗯2\>Mor:Ά4Z{ot _)-Xœw/xJҋ2)؛r뻪,5i/VԜ vMS:mA;G?0zos{l0fߘBPrl`a3 O觳ӖҏMA;J>& 8ŮnM΀Ӵxd6?f-G~6@֭dozݬJWj ?'i}NsiA/|EkYy|tgݝn`~:~$MO75hv>mRg&|rr<]4?ߦL(dcje0o۵[!ӡ ZgNR)΅h(Z DG(6> RAJ0-B6K1?az⋒Er:r'F_.(S)mv> wF-oT}\vmXPkueXP< W ,2JZ!ieDҙn-S2G6\%D [m+RR5dtٲ"@5N"gӶe8RSRfa/Vjg|+hNښc;6ö_+̓v%kv{ڑZ zaЏ?lMEUYy?Ոc8h Ħwl 8;Dv"۵ c@Ɯ[1%ґ5Z*Ud`ȑ^S`KN^F1uȹEʚq2&~OθP;֋aЋ^6Jvw (&HeYTAJhu lM O3?ŻЋ⎾݁ ۓJjȽ Wam-{~4*lD]Zbw$0AO,EHe2@".CwKӐ@գ@UH`LxjG8aQhBu DȨ)R/ ?F8&sqz0Ku A3T;(x~pFRZ[rHFϳ\]ːaog2J\(+qW>Qd O)WMv,nG{/Q]kTmdM:Z.x9VoF̪X ;H`Kc2lqZ"# y"Gx*o:hE-V0>(߶w$SB4bXX"( ӶJۇzAܧcP̸*RYT[?fUI)E L$BARTPw605}6$P:dKB Z#P S`&l=LG,g(7Rtc"_XX!&YcȥL!#T%INU};[M]t)?s6Q|6ɠz>:?{Ƒ@#a9:lYp!b, )ّݯzOCPlY {~_9C1.џo0~"G& ;ٕ&P*,ތzsp109ed稌lu6ɮQ;jQ'93 KdF}=A핪 >Ǵ;Z,?T Uj60G4T}w*CTب 弛n1x'mbVm~'!3J+AG; ÒH+y=.%TFS^'=vaFz>bA%eTiXb>I <&A"˸ɖs^u f71~2v=9f7ܚtYcVԝ'J1Tyx4/i! cgqoQMcwK_8j~ ruvRiz4\0#0IejL9si nqMn֘KNI ϬLA!0x QR$+HN !11Qx48ApR@{F8V?@+EG/cVn3AF3^^Zɤ1LyCs2I+C-ۋN?7M/TZd|mgeDa6ط%De߿IBAy r؃Q^hj V;< lb/BLuT-;x DHAEJrjF{E 8D{Zگȥs8D:ik*K5ƠƿL+MxF썜mD>>qbwk1Coh:sH-g|~o>CnMͻG114YVEهK&R&jnkaHMKN IW`'ZГȃHŜ0\' fJ3#1 21K$ڡ'.ZEιl[HnYzGŝ#)"1Ť_ ZtL̮&l ܺ}@xo7׍uPZp{]p9O&0۠ mƔ:ɛwc4^ȐʉFjɽ$]}XATGm_ј \“YQ OM&hlj&ФyB_-J g;] CIwoM:7hyf%E  9]?L-E%j1Z6hu^\@dg44宯V:E2I^WWr^;ǩ|\ԸN\-Hd$֗.57ioP܆R:Cunڎw7wݛ9= lwF3hiy"RYo(4l[-hzcF~N.Ѷ'Al0%D6G\8CSw5t'>jxR#&Վ,ڠ^;sfDJ$@zÈIi,%EreDeJUD8L %HȥL $NrT!g8P 899[ x͟wÔn&[O'oͮ[U3>s}u޳d?e{ft^ Fm:CuEǓ G+KhWD"S,j oRep )3 ՚1#(T)[FqvSڳ9n+Wiʻ@>9͛G"41Um\Fwoo]}7>;N3N;9n-@c'KiǴ^ٝ;*>cww3ıh6d~zyͭ9K/ٌ-S6]-vb=(czҭ[(M[\m?|d=sQ:rD)ja]6^qM2}?-PwZ@iAGM]΍Д&2E;| P#Hi͘=䘍ԘT謏 *pJF *nN9zj?|nAW.]lŜp֚>9[D,XUfQd63bDg( Jfry.H"%3\@/PH4F L.SE]ej?tuTWGMD'եC/6 }pΐ7]}OUx\>Ԃ׬NI~m\ps]il[onO?|^L%+:ѓ鿏P8fȮ|_3^]93gv+kH[$4DF~t>ps&f Xp*<_8evƷ9vQh^Py4 =Wx0\QPs3;ZH:Cf7e7>YIY%ۥ޲Jtls}~&p\+ uA jkg$3N$( (F&ܸ=U3*S{s)cb8&]d"D8 QQRJo9kltao.T=BYt^uJZu;\ڋ[fbNn&azȻ~jw~>x>x$NGɴ !":xRB 3ͤwT`czqs=,MtP+5Q'툱 $B黺/r~4K.h޸P*ZXș^bC>*HI1$ьAy*Ggq ݔW(Yۈdg''=*#!ZP\1g-H@}rz%;ѷה@v@)^TG:nYQD%΢gs:Qt¡9P K\~DєZS@# r/wolg<T&].e-}e\PqTT)zcfCY?ě|i;C2(x-qә?rW%gI@ bQSbdR^JAcZiAV/:~PIΫm ^*0%OAhô:iCX!5RY? \yl*7y!5.K%IXS V=4"{!bH$뢕!: Amc"G۸Oa([cmMTE494W(%[ !A2[ A&=NcR:\~i"X$)P!(=K,Q<ːٻF$Wd2 qym65=c62cSVuHPJJTFdʏ=[ƾ;.<[`Psf`ҝl6s~u}>}vx2'N6U'4Z׽kozԪU*!a!\*hB"+ YNh(X dL -]4CfDnȃ "@W^;,'-v{< vO`Z-y~63]yK7|څ(<ֻ} bZ *E*^]~)f|{iXל̪tj2W(MP2`xw綛u|ayfiR;]wbye>谷׳(}ʀ9R6FktI)x%24"2Y Fbċ-D.Ͷ*J^;/ʇ<\/F%.3۔ht[G]_#Wm!S\wHV\0baێzCg&,nx;Iw>| Sf99^g .Dz Emx9PT# ek&C/y,k &DX 1U6&^$0X I0Qcie"8-2xi\ igsP6縟~mp`~):zOq K@@.jY˖Պ mV pCʃVQ^UTyq}ԛ]{t%-lJxgc&1;7͗DKCH))N7'lD) ̐SU9'+1d3KKYd*$ˌ{2dsI3d*A)ςDub Ruk2uRkv=ݽHPH<84#'a7]A9[pYG=`8' ݙ|~pW28Y'du&S8gW@a&@ҁtdA {j^ %'3'5(FH#U[Ԛim.Rz8wT9 l] )91sFZ)EF!'6L76VO{뼥{od^?:6 CNV|"EjVE}.?M5$G"hN\M47Fd4YYV%0QD ()' V3qm5u=SSR#r^;'Jo:I;!d2 +cġitKWƥ)q5uV[?s,x/_j% d(8QC (x.+b 9G=p)TUbnJSR&#ATNHHRht$p8vWm:+ u-T-P+Pkgzv@:%U.lݘUw[@`yM4-۪k}b\SXUM(GO:d0b^wUid搅FRH1A@k"9 . R"=7 QpϯL>nwq"x tXr11B8vP@MEYg=Y;h[ͭ0Vg}Qj"ʷ쾺zX݂ 78z54PbHA0 HEy ygTQ=V`W_{$d_CIЏX56;SfqkhR,HMTl"Y&ml}}uR)J >X%BCx7i݄`Nqmӑ&H&g1Mt-ABAH#ωE!W00cf2G9ĔT( 0L`I*# un.'$!J<K6>5KW1M0)A$cQYR0.%)" D Ι*kyt4bQK'|tGN3X%xMF&\TciOEœ F!Ȥ#EȨFk0K)E*7e,9_;{aīV~%FA_3 %jAJ3'5&Hq5*BʖBm;xuхjҷOm mK{3Cb@9izlP '2`㼵 h(8M l$C_f/Պt+jbiE^p$G&e &/4&DAfc)J"pRUPl@W s(|-['Q}6 B p?N/5;{\_『[@۸@9QJE\_Mj$  ,%YjDk''Tǹ,i8^ 72p4r4n@ʸf"})c+ Ձ N#g,vhLcg,5w~2K~Qp)勑.o>5:A&`ehl0#26 w=oo8pY$__Gc{WF,RːV1bZ JXdD FOƙBR\riјuJ}"VG1 [ԜzSޏɚocsRWNn87S!&[OڬE@IL#`"Y_;ЃaQ Cieo@ N p&& 8Ft[iHa-}_Q)(_!xC˳rWkmqeu3bc#{aJo(m(.wmm H}j~>A&X8&Wk䒔?!EID#khH԰l>.2g믟yۉOgzGm//z?/.6F~4X~7_|Ϲ//3;}:2nˢ8Ms1wݝO z,,:u9SAγܥk?OFK k>}S`9S_U,̈\$8C) g_쐅]v5gS9fo/{֛{F6d6gA?Uk?i#su*j.rog}Kآ득;ZU8%|3oqjqx2EyP1ZH)(bO\> r8f'0Sy8B_Y)>tq0kc|U~82K3Vq>Ǜ?u%5pL;SUgϊ/ҁd8] (b4B# oW8TIGgٛn÷mRf~nTM>/y(?z)u]p5t(!$|Ap_{4{fe5Yo(,2oSweo){&1W}qo9A7͆[F`uIB-M/Qm 3l޺T> L. =oxfXuqgiv?eŢD4mESb)Ev|3oԝ^L'WhK9tT2zJ낛\& ni\s j|3Gxڝ)3%2%hilF1Ith/\K<@("f^@ޗ.o$s5֬$y6 [ZJWBOca5X@KjJˣM R0%**DoY 1McPՓd<9 jõqP(I!F"tudEDs;Thm-@yŒ( Gk(DNGjGaJ &NFisA ֿA:IihF╡sW@xG2pl@Uj `oA_PP.T&Xp,HXy zJutyG^". (ɰ T[g.*w~V;e FJ=_d >(?e❡d XnѮ^ ^9h ʻ7luہy@ ]``fr9ZJ-0+Whv@+ 2v(*S{k*S;uԕƬH $\=qR+.;!Ԭ\Ez4Y!.^={C_d4)&xq-)/2'="(JPWݻ;1_|T%S; n~vk$42ZW_X|10^;r>; /K/-*{XGwڗFkrZp)'REy+l󴊜_j_w*'|_^<{"*- BiUe FDgwBL9 4~(!StqCR7yag/=_?rq*Ul1.f݋Qd|T)>¾uxmV':'vz'?R[;XR*gz~wVF!nx\.YQpř2l<9(ib0 -F W:k%'N$[f/.+@5_:FrjT\\.BR(!i+mv2,(LdL\J8scc IJ+1IPr轐9T814 Ls{U-ԦtlCquBgm>\c6)]]`0I3MUgAf9a,0AU!|PT@= $H)d`@K)rRd!rgcVecu#Ѯ4)MVÉK# sܪu5d2ԑ8'"as<B&BmșQ MH!DL1r 8d@ ))}I>/S_Ys(x/NXvŹ!1=(35MgЎAhx3Dd\^y++IT{V䜣:Fr2!̏4@xj23ltٮhΉR&TZy< m6UV]]B<|X!ZzEbdۜwg 惜LV[׍ڹ7iCAO?7ŢN6nK;=8+~TZMq0R`R<$ǁi&b{B|^8;ߘƨ9mg< qWP +p8fKe#JO$iWd-}m%@$mO hy%@6$aWI 8-dp!bI٘8:D%. -k 4>FJDΣq!0Y4US{tYȹWeڦϼϸRJX׼Gexq^(/wJ/].A(U\ƠuQ2!#*wx) (S>Ͷ{{ҝ6&Uc,TΒ%6¢)-G wtDb25: Jh%r8S|M4ǔ&, fD l˹ F Jȉ5a9k-嬒1[D uk*D<9 0AmlLȑ84b{ ghB&kyDI<$D <;0f(;)E- `y hLZze)U\uP1gNoX?_qw%Z]d*Pi5x h"dNڥDU0Nɨ&2"-Ւu1jc,NChxbaEj-S??b=\rߎYeAS6{GGD劲Ol"`"y 5#Ob]ZEO"|ȮGEqmvsV8boyq6w@ aP$DCL[%+h5*a`q|VoǮ?\6)&edXe$!s6y gM?EU\d[o?jŢnW*J- pR*$'Gi{G=(;(Qp>qi7({3B"i@Lb*0QF]\}eW:I! sv*뤑F%6*W~2L2d2x%i @q F Qż㮤``#~\jU\ns"$Fu(6uͣc!9ۢ~w+Q-]_L$dH2ZaZj hs{Qs8FrRN#܋Fȼܾ"CcCMnKu۟v۟)1zx&Ί篞8ot_ߟk9b2h7\:Oޕ&8M?޷ޘ *M>_bz}w o]ܺh妃_dziQӓg? j/~Ϯ=9{6pSnr|yM_M^x3[b{PFIm>i5rnb%%'qB _ɬg K?aWUAipе~q jN_fyy?'xTKtçW #C9\NbQ,d)IdqR6@NisBZguV9ŧa>X-qIzUyn&Hաa~s\콆x3}(v}sG˪嗩6[\dOQi=}'L>EzBc0KGüu!Lee|K^OOqԇA=e?E2W;pUfI6~/Y/JjLC?0wG.Gp6..%󆠃-I}8zOIK?*gӭ ͠*]<ÇRAWEF]* zCݤݙ-Lx "\b<$9aMPZ9eVNg8A5m/6q|G #$7lw4rά"l}zBrso%fي$EpKԽT\Myo>Xgy?kӢX3&fit0-=.fͬޠԩ7zp Λ76?ZV6*9[Mubx ἫG832)wQ^:qO*U~i9UC o³htIF!5 Ul~R)|~53IhŸDPԄCې{f>oߛ 06F1ߦIBM[Դ a6=yZV]]yPNnXuɼm)؜"~\=uCtS7:plaԁQ2 JFΥ(Uڸ+-2Ic is\d1wz`YZ(u@6t2ȔxF(Xn^jsǘQ'G!Y\KPD>QzC( m3mzythӻftl:`l)eq2 Gu!p?N4΂Wtx|GON2⌖)rkɃJ(HYu[s8v凨HK R%eThDF)5Ln lD3iyi˵)I"iudHT팜ͦjdVG~>̼]|&eI2NRN)3&N i3 ݂%M]a$r ;\h-E 'Qt]b+rf9JI6S AY F£8N2!QAAcġxlU* MUh9Hf%qRI.P3o ar,-ey+7 aQ\UQNswW$1).~j-|aN7ą@ ɂ4x-"khϥG; @ȁCoYb7\%ow0J@ NtՅ€;oe2ڕ9{JjUZ1 &40a@aFsk4 jN}J{˭ɹK@xH!!F'AːFw6=Zm<Y& cW96FT2(\3 Fv E$14X$}Ay!9.YgI3Eĥ31>ձtFΧ 14qBYLLm`;㼍!7µM-Ë5X+8~*ubc-]99d)e5G|Z(˷?L*MRJ/YD8??jQqh Nto:VH"s8u;vՌQ"@if$&dD'p#C1Y֩eb zAyoh&ՉKYSڳӝ4qe2@-s$ K,pAO&僠V$tƠ\tCVS? Rþh|}VWVOBiy4㤓paq:NNNPfr {ɩBe/8ob"B9Qь9._w!jEﱂ6k4fh gE3,hXxo̊yj:Ec.f6EUp]L34Bq=pfEDrdIPuMIKF"⿚vuc,ВN)$c͜eV~{*:{|v뱳 Ұmy ;~Or\to8\GU^Mfhz+Kn]ߢ ([5rX{qk%K *kˮnDCZ* r[%Cd"/uP\?[uD[9mn G=4yz~FkknFk~ GݪZoq=-r׬d/X}+!>> pjF 4!ߖ0HWi#(>Ҷ]}Fux!&ܕ,Ҡ^:sf;FeӦhC$NJK5N&Of)(,:*RRr-$id2lh;1*GdK0<3ep,?H嬚Hope{?6 <Eҍx> mV׃5^[pNnhW0+קU{͕E>Pb\]{fpڮRGœO*kYז쮺 ȍkK*oQ/Oڵ=B3kpjiyA |491B$Jk m%]cќqhhN&N!E I)c`@_ J0 "2}Zc&hsWVCqo]]?Կ.EιmFQ&@ IAL2!Y΄xANv3 2b % 2ǒ@xOrcU^Ą7uEfAIhM:ҝɡw?]اg}[7/ȠGb@!h<07HfNIPP -L8q{ީfTFHcLL0Rd"D8 QQRJk9[TVLvBu pԋ<[+Ĝ^OC{ J7 w_Qj"B&Et<& ( g I(+ƺ{SdgEF6A DxcIN w ~ۏq<wVG}`זԑ3Ž2ņ}@U3`aD3>7lHd6Q.x#@-C*Ћ4ꚘK0`&(4QG ctASX:#g>jģF\:tMBgLBc!,ɰ~-D*Y$*I5PRpjҔG-Xzƣ< LI=5dgly =ibqɮzvQ/Ze$}ouFļ48b. px}4s;E!7 ]]wL:ڔ~ꄡhd1_;3HD2`JAqĜ#u"Xv@)0l05B-KE[Ϡ<Mu=>EO)6@W[hc/Se=tgգi%9Y<dԔXGG)@ؤVZEbJ 'Lw%g10i-y B+eCX!5:Q Y?dTu}<&pC -/[@ kmFEٶI!dOHpM}8%G3?Ų5R{ܲHys?ٰ#/%dPTƒRLK682Bl̶ie^gf.52vy7.ky6~,!ɘoVW?m,(!q_L% TgBӧd1z"%!0'o)dɗ- \$! pAM!H`$%!)ȣ`RbZ :2s)Ih&&0|dܭ/;HFWr 䖕Wzl^9;Qz;[~#!h  ΖTY{(솃f9;XÝ퍷]`pMjP~#u1وb|70ք8/dK +埗BFyTѐ+ Or>Qb 8 tLvLv '㔖#@$ "$ 6̵UBs=2,Ӆ^2.FD.Е٦YF:kk*ˬ@%c(^B+ [qH_Yy[wZty]TOcƆXx8Nj=G1 bJ,Nc L/*U+^$C>B1'YV˂Bz԰,Ⱦim^m7+N~-ƵnϦnЁ;\XfխkTA#jQN6YWSXeW* HqӇ[={&G֡^tɗyb1^]]\DDPrxi1)ŧJ)ꎧURC)E݉b2j2R ԜgO;|`e{ h1!cS.NMq}l&YYI*tʢ`)*mW lKS ܥDW{+CZϝukaq>9L6+o^Jo&>+FVR= `nKg«YE[8;"¢^,EJERA*k<wU5PIKH6+F2~@ d<wE5$qwr!HJ{wzܕH`۹*Wb1o'koj,_÷o _O>gS#=M5p J|=z nPħ|^Vԍ^e5)㺰<P?Eq(5ՉCa4Άa:/ӲORY&^E7@9Z3? yzⴴL^եt(hqtӏKk1IEfa/NwolȃUITXFYcD#ý4VV eORYe6.24D> ǁe䒳4hU.*-h!rcJ0gaU(A-o8nM7/Y. s%d1,7 9v@zЙ30Z"CN$S=U+W Y6Ă*%  * Kf|ge@jS됹dƅY7 Ի JP.A/L_1"3\"hk-@-ٮ%_-hE j-%+za$ 7̅ dT@K(l%ޒjh36!9qG˶W]e+N hkE9ׁy:0g[MNyqӁҀ;oU6U9{+n%W[EVlMKN ё4ல4"Dq^&s[grb#8|d$wOmXyv'YdQ,͠s!jH\2)d軘BD9UAo 0⾠<X˓xY i&sDc}6eiM5>5Mc"攄AU )+ Z/VYpXZ)m"-270Dc+0 g';¿HBf胱Zq&+RB`sdptf\)ō v_!(!zmDMk ЌfR˚J^L{ɸswx f?r 9 :c*B1XQp*$JI=qvJ%^ٹ{>2EqI DYǔeN4h*k"D689cUP(q@П&= &A|)xkb4 !hsX8o|ˠ5qҸ:sHFLQщhw 5f:7aV4~kIhQڵڭ5=:z&tZdhJ{mJ, eu)xHБC }24֓nG{G^un i*y(l׆[pTڶا7tf;Nc{7s=}!=S{uZr*Ja\Ϩ :j,0AY;lB Q) &H%f2pcr \X%&n"S2-[gBu9RJ]ҍK8qy”r< DŽx\a}&?6I[%Pc䤕5rŎ*1zGeLRgd $H9`G{'EjȽ{px\.kPd VXs&mG@)B9CkANE 4rd :AvGs)Ih1&>lt滕JZjN6߭t÷), <ЩUCFWUY;4}'p7$F?JtF;K*)`㘌w<-9dv{zzb!.R2#EHh}`#9QQH$L+ASS E?ܙ#3{z_ ΜG`@ ~yrɜVΠcj 6VߚwrWg6Тfm@Qۑ?. XW^jÉ^tt4{n|_6ލe^kv'yx>b̕e4QO.{6$|] .q!`AGĘ"iGo̐zqHYLLöfwuOUuUu= vݨ>NS+<*͙wcݾmvoçS3]\nnNy;~NQۉlv?82JHGّIYNgV7/,{3 !ϊ/C|m↘ 9t n ,psh7O;zbil9`_=:4g+=(ףТ|VnH2R0Ο3gLjozzewϿt#3^?Lޟ^aN{75dvqv8ڍ0ySJX&9wO[p?o31\]qDΦy|8=D;EʞJ|ËK?:#=sa_ 룯SҌ0Og$LFyW0?}3S|\!s4Qku3NJɀQ=G?=Y {/:?~Xdž2yA'Pt 38}>@C0/^'Ό^I8@yHԥ.Y F1:-ZjX?UhׯiS09)nZq}D+uOP/hli.S]# }xx7lLЪx7@Ԫ݀ͣcqe̻}nKΆZn^'o;>1HfK?/StlPk;z m ݈1>U&]ohC|4ke @a8Zw;<9Ql<|4Wl .$d&H(~ [AcҐe$JӋ:0"*jπvֺd)6Q0-HD:9BZJh"he 颬`JºpcτuHXmirNuFYSD~Ih>َl@?U?`ɟ*ǁV8h U}DUpv W _-F|(QX_fR2{Gр׶B$Jk";J3(Nȡ9"pKY%}R  V~2G@)\\\>z-ďT5ठdboG18-(CM9Z(OJg} c7C1YH}< iEιm.sLҁ4A3ԙeC2 D4D[F($l ϒO 4H$D }gJ :A.j0׿.2>Zoxr(F5Qm<]י ;j4_fywkPT%jw\Ԏ`#Hib^SdX}I?k>ŧ9,lK gZ[k?vFQfwt͞YeޏKdZL)(RtN ~Cit Cߓr d}Θ:aR$@,qe=4Չ΁ZT^j梧k %Ԛ\OȅƍwϹIw4EKbboo襸DG?ɐ CG |S*vټ]"]w.dx-wFt Lf!ӧOR-'t@IcnT!Nn" W_ۯ8?~{ׯ)3/|y/  .|4?<ݧ֬q܈]}]}.cއfaX5qy8~5Bڻ^1ַDx̮DZ˯ ê5T܃PDT*,$gO 'sw1xgܕ%>JObOT?s= qQ \:\w[DB];)> ܺă:pc|R9ZOmpXhb;vPEs;OR/g Z J Ped:TVpbO'&NubȖ0هޢ/<;q&zwYRN>->pIRp4DRpPz\(a?lInqӽQ<^YILm@?JJ$sHUia 6$fB OF!NI{IDPQ:do#IXBe pU2Zb/zCFWd&\ x$RəΊWZKH-xJMOF+ءo6 S'Je8tnZ5Aҭʩj"o xB[E۱wܻlzos\mtgK߼3_}M`oˎ0򭞛a6loٺٝfͻNEg;Ӥ~ +\.u:w|Jk/ꣵYv,mZ fpn˻VA)'.^~,qDBԄq5PTK1zά!-qGK܁\蕇aRͣSYJ>r 6ʔ1Alz pJ2M}8&Q)c 2 H%傩DvCp&q.۳%7pv ylOlŖm%+Oh0bx=EoSd|%F4f'-yYo'7>1[$g2/&_y8 0f`nkRoORƻxѐn=M7c\mon.Yb.JSJQZ.kW\mCXخN tO TN tN t/N `c+ l~\vp|C+nΥv(nwy S2|JS"nY܁^G'Sy~L dp[ {G埝.XM;y-eT4[(rף:zO[hˇŵ IEcح.I8 6xmt0uDav6s{w=׫|yFh+`E1ƤW<(2Qf)%PFT4yHv\3Δ% 9Q҄e$-l˹ &Q'Nϳ8WRlW=CR5%A;i7QEm]L3xpR ꄷ4R!s<S)pRF+gc2Q` L^TF`&𬙡)kVq5HΤV:OWJ]H>1aP YyͲ>gvJF-$Q-@G Hwe/r V,; 2 "IAΑH%pԌ⊲@@\&vf[w;"zj,=D6|GgoawVxFCod >zFɒ ?dSbXE^̎,d:7A%8[>R7Y>>~cx_n~!aqj)-0Gq9]S}-u5J$a:>G N}GAGȈ KKUD˼dBka()r\Ef B!.26ڧ}Wڅ{B4-K IN/l J,c1X _%d6:/)pЃ1  ث-l-'zc_ړ%rg^0.> j|5هtUurqԜspUu1׌Xr L=WxQ)^[y]%H; f%6h>i_[1 B6*Z1Cn.1ۘٸwW[ AiIi)awX}zy6Kl$?vxUK'+f E|흇JO/īH,h~muv]4 a$w22JsEQBlR6\q c6y zN7q.]֥H^gmmTe205Ƅ̼JYQyz#> )Kh X );X\q3΄ғ 6Ë }$}Ru.}.nFӯa|qV_QQ+`PI8!_5ҚHCj!#$KzTt3yæX׈fq9i2UxT^fK=s*)/Ek@. \~:mjֶso>,pX=R^8f$Z!Ox?XPY .DnNS-2W!FSiVf)͙ęT{*EukG38OO:О0qv4e>jxjEOkY[aImM ;YTkķ_ []4.L#43 KKMsMg K4,zwc|&%&nB y mzHJSn~uTc) VkY%&&?*z.&WoS [9#9ibU4,SQV6j[{(`d}xr=jVsw~ƪuE۱b糧 (^٨% C%V Nbmu09,NKqog)g%V BQ_/rkVy棃’o#D`Ejp"Y5btj{&uR]xjzkf}vll*nea/~) .j8=|ui7TN:0NX)'辟tvpy8'XMi7rZSnUKnq-ȕnyrFI^4T zppẅ6B wM.Yls]h*`05ؙ74yUo|N}[{yW欤5'ܜvvV/YVE+SnrV&xPv#ǘjYߐot|[LEҕve-Qdg Ǿ ΆA Un]ImH-haV|h5/C4!TItuA Pղt([:RҊup];+DNT:DҊ2) + \%!ҕQx^R]!`YNn1 gbAfHuteֶ$wEWwYC+{ޠ]+8&JCNWkfj-\-Sz([+]6zP X\CWWR :t(ixDWxUO:+I)thmT}jYJrErZ ]!Z{B}ۻY CT81\fҀڭ[wwv?v~}]8Ǩ0տsiIX*Hm"MVExINLm P&Ne^d4ъ QRj WRtժ^e uD7DW(Bi1t彧+D)@WHWMtAt.n9g^i*pJ[()µ;=]!J6C+c%]`YЙ- Q*2Ջ+cB߻Z zpkVZH]Ɂ6zJAqu+nT1tp*t($)&jR 2wB tut)$uVZh%뽺BB tut%\DWذb CWQwBtutŘJk5kJK^U+4vSd]LV<ɊbOE!rR 2a{^Nju֒Z%,y1wZoJ۪<8rȁ(մB Y.!"]i<#@CWWCWv_:Ci@WHWi]]oIv+ >,dEă>Ȣ#3s)mr] DIu]uݒL0娫W/fUo==SBWϐT Spb 4-&UNWC^*k8''>_CTq(rBWzmqAt5G\i)t5>խЏC3+CJQX]~sW]] V:] d!]Y.I] !,(^ ] Ou8vboBWd6=zKe4On?()N,[w-{:ٺ}g9" E8btÀWa: mx*q("9 `Vn1t5@kOJ"gHW %Wa:7W@T7> %kYtC58K+|'(k^ϑCX/i1tb{Lt5o(gIWbˡ+eZ;ʠ]n7*c?́U`GcVc[r~q1Z &/ߥmyss{MPv1-߯дW_FdP3}QwIg ŦmJoޠqWb}\!XqGowM;b )bvv.퇉o7`M!~c/|2zOw3D fW֬ŏ#{~6~Cɪw܉ߎ/>,,O@EDr21O3p@}yycf{#G 3[4.G^B=<alu/t麭kզ[L/|&vg9d(P޴ &dr$gUn:[BҮd /Ao>/7 ^#ګ5˫Vߤ@m痵.%gzEBɫ[S}ˬbKQ+G}GS&U5fх:kU:eذO9WS5{U]*3SX4|g?.4n]|" TĘ{'dRWYZ-hs"tDK1heD[K!BMP![%k`DMԴA3j]^u9E"gqhѵh@.o_BRMR̥fcהּVYPR1&34sq}cs1 EG B%ލ=N4YW1tڣ7'!1NTRYTC*)IW6՝s2J+ lt sa;z7EעֈSB\RHHHI?̷ ҋJSh#JK6>`DmT/  eZn6kr5J`1-f͋ŖbS}HFu)[_c 37% >{(R.I 0kjA 1Ȏ ўF }m.5뎼0Q#_f E US`,ܢ k v**6(:ݠ-;y9xC9ۦG;hADUCˮė b ±&6VW6b"$l00ձ;udzEKxٶP8Kp" RLU쑔YlXkYd`UhS%vdZ[ElJ=RsA19vx X* kRTL `6+(\d$ QAQ)tJ௒2 VHPcLB2m v^Eo[PB]Ѳ܁pYWBnTzCZ dܠQ (0,LhD;iש;".FC(dԭ9ƒ;.鬛u0wT @T|B|)q$8TR 3@*QMe|7ud&v/ōbUڛj((SѝD]`#)#MEUP4kϒq9kl eBRFA0r{&ETݳ.RVK.=#Y̼Bj|7 !ѿ)9o5R.8(l,@E@H 4vM4y V*КeM23 r5_ݠ=bF\8c)"9YhQ1&PTؼ*Ca:iB1"'q;t&Xv5Z u`c{.0AACZI|txḰK<<=K~td 6Rt%C RGU!ˈc*Jr %tF\8G(Z R|$ LJȼ`|B6fd8 W.+X2]w˴v\7*6vՂ ٣;w669 6IҢGp}PIHu53ZSP'ՓFCoׄ bL P.z~;=AzV{20väDy آk"* F m})ENJdcۂLv0Xg$SbFf`J;Hoެ7al *'~x+Aqj)5NnB0#z7(V1Z8Pc˜Q -F:j#Pw:H9-й ᧀ:ʦT8FՂͨ19fwܬhZHF5kփ*M3| ڴfҼNd &ci@P-źC9y=ݳjyA^ ьSi,qeAPW ٢WqdͨAi Y'XʙrL =W&3օn4CS3AO8QZ{(=mGPrm⊑$54JCiC`7k̦riw4 J.ȎYyЬ7.8I4M](IWKl$dip/(繍*?!tAyG58FT1f˾݋7~+noXhA5*ž&HuRքNi"KP;7TccnM-uuPA2no~q ^_.bn)4fvIcG/-l~Z㕫ƃym퇗1=>gcs'n]z"D4noݏ]wwUDĸ1.=r dONW8(i8w wt叠+/tءZ bjuv)t5pt5P t ʠTK+vߠ\BWSǡNҕ[j&Z ] u5кWWBWϐȠ=~{~1ݪݰe͛_w1n3g.xDGGv8V Ym.8O/nW>ޠכ1BS}g^W5 ;4}rNv?n}, g?Y-ch홲Tnu D{xn]koF+BvО@mMP5FT]&/bʑl]h6C9<q-]}skL p4r)|("({,ɉrBkAM? r'y՛~,-+?F$pF]ȅ^z[d:ZRѵrG[ͤpV\.'h 2B,%&d<ȧYҝ>jڨ% %VNba'-1dqQRYKFZ!\N!ZHQJEr/08Bi]pm[ ZšNWҎ^ ])!-+lx{ >Z5ѕ*3ݛNWR^ ]B405tpۓhMs<e%sCW-+x{ VD]̞Mo.A3d!=LWUAh +W0 o4ЩAss al'GF%Bʚ!JfH%FrZʰ󢱪 m @x`>HċdyYy}=~oB#Wy?R,+an~\e)4 Ųߌ~V~eX]*tw)+u\>'@_׺M'#Mׇ?J^<--g}Guo>,riW߷wk$ W{[jsP۬km93o}BwG b7MEL0.n)M*Q7@u4s`[FEL1Vzcb~O`"tпp(K@庼6)`q\JFA,zJx@/zͫ%,<A҃^L2dzje7nSxHW'&s GЪ,N>>HC>aQ3&d4(7L6j0 7kBIU#- !#+ $K!dZ)cX1GH}S796w.,kDJSr @Dܩu2*̖z=nR\Fdy i!HScrGm`*Sfs%gx&sؚ5bmp׈%[O x2SS$,+CEoLmvV[iދ{ubn@(,䬄[c#/&,~>TkubE*(t "~wY Cs&8S1OR8NPjU`>ETu_3wT4eǕB]nNm Ŵծ1p]jj [-Jm|R?pjBVn4Woil{鿝fRƪ-]zؕJl^k]Hv>{zb9g UΞt޽ i(!ln{9_vɓE׵ƹdTeT`N[~VaP\"LD˔1[ "E)xPLÎ3I͹>Ay8Yٍm apDߠ#b1#`E{Ea|#^Ai]@z"@w+O&塝9mG'1ǂ6ӱXvǕ5b8Ljb6~^ =Y#ۘ51syJzkB'렌r8=5{ed%2YR)ͩ<^urhlrT`PIǀ(!""' o\^M9yMVOfQ=zjT/_S\W/ jB2EK mИN2lGxH؄Q< 7E-evDpJ ¢#{ØEf25z|ʨFgd5LYBc6h)MX2ýt炘lRF)8qgJYi*~_Rr9PyUEDHL<3@- kNxK#2;subKރl<A%pLaRF}10yrNZRM fePBXU'jZa6.NZ;G$OLep*6b6k4^2Y6LUNɨ61p"ՒuQ qLl>>6v~ڶbdR9C͸\`w~\M3$+$0'^uD:NeY(L5,Hy[}#"m>IBCO|ȱGkv>t}@9e;Q o#soH3",N\@9(4ΨFX"ƟuYȱjASX\p{EP("w44sRH`*!K69S ]Zα[^n%|TS>nW* +P2$I Ӂ.6|OYsQ"Q22BҒ`Q2o3ZJfʲ2WYE.e?żS4Jv;faB9=or۫B**?dɕ! 2[i4X $C F "$A J "|L‍&?dy8ڡ ks]DP6@Vf IkD{#J!M'^R/|@!W< g'I\XgL!@>Wx%eҜnSVE&ϵ*2yUɳ,b:1.:BB2qIC⥃EUZIIY9)!v|83gPtsFU}Qr3FeVS!&x%Ǜ:lG;.;S0!gؒk羠'}6Dױ;–0G}=ڻW [pHqв vt:V#6((n:V/@:%YY!9dItz<+)gu]G -,KO9 RdcZZe+ϻ-|jg`;5SűIoXJWܒh"8[,oж xE-s^t ikҁE88T0QjtS&QN닅(KK`!I R)*"339W !똧] h}F 33G32p37`$it$Y-RZm輾6X~nJ>C8z AȮ&?ȚR$,}(|X}zi͜Prp r?i8M< cROSs"V\9h!2RJuU0~Me2(*U cSI|Umbڑdy(,5%IRj4B{K0$l_Wxt!$Gb]Y2))i@r` &Iw4^x cM+6M- JC8:VNp.pdj*@r0FU_ȞϮ',c$4Z LR1zV)V)ZnXH9IGS-)z.'ְQj둌1|_-kL+C[וw grrۏ;^:rB )lr2U6W+n,, {%})TZ7!DXmD!-+y9א#RLaL'Q()Fwkm: X[6Ge|Ɔ @ .|hk J&6Ud%e <'͒΋gpPR[$ 1;ӱtF· 觰pQ(ZI$iE& VRX1\AX!,=1lRf:Kz5JtFNbL@gƍ1ܩacUshe6;cLt76C=381=;`R7&R٤'b662W*P9 1id.EÇZ'{uz^DsG>kDJ\N,pZKɃ0ZJz#,w62 31ZxVԚ H!Iֱ7YqKI#̝,bOXc]B錜4tuGCzU0ꆣuU*7ޟ+OYͰغ-|  dwbqN͇\^ X[+X\V$sڳ~K驾b+ٟ]'= LZHNڬT 6Z!8SH!IbЖu^ZĀhtGB+\dm29/(vwFN7'!,+Ϙ"IţCIwg^ea(7Xc7i2IήG?,tIZhyƴtYƓᢞZJCaYJoI!T.fx`T1 SKl(#BD)^Lݡx9foLxWdWT4Dr:D()9wh$9 dCa\pOCQ-.BޮvsОiD?3F?L\ίBkdbٓ]mOf W:xM;; %UY~up-K>>q[¿vJgGFTⅺjY= NIJdп|Uۻ~e[;mn#t"j[=7|:0ުu{ʿ%wzQW .t29筧p͚_u5`}]svmyR{;x,KPn0nSzb,ɲ\ʛLI 1yi"s(oY5Z5 /_h./_+_E(Ew>8;$o'[/qZ|qz_j_O/C#t7$h(3Le.MD"gBzko5^eiWх.gYBYƿFҋC1 ]vk˭nQIvÍh9ltjFKͷ_u-ߟ/NǓ V/j:5|Rv)'ѫ,.ں|vl[wDe~0"nl7]W~Gzy ?K؞CΛv~<Fp%0B*W!i#_x-D==8Y*"Td+'d(3#C!q.b$1V !B: hODFeŔ67ywn~2RJ/ ;NŸls,[!3'/TR(t9` TPQ1&rC}BϚOA$PEOK7QҰ̥3r *3X}8mFK\wc8k:Yi<`$u#Fn!"<`TV9iᎲSh9Ft( !K'*L.9m!` 72pUȤ֌CΰJg3x.4=6x_&^. 2^Sdm1ro;T||\_~;h%t6̦(!u"!De,SB$̂L1mL$5a1&cR)MFE _j zɜBy$#9',ݯ"gƎ 21;㎾hm3hnkjA #qdŦ 1Ȓ1Y!s/tA-C^64Wd!2䊬$ђRFpp$EB%2цCX:#g>lU{S+xFkD;hA#nx֐1f#7!<IJg%7ɓ]+;@kwd˺Eq Q j%r V\L$yt n%Ȓ&<"PL:fgT UGҋ_|qSX/A/zqmJcodjyQ;!3ɼ1%(UQ>zqz7sW}蟦:sIlV}7 mO5"F8?kG%h`W ^ʣTBPb)֣@kk (YaK,xzlZn3'ÏNlRڊs٭ /vMQL)k^˝ÂC/L eb֤4XμJc&jB|ky-7ӥ$a kIY'J6[UPywԢFE>ܞ}<xlϸ ժ֝"2LD ѡQReYR"c)K^ $. ,?4[ACA+M`4d@+2f"`ȴ0bH$%1dxAVXDO7gRtdd YdR %2)$L5EήEgBnNrN5iϕ%᣸CdV*SM7gT1IޟҩC='Q>ԣ?/:V-%ɃO>FdM\x~rUpG@q0L5RrB]E<:/I ^๯m0$(SN/14OΦB~\セ&hBq\NNO 䲐Ϲ9oFa4-^fQ%Y}m9^//f/UJrF!DIsvD7\؃56-h8+Q%(ǣ;O=jڸ޶|4lymp+5!"2Y;- 1EQWgar=8y[wgܚ6WXzݝܷU!#;;STFI$V,ZQ=$%>B\N$}f <~uQkOcA=M7TN2u\z6=cEyz~zYkϪķϺ*亳-kRߞuN} ˒>zj&Jl9bcŇN/OӋwo-{ݻxۓwRw?~wo ,vNxթ7Yy>/g6hO;k{iM7̫gSzrwV)%]i5VoAl+B'#Z¬2,VkiU b4`okE;0s^$IJZZv1 I;5̵)Ey7pIR^^=_bb|R{mUAe\5hMR S` [%V(-Px`tμeb"ҷX-cۧ;/ccBXGҪ4tg;~CwW6>4`t- 3AVhgV UZJ/RֱV9Nyv.2P .B RXuAS&w&h#B٨-)d0g({M3qUe\݈JVJ- C|ŀVr7[-SWVMnd^{7yN5ńmgܬiz@nfޛoQz<8GK/ϸ>}km YZqP=^"Hs/ɱj\Qvpe~{BU]wܺuW溪_/ӏ٧>[bڈ{$~B;ʗ;w5:VGnխJ {:xw~ &{}?^ooC?_[7RaR='N.xMgB! UtVW`25dy%C pAR2Dnw:"uEFcX7N7 TEܖE)hQ^jl]7BPIO B羐,* e(md4Ty(*6ueue\}K/ǴKk|/NyyIX5+?֋4qn,\)9|5T_O`l7I@*i=-.W eK.-.R9HW,5y 4]^WD j V 銀d+6D+"eW/DW[T=w(QZ!uWu՛ ޔD+TWcziFWk]muÖt5])'tAI6"\FWDkl"JH4#]!FW]iL\vw]!18F]|YbӶ'|ZϷcS|J۵Zh9oΦan8.OyV娮i;=%]bxwteK-$]mBRm2xzq?- e(x٦=냦VgK,Κ;%`"?! !Ȼ `ȶtk\hrIԎ"]]pee&@ =?_~>rv8:_2HCklۗ9a%l•K@:Nd HW ڳEWD7uE.j^ =3!\&"Z}'2Qԉմ4?ͯN˻ 出-wcJ]]^5=BrN!M:bv;%TdYYj- yޫLwݨv^2E>w]Vf`;<2ʽsT[$' S]XB#2u<,U&ySKnqtxHCZ mnakٴvezQSKn-9+`Hz/`|F!gOVuE6$]PWNJWN("\\tEO=kw]IW#ԕ`+ p-h}RaF x_EWDk"Jg^ԆU ]7h+n j9~:T]m*0U۸RltE^s:!cQJt5B])%tEp+}ovQt5B]imtNi +&v](omiA H# pT.'QZ0ԑ6yo]?dbFF` W!yegmx,mS >B\' ]m}DSu崖 ={>}Ȅ;&߽hPIWcԕנ ltE*pluwS{ x?Z#bQBz9V>FZ7|cSCgWpD@~6JЕNڶ1)FB`)]."ZcQjt5B])$+]!WKEWDM"J/F+ 0?p} qdu2Į+18F]܌NWP[\*3z4zLHi8n+zo`u Y.[wYEN5g>.r('{d`URxΟQj F8֒C=5K;hmқԎa; h' Dh}H B$]PW֪#]Y+3dhƮ+i u,xʡ` &BZo "e)xFB`-ph!D1*8-']!FWtm>BJ%UՋѕٰ͡p`" jp~0z(md2IWVtE0;| C?\h Į+t5B])/QtENT?}v^J{m`+AѕFKEWH ^Į+ F+l|I %0!)'1n&+ LHVt#g7V*r-TAۺrs0MAբ @*/J\H(Z-w B#:mɱ~K׃ےz^nD.gggKN>L/fǟ.YLc@'_{cYT5W7wwQ/fw1ν.Ɋ3k|T9lDS?=x|?Sޅ&%@{#Vyrf̦A`ZY uqIY6{j|RgIY! H<@\e-]!-us Hl0NqwljǮ+@ZEWH]WHu2F]|Dࠁh."ڡ&b18F]!!]dHձ٭5! ad8PJAyA$(!fӐ#\`MN.QvJSCn< 9x 4pA*."ZS= ir  3u_"\ehuc$]QW^aszшt6zht5B]-T{FWdW^WDҘ+aCꇫxSG:]m[kCW6j۪ڊIW,d+5$d cQ*t5B])tF*6"\hjGiRv5F]it]\tJuE)}'Yfk %qo # ND9\#3i' q$¡^+oң@kFi>ĵVpII{@ڥai >d6!^WDUueAIWtnP\t m"ضgIz]9UHW@YFW끋VK]WDCu!é ltEpyyqxtӀ"Ov"(Iz1rVkw1o- 8 >\. <mhvIJ:vCW.jOU8'LfwoL>>/N&K)Ϯ9 FW?gx${vA,ESٗO~ۡS/)}v(殏nO&y#YQnI]bMdV,d4gX!:C UgWY5+?֗e>{4;1۠W]fZ~XCg>b߽ֆKY@wo`{}^_#iǻ˜1(SAgǒ5䇟VKYV}:@1cޱ{!J-‘ wcp9:Y|}ʬ;lS~:=='ٖ5⼖ޫȹ d*w0_~#oxOrGV'j:EO쌢Ƙ/5qcZ" صq9fu7F|X+po}?`,:v Sr?}f} 4f߈1/'fӆ*C`/GZ#bA0[c%[nӳW_~9­)XjZrbJcսC{|.Ћv )oi(m,vr3yf[iv$-y99GByٻ6ndW MU~HllRq}8Q#sC I)I忟F$ \Π5_t7J뉼Gpur^$f:f~w*PoV_ls鼅b}D`'_Voh.&}jj^^^ټj'ŋwӟ.3{ݮ0_ʋjjձS-SPVe%-Cy$;7'!4@yA].g>,3A8hw7.nY5]~ċy q"apa | UaogopoN|K4P:ur؇{4WWaZâZq-\ǡŚ+kH9z"﹥BЧ+iG^Ӻz#d!Q :jmaD+p}L-A*&;h 3\#ӿDFGԥab +Չ#AebP "ZqBWFefSW Jy1P.%:0R܍W@.D)V {4@TeW$)N05DŽ9F%ɍTH%L)YS @Ygzg@5(E@>mATɑʵ; /m" $(:,tE@ݔ}:ɺ9jXDY2Ԍz?!^0/mbFm"E*jrANҵ ,AYKP{ d'9g uU$D`9,rqs<#}"F{qbT*gLZSLbl &7`=u9˕$4͍"˚UFJpi[U+ 7̒ꤲĴUpڎ3k-k*vSњ;גx~rCzQ-?.P_V͏_w+ KT;]ޫz|OMVW'OBJCYw֢^XQZRTtMGFۗ'ufEukf9 jDHV+1J=#mW-pZV3x;tzWt)|Goo9zP>TtxsN36*i4IZU(@ H墖NH thODe|'Ӥ+;uG}5jyʎV%ވ~hp?DnٹEȚնړ am8N\U;a <k'mgnhܜ5jōY4;Lj>j2XZ L.SHǀ Ǖ$C9cI3q5<gy޷žӇ6v[C~_(6MF1F0/Ս YOGYKhм(QsjG/ LR$1 @25J{q>qE!s5eAA`M:tL^ik:jПjRKdچE},jљ āltn"1B|r{2֤C_^99k`p`_:m6NLX2"1rf$1f3l?Gi/98Io$F} N=OFEp5 *M݉5J.éJ[!걗@Lv)\W/>z^k0z%9*^.d(d!D,r'$.EK@avP`k/8nXENQ ͋,2qr9J[ C$AS;l@f$t"V} 'g ,b%0&A_ "j%$.wF{`}ASd3Q ɔ56yr*LAsq1ʦ:hv>0_AOؒ).!>)=ӹ;c:yDAڵ8igQyV) i/t;?C~>D[Bm]JJ#!fPڶդV~?)^;tiU#卼azx5]PV\zQ?쥷'Rו-%ek%Jjdk?9 eJ`T W+Mu1-A+y v+˶!.k36f<~BXc7T g k-q@|AAأyYoP;m1uV.Φ+g]w}GY"}S@RÙB3_؜ɰ?-_ǽx|gśQ b6]yjq |y.PK'[8QÐ @'1>Q;a|_Xqc>%0;Xb!z|,CetBmtov㒝UXn*_{8{n3#RIpM{ϲ\:g/Jf8qƉ0N܆q207hy)L't4Zw55bmlC)\FIP -Vuſ4oe:":*poqZrgpC>v?y*)G~X|ppLÉ3 '4 MP+ !5vL2:uvDiNDႡZX˫uT}ׇюb(P&%qK(-AqNukhg:4D7@MDgZh+vH| `v4r^hlcPb[  Đv\s2RahZP!-iQжfq敪km$G˜;ؙ`w[F>JɲR.W/Si$%[j%3#E0 F0H /$0V_xi)7c %+seY3-܈Y 4V֘rōT8`mKBP4wRsJS9S$J0B4cd85/X* )h\4 OpEunſr^- bQ|ANTjqֆŖ ׿_ Ɲ>VU1%G8O@gƖ$MGx{sSkM>mo?LLI쭝`_ss ,JEr])PIPFK{6},Sڭ8ݺC_(dG =(#PuRֿ i 4U< h"53(2P>wnFNQ !{w?_ r_6xoMMptpo Ѣ/f}jgG<|WI^;PRα#itōrɷ}_yv}J 3 FNK.&aq]m4P;,nWU3rWn,o>mo˛iN'b۷?ݗ߮_nF5E-~4$X<{Z,WM?n|Z,h:ԓ]VElq{7ymw/JͽoߒEi#t0QBGpIfc|z,>S͝?ǷnAX\$xmCJ]_]݋坋L w m#rW2}K_}Nfܪ~_/r1iן?$;!1*L\7u!L"!YY'چvf#W^b1T L#`_ro,M'Dhcc( 8؂h68L.$}oP:եN3M)殐hK}q[Tz+ eʦ23)IG/DžN_,_i$#VX@_'EamZw8sx9F WKnFds<<1`2Ю%WBn+*Ga(7q#ngFQ=W;el}@1 2yTսM#ߑw;^D6=;>t3-ӔI kCwgE`q'3@M6]Px{[WqՆ-)f7rsGH~HBW>sIDR=vLڋiI\e@I3M$ J,,I u3MEUY`> T+D@;/JQ!Yu>45҆8tܕ#)||e84 /gMqwhgOn%;m"?ښ'=~'0,Bϋ!l!c9>X (#;kA?Dn^$Fq`!ôD>ꢣVj"/}xg@b9Pc>I7<>D/e:V9FhɲBQYy-kSbtڟ7Ohsyt׳Zޤ!YY] CѮ;%;5L0g REC~\ >H/15_"VN{{`=RINぶ'bz%b-$E!LKK(J*lZ0W0t4na4kQ\H4Hfζ:{PP1̜s©yj- ɲ,R(ySRpPïk>_qꍺH]PTn\`#lF<Ìel@)aԐXRF9,y Vm{;;4Q7Ln":gy C eRʬGptȻd:FMV eMÝ1|,u]$\Ĕ6M 4soV -㵋R%?E8sɈvY?ԏ}gdn~PP/䠶pP7TA< rv?ALf؄mPK4 F-"yǮg3eF !ֶá+~[>1㌢~Aj=2􋞰3S ʙS$ISh쒿cu}|eP_*ҏz`o48(?::lm[&fDۏJyf.t8bf:3$i:0TBx | Lj])U 1ZOWݕE+svN\oZX7 j'PR)yIec36 x8n٨8ZxmU\Bc0Om61,aJ0`OS'5>-QHl:8,dtwTD9/32 v'`BֻՐssb^x5aiސ0v6):+M5,t]{|ݺj3YtIfpk-9u}ڤ?/乚%V`z#Wt&zÞ s*Pϧ[G\:0LIdu Vj-(<'HU)b*FQT}OAeu`*; m3XN>k6ԥzAeOi oI5<'z(&-G!cx7"|8@)a0nk*g2pI'5\d0I07OT܂ ,X A$CɕU-26W s2C9D0N{**y WrTsD0 ˎ4F1tꖪbX@}eF.y!I>[ 0v JgYsy( m` ὦﰃw+d0ko ~`kJ-yIg 冂8I}8B0Ĝ B툗 {cau ,. @ń1Kh>e}c_(4{*2&f4GvX!K!i@sph5K{ ' H$m-If;lh2Ռ^D`M:$ZGLh>{!YwcݽٟA1Ӏ9GN.-K aO/{٠iЌs?B M>sT%l2P>!2<r_nKǼ!t;*:؁vԬ*Ԡg?g RT 4z,hsOnБ GAD;>;&1[9%'ՠШbb!MCn]cZ|]NKjy%zqvI;CyoI~h ud1cÿT^vPAA3>bU\\F@H46ȇ1d4֠70)˗5Hi._ɷ٩@_NxQB+I5RlPv qt $ji})sl\:gZ7ݼw3l ul}sgA\gZ)Zc+)X` ] >^u)GRH8.|a8,Hٞ3ti.?\`}W[2eS%t[.#{"#upRD!ԄJ3O<:x˕@5Zr{ܥ^p4@{GRv8V /0pj^;=_Lݼ|(2*roG9*P0hok٣_ZjK̡(F}Iho;jg=8M6%d/-p,vc!d-b6SeJdV G Wܻ`]:DGW}7,e\9KP%K7#()har҇VqDž-/Jo ҟH}K̝\IR ;p~\uDoU-7U AVp+NcΑ.L"x#yO$ms hE͓y,e@ U2[I5sr'Jaom@!KisK^Yegh2KV6 -h2c mZk"keܾ7iC f伫z䣻~=7x JSy(Hb2qM]ڣ4u~-~1s[ #W}ڃV ŌrePYqmFRkA?{8r0O =e<$ =9`wFWK{A)RLI/i  K"WW^|: s%>qdTK4 ߴI 8ܡQ롍ĺ}'9` 0K}5sK35դAlfFt(8DE*?ɟ>;_ ђ*ui>1&N'aýE%qid"b"klUt1JB< +N۫/y( 2 C΂ u:LJSRR**OPo03Ӻx p 2;W^F,w>ݜ5ә_SAqKc238}0ex1l8J)z{c04B~gǿ&8+_bxρxoleJd?y>{ǹӼ9̧Q:ɰ3DU*E۴YAԥᯋ!k_,> Sc !d~x͢E&. EƋa]zk|l##A8LmTiG,.l $!`1V5? 5_fyNPT/;ta;qf#ː#S;,O/!iRDNWy gݝ&\#A1+)V01*d:Fdw TeVQ\ͫI7YD]$EZ0 r 1 ^y'yIhð  U0e{k,:Îj+~ps." l:A/!D Qǵ!PSz<^ͿhB BtH\Onx5{kwhT/ -3Pgk" 1ye2{O)/`V9B?^pWN.^Yy8]H,O'0WrQ/p.gsE!;8p@jQ1&-T * ei8N21ht0=ݱG;IZC Ј d-D dhA*/a9(K"b[7gYcvq. IN*θV#$IϑxbTˊQ7V }̑'zNlR\P I&VQ V/%c Ih?!C} p+Fwe`.>Sik/QIpi3{IʠހQV֡ '>;(ZD y$1SnaXZUVT sQ!URsj4~j:*Tw@{!?d%a]Z,5L 6=U&,e%2Vנ`7oCڎcqz}G|Le)\_;1cUߜKUbb%4z zOp=1<2W]HT|[:y5vXi؍۬x3Ȯ?e;3.˿Ux`lճKu 6%!0b);+o@hr'Y!’]ּk0jG/B:P ~"҈-W͉P X$ SXDH\6zTBF?{g_"eODB֍EʩK hRJ4 \w<ypf6}IJe{sT?My׉d>[e?X'2% B_p}r[f åTg~F%h:ALǯ:Ip)6k$eXFϓF 7+N>!0r  H U,IX1dl~ lЯ>[>s[~~4Q?¸䰣>WS"꨿sb?+~/tV_W n[C["rK)0aWWZWSLj!N\d_ZcH1W &UcP Q jX1Q(221U^$Rd2= VPʻNp7NA-D {y rDFE֡쏣an!&snkۚz+i$!JWFW!!jwL-9V#"qd`dC#@;+oM;sUC^5IpJLI[gQD&/8>W0֡Dl҄ p s D%4o^}_OoԅJ@+iBشU![.iGlq֘&XD,F-:!.1[<eb48IoaC^اȽR+%]@.:*FQ]FP + ;ك@O7bN w ꕿة!}Zw՟C6^.דok~ ͨZKgn")l'ڴ.7WAUFFHe ~U\re|KǨ`.U_ZKC DdD„Zgec9N69Rna2;AVwUtܙbިVk,R{C?oI*Zw@虓mej_ Vt?K r *Z-cݜ2K3/fztKseb0+X ,}3.߇~)"]LټCf5l=XV|#ƪ$]뷙u9sh9f*%$`N ;P\ ^ZMMBh/hvfP E\!4N8+'T%̽nJ7I{IItZh#_n$v2t30$JOh r{V/+m)%xO<Ùʩ9'W>(vSt7 x:3F"W6fI饓- !XE d Ax|z%h[ZiQ j^BtuO.t^3NM.߳z#Lk & ]"Ɇ5i,7{^bv。\#oXjG/$ %4}[[b/W>ƠLu %#Z}矂aq̸6\Di!09=.|@l$ɛ<[*!$9X蹪^S {}a/L?CU ҲeIM6*@;8g2R!)6\ ̎Dػ[bo9;5i+P~_`ےyh+)\zz4OF.qlovf?J76G llv8leM{O|mw;*~ "8+>A=({<><>VP8*!IFZsٲF%9l*ۺ=/ǐ/xj6 ?dZ&R^dzUj#[~,,?YJ&ewU{d){ ۾'ӳWYHHx;#PI"ɹL F)lh;xL>ߦb>ivQn]z81~%y%TR\^v)xznp߼-m6 ݃Ϯ^Qm}M4vo!ie㠆UͧvŜV9kd\O4H"`jy?3b/t*P8S^U"LޘPF `_MD^4's 0,y[ pg78Swn,Qš_ehV38q=]] A̋/}La;^ v` 4% o]֐nL36Mn7 +ct 4NfKi_O3$ {dh̰0ac~8`" )V~ L6j8 s C*2cA!U}f4R!8i\gh_nse!U,t{6nVRP02ԧWC^>h`IF=O'9XBVS'[Yd/nxW ~'#\݋p%4zUhB~!8% pm@< cmU.ŤHhd`A DUq,†$>^}_OY δ;i@U![OŮ7(qG [QjRq-C %,u(㜆:5 h>)K.@w!7V7VUaus)C6,f +6zhb= .j)Z&Vt9Kcǘ 8kY$r`]N4GI\'qL:pILL(+]VH̍e1j,B'Y.9e%,vz!z}2 =ȱ{q~O.]z.;fj>we=#x=X, ? xg5J]Jb݆#uTQ)%E&8]lud$3"#b^w[zFshErB+Cd޹ۯS$Pi>6R1.ǒTb&B( [Nz _;Є X,\j>h1(]MP#M@ ذm&Qwm6Cyy~e.ĪGib"(~R88d#'!2,`H+YR&;-a6Fj;ק1'K|i9@H)6X₁zF NsJhtl0Nyx 6J޲E`Ogٜ͔lN E}( L31]lKsAk?"w)YB&rs_]A-FKwy''Uؘm4q ?iȀ!bAAۆ^i^L WGx&_|N˜z,nioSy_!]B YmzQ121cv%m*9 `D.*ו `ӡ gBxk1;: C:a҈ D pH QQ^sN$D%4>cAGQ&4F #Џp>s$K w:CxÄW02VKy \CD0(o_wzjdQ٬ٍKc$ I;*i]Цٵ?_6w|'.f;5NҪ`1{Seù&..ZpgQ^?.n\ˮgoz d{P3(.d-iؐtgDWF_bR}զ~=0sm/fOv.f,yT3i(S^ MUзT=IUp*;QZ(n=_W7ܧY_:ǗwnuК~?O .ZA]TX) MVGx2@T_D=oiTCi!aYVM+!I~LhO[a k~NN_"$†!"ŅhrQC-Gҧ"ǝVTJ J F$H,̣9:>wv1Ml ng׿3`+b?ZO>zI#{%dYQpa>N18I?bgۯ_HkP=y}a$ ' b/5 ^m%Yp֧nI 2˿|lI9۞U6o6#x׿ ` 18 0 ctа]o+sY>NN|'U3g9!K+sGw3Wk>fGfy-|30iZ.Ӕ0gOL DMJ:h## }| R|=ުsL2"m*<] 5n,w/&Bi30(! #,WTٞY0`u$WQ|@TSqz6u[u`֠^Ωro,`)k60G&2AR037|q>)86P.؏@h 7U kFjodCZ  .p\bamhlYآu ι1Z|:iՑA;rRr|õXEWn3'zQa#LkeK8( R$*/QvX hs!\.jq_M*kp`nGk) h8K8A*bI '`!H&*LQiK.4N>żXj(CPf# "e_D&p /Q|9tv-cZ/;IQ{{=oF lj9bEu20$#*c4AN/+mR)Jo,o8:L5yj˽M8'ٕE,(҉ /t(qcH&6\.5 z q"9f.~TĖE (NYӜ kOc!m#aǒ$3H a|)=.=CZC=pi*~Rds:Wr 8fn:$/] XU(i2T|ǂO%b&βSv_d1 ߂ E[C>+ u)J`I񏜆 G%Py>&]*ղ]z|^ G!š["[\l;dhhW)񗶜N+a%U&B{k KD`acvR8ZlҤ4˓y0)ZB1VdK {zxzP7TeYB6J.Wsq czQɛ]6.zsnLW CG}(snh]&;M|S,{jL-"D9n+0[=tyv㮟pRСiޖ*Vq↵\^TQ',RB6`AyZ͆NY4=Ӭ:r-uCO*3VH,1j_8eac2m+^?6*xp ƚ( Y a?xriӏXJs*#8` zԗun3Q~.Bn|o0GӚ!}q }ĒE/̸>!a +V&D qҗU}PJF N%4:֋/)Wd9eRF)w0*ʻg; O,ECTإ-ԅPEKLT 9(F8y"#CYD㠠BnC#@4y,,F[d(MrѸllx񓶅*|*\R&Qd}yY5<Xͬ_FSwS-WJN`1%4:(*d>ls'ljϖ8hx!{Nt1 fO|-B(GQkB/;5okU1N9fy¥%ZcNt/KhN߀9rue5W(X1!BܽJr|7*8IxVFb6.5qIs;$naE_QsےSŪD(o'S)u$Ho+("0o::ܖŁς,rMsKCk'5 ^ |ȝ~|sZ5Î)%nv>_S_\SCbx ^/ n t O\'.S}HҼan+ C?G]0hbŽw05mTc^_U0wk&};a5 O.o<' ӂ:|*qָ mq"&UP"X FHPcNfkG)"}$k$8z+vPGR!-1+g Gn(WE)BtH\%6*=XrUԸ2 TvYhX6a :/~uYCd?se}~7;o5aFW)!L\ MBIӔ8CWtLIǟ\&<>J@HI)6X\B>K)! C0S9 ~=Vrf拕8|ӥ1v@P~ikws;+Na(Vcʝ8ld,axK _B1hfXL)opQx!7HcZ1ltSBXL9410 d;8efY0࿃!`9Y JrÏYA%i%dhe10gӖ—/Y6_~}p0EHΙ Z^΃(YOۚNG8[v~$YYpgK~eo@] zwd9{^D]qcW뱹~O1~Q,_s-Av̴ 0#5Z ;R1"@sPJ+)zy uQr5Οjup\jrfo Nk ]h8cX78>1{ߧߎi|d 9iu*IZ=tm $4 GٻFrWeX|9;=9tlFv&:̲K$K] Ꙓ2mJ)]Qe&He}܆p 2)ے3BMG_SFge_Y%4hGx7J\%е,DX x1Dv_չH3oҤ!blPD1o$ KNݷIFg!8I1Qvɩ]@B]@=AYSW:b-dt /3ქ*|!cmF`%$C+Xs rg؏Eg 7P1 B!AIq0aBIyaj?ˤ modcxgtdn\X$I n4]TƯ_q0ISK  X&\y #QG⇵B7Ky11R%P~,jHg V";]r3ܕP$XPM] VV72=XlS+QYe:Tt]Eէ 3naC%i6X^?NwZ5V h\kBxo/ ndéɁnxX\+Åi}WŻ2j-(4]Y$9OᜠI<'{R3? Rzp|Zt8"(:EBxIaⷂRw?L1߃wT_؂+r0?6u "KSG9 # s\*l-:@TE 4:xx1gnDKPOYހޭОLoB^@ NM&@YA$$g`I"P4,NΞA5)۵Ɔ {eg8l6pt_13ޥudx?ٌJQ+²2-F=j]t)]z~LvY<ɲQY0԰2g ]w!\gl-5JYUÅ872w*Ӫ7-:|Q:oC{ꭱA vi.h&FXe]}>|.1]?e5fx>+|Z˥T5W9Sd  |ɈQ}j1К_,*Rc-`neh!]J*X1ƲVˬi #Mڃk/}v]=EZa.si}v~.YKp ǚ)= \2g3 4++2Iv@u%㯕z Fh;}CWcnkz>LI*?խ EӢx|a[|Yח>?Nٌ8gyIF|[fmֹL>nIGdMJ6U;4UTt7o#;$G N-K߶ Ƙw\; 1. {C t4\aT}AV6f1drK~_ȸ°2N7B&#4֊x/WE~lȭrΧ9nZ]C j|{~~,Rv@D.)L9SejI5 yVlg JL9s4TN]1tKWE= 4`RBaPdY2nc3i2le\-+{%ޒ|m˧/'BG7ޱD"(6OCsPi˝Q*'nP2 \ueSW7z Y#(t7@v"{^]\ܜ/-#:7Ō@g Tyҭws*x?SB㊙ ш2Ҝv.PQc G`F N蜱@ed|*<-9Vk[ ]RIEAT׆F>w󘬂{~M@-K; r#d7s? Opj8zNܢb+hl?OTb5 }6IAbȖicP{JBzKYJt"T'1×CN5kq2"UE;8">LE-N~vemϯ5G;ŧx[/(2w3R~͈R< v`cqGѷaTIm%!H@d$dp)4׸fv2j? vEc9*zܱmeM? Ҭ--*oPl]XF}yaA1D/R2kTkwR;'ia YxAZUnͲjYy4ѧ}> iLlk ]FYs ?dݻVk Ew?} -f]v5[bӋMW~3p7,}*ZhzK.W-7 3м>h&,&=ߟZS;J:1RnV*sf k,3b6\B .^ՙFqR M]&K RZ@Zp+5U+֏ 2X5(x<\fi} {iʿN_VP\jct7 ždjqBPn-a6?dw7 6-?? _]u(30(aІN_W7G ?hk7Y2oaoNz?qf<ӰGos0R7qNZgyv{ by˻׫姿Yo"Rdb ͯWTf("!`31%;i~d8>_JG^MӯW+\5XiΪBwb&KՌRSrF= In$λ(sFv^)?5s+ h>lQdn=~y3Ctx3ގ=9vM˚u׿yt HF? |K\+>?/>%50K$[4`p_: '?ýJ}rB )ڪ+'ԕڞ? 㠶pbZLY.T=ѝW}ioSq2l$a ~.g(g(g(gjL¤ yӌ e]Pq)̛T) ģGXo s~MzyzBil}^]B׽ʭ@lA-MF6_\pzKv~ӻn@햻y| p/ 襂qEɅWƾZi "XJ C}(%bZ՜!y9RG4EZAP$1eVLR 4+왪 "R㈻{bǸGΠ F]b$V VmE#Es(ƒ½F גjbAg+f(ۃ⤔Ew !lQe7m|g ݡŚ.yjt.tW(VhǗnAlu]҂+h]HO4:sm۽ }atɍI&(HHI%<ǥ*FISDziNX#<0 ? 4c۳~ ]5E+@A@9@vrfXdNn`?"yyd#PhRڡЩ8uO\PZF9]x2) /q8%4=W0rsw)v :f ȶSu?}:9g.%2@hh[ɝ&ginwmmz:8m[l 8/*)dYlO5EJCf4F0";"B* ѰeQFsSzz Ժ%í{H m [<]C;m{4qtݰ'"JyY) !Ѧ)C1w[S43tImeMPO5ZߦdžB99|b`W^fɁz<;KY˅PPyr!-oжa-$RG`Ӣ6b}nxtB7#_|{x19jꪖE,'"fK*EHJ5)k$?z 9C@d[U%ԫ/ou^ ]a=NI^"PrP%aL ͂Ro_o[Lip 3cn#P{3^;V,VMjMWY iFx3^lo۞ϟmO[hiZ0CaC\-ytxN[!E jn>}"h5d 70:Y\LN;}r"=jS3a22+Q{61j,l6feS QVO?NBq\[91FKVg,-\AHa(%t1)'ku+k n9+aɉcރqtfY')|& 90mnЭ;CߞÛnHXVVךi=笑(l v|<γU7HhRbmHo?;BL펲fFe9+uU(VDsAt,;..41bz81֐n~fI Zw '=h)oV|^b֢PтxU 1d..'iE= ]?R}D E{=]\ݢE.3A42ǻ9W e@. vP R435[. _0J=^׈+KdP%8]^~2.[JB5;7v+'&2>tEXRٜ$2Lg|,h4d6fԉ/`MƢm#3Nޓzj&\. T8T}γi=<ݘ?dk@b[8; : IFx .S"g#Q+p2̐WobbO$< 3{oQhܝ݋vu;6uErjm Njڠ"ff#M:)G{:+k XcѦ$UΠU5uM_Bmq GpN1wc0o'zu>{^yWN3ch?z7;v>[OԂ7Z>K2 ngaňP_.\݄SY3I +2)B%6OsFzz({FbKڴާM} 0-?+d`M/}L yLi+r-eA$̧_Kd2038XsP.hlܞ_3+ϓi`MmdzDpg;5"j2_e(l\{- <*'g1w1xȂF3^ϓd9\{pO۝VE<CpvL~ ͼ`o 6:S^Jpէơ?]|^p\DCȒ%.ȹMe=d"uOJ}4 V\sZiRR|w0BN,/9k 93d3b"mh+ 4A;;ӓOqڼDy:ˣB׮Y'jsRgY:'uIsN3jj-fTd: dINF:2k"/IH%+jڗ}%t)`d$"HLDפRI"Ϯ'>Fv} W;$ {:mh";}MsIJU_sqp1?:"&!rXIk%AwGQ ̹pLcA0$@Y&MhϊM'M?xo,ètt9I->]q鸗(2"YYMѽ${I>tF%t ,yԖ3 'ȲG]MҚGصhΔkC]6N9#*;uu÷Qf4In8@!ov&OZ%և(=#@MBNx \]ZG3.$G!%R:a 䠭=v)x 17')cEt< p#S @'hE&⤓ɘ^Ck#\Ӛ8XWz h+dT6]:R+c]ffRa/ Ymj}Vf6o>KJ+ܐ}At>ůf%Y'b6@b@29i_"BQ8 }kZ[moדm͵BY=@J=E«ՃKg&%3L:d{,fEmn%~HH %4$i䉃DAaȏE^vCsp CV9,ɭ$$qx>Ό+, m#Hl Bp[i[Ũɇ%K]؏ͨ蟕1%mS_ )TӶY<2 EUlg|WZ%v[-6YmTH"3>B9qUG j5`!%GZQGb6q1F&T褉)HN+FGɞ^7nCpH =,/C/vFZAkȩ l IZx@İ8Z9eȋٗf[:zqikӕ&0 v9Q'}u!a:Z4cIȃk*G]+إZ9 AgjTFz#]i 䔧@ QxBj,WL>FR&\Tkt^Glr̶<3y?ϫw-F_j@ɵvv|s.9ͮG¼tuxdvޥs?y7}稾)ԀM.Sc ӿ^uuuĻ„՝t|ۃ.#IC "v`B`Xi;s0@nsT!<3FHKcJiZW*|4׶vd[iRdr($;EE%e#Ox?Uc:;~:Тy#z?fpϝ xA0eKOlj|[IiA+yvmsCUu' A-!-vWt,HXBrP#|Kk"7yN?C\ίΧgV:/@m-bfAxJ(@GeC_qͤt,sP ŔC ήʆIID6ְQz>c,07"ϻHT\7ɼO v ҧˋ6 ]\?ez1Q3v$dvYic6l^PEH:*+pH6" ''SZkk?lg mS|R'$ 'uBIsBr'Zs"M^LBV "\E?A"k+.-uO_B*xq+hj}h[}`iŊ]obI `}SFۨ}UעwG+/ϤRZܘ=?Z;S{мArqxkI)d̎ s[Y\GJT1A|~@] BRmw]¬K}ܗ'1ơ=+.[(YǦfG_~Ck$ bW6Vj&V+wHbZm5Z+ڍn# >yi+݈G_ %q4F<:͊B:؈uݬ %GG@ Qca7+{AQOx!/_SPqׂIZWP>:Ch tG\=Z>W;EvMyV,lγ\}ZT}=WNݠNTasq`\^3hڽ!hA9J <+x*dvŲqNK爦9~]C̙, io_T7~?Ɓl|宛Їs~I_^ӷVNwG3y~3-LHЧh;?m|\p۫+~}%r ;d49s<ϼg3R n^搯xMC'1Z3DM kt -xΊR\m40,`vqɀ$\!6D}˱u2oɝCZѤֵ5 5C;F9⻞,?b CH7cNG.,9z-Sj)T 20ymD{j+W }R%5UxwݚS>Ca&[o0c 9L9R?~ju,֛;nJF-L73x$ b׸rcYz5~~V)TdpV '~Js&eB^e' cw94忔rr2JIBC#5 LA+m2AL)#8 }\pPD#,% Ԍk?ZՖ1"Hjjs:L:qCS}TrvA$ǚKѧn{KK&MJ/;uqԻS 'p**[/D_֒fK\˚74,(!*@s`6, q뼷` 1bg_(-9}A9lsl1:Fb?B_&=f^M.n~X>LgSIOڀLJbLgh8QN,ev\5=1M5䎢L9њUdN1#gLUQ{$w-ζԽ]fD\E ۬[xڄ' a8SV4 B۬HYu ^׫ղoYs?9$,JXK5r\4.DeUmfkٓ9c@.$:3w "l_46xX6nx30kAޯ;Fb ={kn)25V?iȯnuA߯ϖ[;e/)Y6U l+`dJbUf+ZFKbxBmڋuiԬ1fJiCvGe.|J%$w!cC;w?BmMY"΄a>_74᳠x: Sͭ]xP)HSrVf@ȚPQgT/)bS6rd)oeYa&ޞ{QF2#۝zMb @]2+R0})kZ݆]>7p~[Nr~O1{&2Έ?@J,9nx޹:SccO~c_J\h?7g ڜ{wɛ޵vGMGͳ};B'[^ N Kza E 5ǧ_LL Zzdأx~VO)6dy9Q-\Тf&ZP*a_!X3k|j/:,C gVP@r @cbV Mb6 RzЛ=r3-mfV+j1"yEWk.o!h4]W,~R*-IfJ:ľdJXA휼n/ênLJ"?i̊Ģߨ)XTTľSC(LM-5f bt-_1:2D['m>F!o6f>R#YEͣ]>4|y$[,ȟ[ $3JaOL/oĔ |V iuJ I[E- Q!R:XZEׄƙ $T#0$5,!,V^-,5|>|e<|SA~j?>α|lMm?*o"N֫zw){Ch?8|?DVa*^oYCYt:^W/t$5a9~t[*~W#kyv_ ݞ[Dݧݶ[ |{&af;HȇAaVdAϞe][AjvYPi{)NsBv}5G85dƽtܚ 316d%&ofvw3n^g][ С>lQӉPJWӤn@KOĭB9#cQM4cJFv䋓 :kDU"RE%ꞀֱSchDr _l~+>wx#=+:<>8);wrR#DGLif! Q #B{Be-Tt*Nuq5ҍ ,м5w } a'^L(,?^+0o*캔 Q`Ɣ10V=)߻[-zXyˏA-ܬq5twPirj+ P|g|T+O.lbG8 8ԇ:;[aUg(b^SGFyf|7$ջOuASrɉt!{8ɪo(JI^X]C@:638P##{Ώ4ӰxY=/sU}uvGv&U=1+G'6 *=_Wu,wsurصR/RtVTEPJ8=qD ݿ6;Of,z;[^ݎ64woFܳ;;1M?x[Œ?@Cg=e+Xs` vzpUx#vc h`{-Ԝeh9ր1bE X{Gvh 9\(:/PcS}h{X.P>$1>'uiR>frc^lh=c3[Ï{xhR zid {~cazM:HHG/y#=躺yhX>iWf CvbsR#!9"FG)xԌ]3/cll .x_ֹZZ>$(hhZ$Qk_ 6:}u;l{|}KG^^F*d4 Qۼ.r5qH6-W0_ms޻~\pO{v``t6 ?o#wɨ{B(w׼"'ug: <(܏;+ ;9|#{- U p@{FY{K1]r Zm0ҧPN9:~$s{ś  4e#ѱ{#- Z1˻[OB VΗ]IE:t'4LUXNM m*x_VmJz̒(\aҨjy7ɓgF9X)azuj7փdܛ昦}15ZXI#g#LrT,j1v YB+0eH9l*ִ#pǛ?){2O\:4OoU)V^ܽtڣfhn[T_4 >S:7+cҀea.<'WOԱ![AK^,2RJWOAS0L/.K Q7Ů7NZΦİU V[d"7"`\Zi1y_mZh6xȑ{ DJQb}]5udJB1ն\g _SA}emtP[)CqqkBT BjDƅx&JNMXWy'f ##< ~-њ={>W+4ԅ!-8caLS"d~v^^it4D)>x9~j?&C QOki$ fs/Ckw(̦r7kS&=h l{@4grM]N").٬CW$PqILSJ5bZ>N\`wxq(S1ʏ[EQ: g~'B  tVpu:1S$Qmd޵5q濢kNK t5/yéCʥꫤ#ɖM9qU{м,D.ggwhå?"}:^8݅oabÄʊ=*}6jpYrva뱅.lq.AkR-ĥ꣄`Q>ǜd4[ڜkd1>CYSN}YuUd!)>*\12#v3 /6oKi6o%b}۴P%vH)2aS]SwסT{fz35 ވ|,x0ZEv?Ɖ#F-+F% .yٔ9TUmkA޾Xfؙ GwoVZK kgAS ,bf;*U̳Yt̻D5'::%[d:m9lLͨM*qߎ fAdt.7H[;t2hO/d:'W4L-ӹe:3}îO#8ei{@=iKOB=4^&Y8aZg E< D#ĕX,'c[Czzᰉw98 %HS %@:`5>Gş)E|mIu՟9]BPl0ݦ%T{0 (yCX[q-ȅ{;xXa+>Byt7wՔ7Ը9)5nwJAgVuY##̠ G58jh:6GVRuM4tjQGXs2#)cjirx^c =IҤ&RuGEL d)|dN.,zE2<Gck9.3H$~ݛSw$n'j吝3q[<M*/L٠6rK.7 ÖjSSRqW"Eu[eKn{L̠xWp~ZJc'vvɳ[|,ЕAقckA޾gE+Oȋ081ȸOI͏"TH:yZn;wZY<&& əB}Mds'cK9A?Dp=bR>9s|H@*d*mLBr.N[erd@zX5iu`cMT' ;f/6\ VƖlPa]i6&r_>U$X3D.F`gΫNlBCA$h.ޑ͘5sO\*ѡ^rIU!> K Ӏsf ?4_ZGI/H~)p+ ^W S 0f+"+W<bVdA=h*GmLAnzJ_{ ǣ<⁉`[fL*z8ԉ[=9ojn>3szݿxyz~x/'{xmVఃmy0Sf[pܾ+aT7w3k~]v,9=7 ̬tTEM*"_Uw>g@|xIp$4Y4ג}bP~L:*z:1ޫM<e/' R HL+` z4{O }I=ɞ/4I$l-/%pˇчD h+~ܴrlRYу)?ǒ}3|ΖQ^nisH NA mfe+lOE0;;Nd|{:Lz 1*ma6hh&f }Ζ%h6aȷmF XƙA{L_B` =hݠ 3& \ ωC7hAEQuv9eM/ڙc r{y#6hRxak\lo.8mL$0~!rFs[mT[RoZl3 ,v*Ж aV[Q*H|I(uymd-{Lx;8wg('K<{Aӻ^܍{\zIM_ן^\ ~2{+ȓ[!I湁Ny9d=爯Qpvs~1QlOrnSS0$ }ODl;+ /7E -~?R0yўo=KIT3e7<8h77Evt ;_-պg($BL~ R/ a"0!t DQǻ_>ݚ/\hP V={d`aο$ &ꨑx )RkI I̎r6ߓiJKG!o>hj|I(cf JOT=W<|-Ҝh^.s%q!fV>{D^ XKp1b+㘒u S%7l 80 iTzSw߃EEkTנ}XHh$atR׆'P(Q%/\cH0ul*O&-?.E;xĖ؄>A= L@LA󽓅 MCHc݇ypjIݐ9%S44sWQ]?/t75τ6Hg`IH;}H-`YUw+Y{g‘$⒲]t^Ir"sv Df| $ӯۯ;=X/gInbNfi}K#8ѽ9Z7r7H0V#WB4`{ǁAǦr[ *!S]PvHtR,Z  vR[)/V8zQdOl.$!d1m~C:GrWw~rӧ߷%=zM6'2+1a +?R/-Xb&-P3X1גJ-4!`Gd0/QiA iNAk6=@oK'Z&G I=eҪ# h )]iw~AzZp%Gӗ αх7f~]Ng}ɁmT)>^Տ@zsC' M:_ Cf ;>_Hr^T֣ɋ͢_r5t.5y ח}κy񣊘!/Vuw6رƃg-| zh p5;aczkuw/6(kǩ*֔Kqkb@Q hm:ɘuj<7uq]I Udk*(Z{G5`aܸ B74BV ]{ia(}p/ڼ\Fd:|$(3ٮj2Ƈ,VQ,>F:%؃"&ZWdnVg-leKAor@GL A[$!l]'zk[LIxݘ*=d31# I "ZN de/21V*4y,G`=Yю Kt|%ߓ7zxW/?{ΈMȞKun )p$sN۞<7eMr yƭ zH }ezW$W"Ũr✉y:PRf=AĻSҠ RCmth0fH,$!'=]oUXxk6-ڼzT ^n Lx\}&<䄣nWGrݘ)Z"Z}08̊(YtEqN Qx=4xj4=w ) .ԗ㲒hD\Y9 bߦ9w%46CFbp I> =MmL]d8Kv;0浀L/FOcFx"zpaF7yM!bG`9 Glgn?68#()cuα;xEtV Oxr?U742h Aߓ?zBQx>KS $9ٴ:ɖ,X*%ꢳNM2cTe,AkGéePW,[Ps7܂#zyf4_"9΁MˀFQVcKc"^eY=}MN6ԭ"rJ(Ћ7R6>ka-:ol7m^naO ;$c ~ZDQSf{y$WޓQlv,ؕжʘ^U71h܈[b-R$Afbd|25%^Q%4 loAZh+Z(Sκ6-2&[EB6^r+;29ˆ##Wf"6Q-O:f2Añ%c'1%>(G5!Cݨ,=(?/p~ ec>LDլ ٭2*5ߊ.gjlLhbk̝ :&8QxGD>j==Qgm6Yy5`, xwv rrШ0^N?5H~M!jwٺUiEBԦA%{}V֣CL** eEXela4ҎꡉjC Mk*!:[\ReMר㯗!ScwϷb]HgeKx,_r@N廬ňT%`-tf,O!(XtHuĥ^[Mee гrn%LBGq+`]cݬfSM׳d@_r9PM8%v#CjoJVTlDߍ&Ȝ eɽ){o?xƀ#\j%XsD1Q̢ͫ1h5sv {,6w1\M6hnjz'OGH!ݘz4s,5rifW.yhV qh92#}:rwS*s1Xt5߃]Ns9Џ`9,WC/-;QN5e[^FlEXyAG|{T3l /Aml^}`pKsP_/~Yyr#|zӜ\l?-˼u}/51~Nל+=οID/eG|e{Kuib)wtq ߝ}"d5ތXV/Ōl]s0%|hBꖕũhe'JEeRr/;0=z\z`c&zt*n#0a* i'z.M7ѣgIvыYxKz ylMh3ഞsvy YO;S0JE*B@h8t0Uݰ~Tknv/^M;yi0z *q\?6?] &3A= Xe<͒'&:!af|U%*:,q!LuBUx*\L)Ꞛ`@\aN^ ڭoS|mf1y{S~6>¦yCņFoesiSpK .Xʁ *wW8L0iL:5o~.PYB5ty ȡ֩ Z XvFq5tHm/d};P`dĨÐȱW8YC'kՇlS{jΆLU5Gj-FLA;7LD_:i$휌ȳ_y@'h7A ]ozZkВiມA⧏[R/?ZSs|WU.~\_ξ~c99B]?E^套\T6,q?Gv|˕jGWc;|)'mg闚{l=5@bgzbQʦ\M}G{/~?gjٙ4o}__Qv._ehuS(Nd31M3;fhypW'K~vvgf?oUKsZ( xZ[ߚ~]85^yOFE WB˾wU~\fÏ?9y"H3Y\kr5ގjY8//}"-&8wϵSe<-7Br<}q`z f]`Aҝ,v,lr?oy~7*hQ`ɻ6u/_\! k_{p`Syb&{]nE*)UnzG;]l4#7$Zٵ?z0 IsWe|HՓ?T9$CSre1EEزBYുI-=Fy=4644>I/–nZz:T1n99ߩe)#ɿBm2R?"fȞwh3؋R% 9$EcEΣ*eKfLe{i_ dOx9 8/`(XPLC¤Im/3gLXiV wLp3&F4i'$f9@jT]oDUI ֢VipO2?Mc8Zs? :[BH[9 :ҡO5LB{Ei1V]0~j/lF"ɪXXJk=_bsҨ_,YCE} +msa#cj"׆&24Q_em)&^cQw\L)]mh M} 4@h8B>W6S}F8Ԥ0"a`29]* {VֶoGSP|GN"8ZY;&Z mmhZ#jn =U_}F)VW 6Ae`KqnTzOV\6 uo =4IC:E5_ZN b&E%sR6k1řϟtjbJԇ w2#nD!6zi2<\!h unI 11k1_`.c5Y]`"D9C>DO~@3nj[gOax"Bf联&22rE/]4p\G(#0zSY3üo.:'zMTdw_ {ٕO#^ !R'kˀ$A b9< m71]/MGW7/noL1' m0`"z"w\DT3 y0? QZ0kIe3c=ʪ\*qKb auI} Y\=à3)tQ=yq;Jrh[1MW]u( $¦ l|nf&MrȈΧ03savO)[A)hpPrh5ldo՟ab՟й1OR֪8OvsU{n}\Ok?6,ȑ}965KŠQ,iɤpa6̵LIÏ%6'XԧX(5X,Wg`Mчg,uin;(V=Ś",XŴn+8ĜQB˲jz|&+:%oÛq,养3ւ:2fR6ri}=9+ݛS_M.'_w#I,]MoMnD"!=*^Q6|vh`*:ߛo֌u i 7ɛoQ)fy(Q1+B! ^Q_(ě<һܳ1PL7<[S24A^ysPA;p5}ZFEL $&qR2:Jb#a~$|KK9nyˋ'7<4v>y^8fvH_dت42Lc.c >y2HƍC4^ċE"43R6!8N:p,@ay\FdVt~ }vval,,UNeNkx>c+IY΁(9#0!ӑG\9EMV[7|FmޭkF>[| #*k.&uYݥmTw]n;kqk lh6ۂW=yrTaG@)m!b/{asL*ڭXE?CSj%rD&'NhN1sSl%뎥<E&(Sm5=g"~(oi^u_DJN zXXJ)t܁1.e:((¡q152GuC>i?0ermꌪc#r/H*PPx˴#fX(IGB4$~cI,SlH y8\(&paEpd r'A M)[q! *E4iVv |4`!:rO$+D$AdJE?#ui|iO$r;d:Y渑.s-A<0NKB(, 1^H7M4x5ҭ16bgS[qjhoԤ*y0_'H'k.5$q^/@^7pe<π `km-55Q)kh*В6.2"WG3a)S‘)cAIURftY(td(@y@Tf%quZ(Le'<|dN-1aM2aKYosj!4(cs QJ/, D}65Λ`Qq&W`9{:߁{'3L cy k^:M4fD\<+3,~Lc/HuqwA-\ )y^l]|/ 'ɫMf&v9,Z#X~ۻ&FN/g-׆c2n>(7Û1Ymga8n|z;c<_l"( XHC^GrNd"(| X!ڒB(vo$m'5 Lj{P(e4Ĝs2-Ӡ1:x̭sB$''}/:#a9lqގ%Ka')0gRA 4b?{6=l/v^^l[w(";vLYb[$3<}׋%oKȃ{[jA!\pCQ`$`F<SȊ(CĔϫ(EP+l1.%[Ei="$c% &" Ҕ({YJVYoZxs,[4~B.}9wvfv5v/?@W̷LZ`XKX2м^4GK3KB8:M=b.`D4ɭ 5KiMcq)=B~ yjub^NL+J82=X'evd,|_yU9"?ԵG p-J H?uvK+r,2[F*<OFL9*iYZՉ_I܈_& 6U bЎ8F&9aI5.ArcLh&q\Ri 3jzܾP~P[2U)B"V k(uKI)6iO[X˝$%fZy!9gPQGKCiRS=&h,141*6s8P~({A(]o'F+"uBЎg7AHOT$u,6MH%cI)C[`:C92ce͸dp *A0G قΈ Έ (;# Z@ )F-R8Co)‰8ARhX4BDsa,STQ̠UE|C$ɒ.̉ CZVj %!_o3UxEeC~ ' %Qh?"bT8ۺ C] p볖cYK(GAG\%qH&WdP1[wf]5ҫN/Jq[ 5kw5q`*pe-iiYŦb8eqWj_ ^T ލ fPSY/-zY%đ #Ul_?9\ΑF-q/kObT؛)cPk 6hJ7 _1+>eA$Ca-Zu%/`1#+7lY7e#X8"0PIdDJ)IpMEÇg}O/N ,GJ/Wlg] FxN6mU@JbQr ,]:3ͼ~xΙ |rֻ1\us9cxrć ,6/v ce4/-aD1Q*"q(raBjaK:96#M(u/FjBIDr"u,?5myk Z3{Ҍ3HkQ[_k+jxe5*u=ڬK-B ҏ66H`#]V;#sg벶?=|mtۗ"%u*iWVի܈60\UW.y&w鼛:xk3?@e *#i)(f(aIL` cl {=cM-LT}aJ` &NR13}ا _?pM*l%[?DK+jFӬ۬pƬZ,1-"-\X pE-ML(0RkeV SXxEj"$0Qlc-M }Zpa^ dZSmq}^&׺Pa6V[FژfA>ffUf%Hb"$67?xN߉;~']5Z $("4&x`l萚:sR"aEm0l⑭zРwM]>\κ9ɨ[AZPQ A& c -BS9w 4&$3IHX`[(nAE#<+0ԽdHhhƪH䀤u YbRIe{.~ v.'hW?N*  9ZmPz *Uj5y, r,wC d ]!JND[y=$X?}0kQ*^*hԑ\+ϒ`s-:ky##@*UR&"|i>nzW-IDI$"QZQr_lQR)f %UP*+;TRJ>iOOLjio*/͓>-m.լAwDDb-#ey$1%ZJye_zk9XqYje5.B|o= 6e* Ì[$6F`"S9B),b|ĎswM &S[$([.Zzi 5&f.Y_w ӳ3>-zO'g;9[_2 _]<7RK&J11!ӬY[Z;~ӬKy_1Fg?Ì`S`ͼ'p)NBz#S,_vҎ8_wmM*F2]}vz#y8S 8p$bD7Vw,`!n$Hq uy$|RH (fmsX!e&OP;S'HܕjWCNM' (^ut-ƞeS$?e3xVXij8r/BQpaDj0;+L}AwK3(q78xZ H%8$1sDK,)- y} 1ϟVId0]<黬WYX{{]Kϓx~=_lUC~uLX]zf~nz5oYT>q7c\o=_t| fy,( _ /Ooy=}-Hj:#%_ F_N|=B@#Fa ǣcBLx:?,_/Y2\wͧ0:alX,16(&L߻dh`"L"*T#iCPl0c? 'k |—D[ ͇]*`d`5wϜw͇tQ,Rmoy盳ajX9ٖecmo9iHcTL D)qR=˜]=N&:fh=& !HxIC"&]*}' 4Xmz:H!bF:EP=)j \Pӝ en)$weqI~,Ej}ذW<%Z$!5n6bU-audF|WFF-m0JAL71VZeP x:Tϧ|Qtq¶veYQ²Kc>܉Y rJ'sk\.ϜJj3ce^\:v&ux:;2n5Lۺpvm0,xl W)ڸL{0s=k'ؽ`Kig+e7~ܯliik >կYsP1(K5cF FoB:jpЈ'rYnuVMQj n 43D]Q;v7ekʔY߬kh%IRݺkp6zgc ߕ8Ү$.*&Xl$Keb^SnFsX*]ӥ_etFd` "c^.Jᤉtq4.dNR rsص?=X3X*=Tklj \јFJz"=ގ*T)۹O6YҌf}MJ=8К$W2K5 <4!@q%՞GLA߉xRdžMA," ʯO啫S*?~kˍG}\sm9A|R 5qH̲jG\ǡ.PGd2#(GSc(5{pY?view[յL7埲,Hy9UdzEEEkƞgfY{V~@9_9 #{[y吹<b(,y0$5xށUtOe6//WU$N2)GwNɏ:b')w7XkYY,yxj/]qSf^j@T6[C}%ċ 9ǥ)WJ@y);!pWYwd(f K!cf\Gx^9Ym1bH@ A^7Jh|fVz'=u2 !ZqE/׆*9*ۀPQHВ!JfDNഎ$ Ow+#Ə#@nί05PԉtOh012Zȷi7<&p]2^hL <[r֌"s90d?#!IkpF1?a4= J<`忭 WB\ 몤_Pj̖{ yxhҼ %)*Hm\ȞDKI iҩZKC;9)ih&#xRkV4nq\w5҉vOJ?ir]Zyq% X]/piIpHK\*β4CUɻ.R H Iy"EHRǩ-OBF'WT# RUJfeGDls2zA "y4d 3$T5ZxA qHQ\ !"/)>UE$_oK)뛻iT熻,X@,2 cJ+vJL%i9$Qdu:%$҂x"aiRcHAT֨zZV!VEgUp[9XqI7"126ԂKJٌV]QК `ȒtNeQ*LdzHs 8O-6ԉvO#> 5q.xr@ObNKqB @gU%9H)]rZd HH/*%>O&*=Ȉ^[X"וQs5x"*Ętѣ0fBVQ굍# 8("WIQTDbQv3a4rF]U޳ HWYIyw+;tV WĻU$Bz +\ _j_MZ~Wz/_}AՆ(2'/h.Z-jsO w?kpwuEחm0F=}rܘSE1/V۹u=[I*~iV>ĺ9m b9<`RcvAwq8j q! [58V@4jJuJ!Ҙ]Tk:ٗp܂Pb5Wmi2#PFц-Tt`$Ƒh%8jF vR9g`.5?-qeKmp[2R̚[OFWZLIIne=me ZYv<i&&7Wop6 (M]E}ъT)'' H A Z$;aQ2y'r5xrLKE2ըs}$8*)Eq&Y$ꉲ3=lY%Lvr2ҕeR,fzIo$K똄ut*A"jB\{aLg>%K"OCdO2QMlFQ2֤0.%qrN-)`U倫+, gF?b炧:/X$|F[k<^X6O1E@z$cAc!(OR IP[CN07\C2hH+4i,BK̸OZyHH-zrx)&/RDR=hM4Riy$y8vG9j{= zﱧZ` ASk 9"IBhTȵ&2Edm۞)Kebl'SrPcgxh͘^$0yU᱓VSer$dP OC=#CsݦOv)yr#m'z d'q24M34ũ5} 9Vڶ*;*zbBT"XUNH2{ڶ=SJgwG{ZgvۡGY^Jhl8e&Ųft葡Ąe %}dl /6m{&ɀS%v@ʑls./5c$瓑ф$RʙC LȳL@#@f6DvoӶHHFq``ͭÇ;^9%H.zb9%D/wN!i9 W^B,[AzAOY%Vx{=m iAm;Wʤ<] wz$of-\̠o- n~,-ckevfznm.-Zzڶ<+Sa8ۭ&G2CMt)hcM[ҟBnz=hAf*#\w!8aFv16m{FMَ#ÏfC'o~c_/]MbSWhSX"g'k(֣r`M;wiw|v/2s&%3-* {AJ;8!Zn< _cwGm Wڐ-oaː) 1vź]S?͊wvCTKph1I3:$N;@r$4^crIr\<%7ۏ=ovb!P.)՛䀫ur~Unm# `Lֶw1R\YJU ,{$^̾;(XMjE~HwU"տw'Dw`=!k1Α˱2|u+ ?16v>6\)#3׾3MLA)A@ε~_ݺvo5H>cף$-dݧ``[tPm Q.m G\\ȝ|DE}>lIIfrCcN>=hvn=zx 31 Wat GDo}dNdk齙߆߶'>2o&!/YnNC|kq.G Ԫtv.2ھ.̢Lw{{꒔!g˜&1 bOp՞9\Z2gVXb@fD.͇:$<`9[#4-mGe2M:в, VYKi~E߭~~lJ 5u+_9I hs.[T3H{tVWώa#A 'Na+hxts?|ܗ+L S+%-`"̕>)0t'[{R:h<ȢW=%WÉ0m0%O6}] h=$g?vS Bκп)B>/r(/-3< ~w7Rv~Q{~@9_elkmH /~T Cx\fB?mƒȐ7CeLσL ~U]]]PVάEI Obf !i @<oM%弪vkŕb*:^G&-N[WcN~+*ԞVʱG*Mm;e=*TS'HlrQU"OSlgG"Gj .4)A2L9EƤxWaV,Zrq5 kXgNZV2fǎ9CA rcl.*s\@HU,n Xy]\O=Ղ)Eέzs%BDS?3_+c̨1*3b~#ze_U3Ս]_LŻdNn^RWffC_%|.rc_0OZ2Y3py< $\} Vj 文`g`c RF1mS\(a'v#IсOb!4`@9.ZD.d\9:\p}>:O B/w6~;y+i)^ki%.!HgF`x4"^2DTG 8DE4Hw Wg=+=Sq{ q*f"d6/_Ux?$,QgM\\mj3.e<JH HjP 6[j_驛džo`RH NnɌ3Yq |`2~ͥ?E;y3ooFߔl.R 制vJ /#L!d$&ΐ q#bM+29_ԁ;uDq3MuP jk:`#'Z*`5ބT7BIMPkbΌ egG 0xgM8!50'>NN$IGAbFAN _uŠGB䘢1"!2Q {'EsF\y9Ax! O='o 2Ok=gm , >9jAkZVcn1o?k=?LIPznOMۚ6_sԷ׳>|q5\ͤNZ|imfz,EFϭ.ē;s8*]KT*>Dﴲ%yzyy$Z;o}8)`c%%VB 7 ǗRfjשdiO~t9ir͚6GWpTyAQ@h~+zDi{ QbęK"Ef)n?6<>m\En[ml* %*s]qfNpvO6*E" 0TbՠL FxM&H"0o&`=p V CBZ6K+} մ tv! M8TSb tb #$7J 'aғA:c)Ta(+$F!" Yc  EqV;j iMM'e`HAu v \L h#ʪ \{!.[^Ş~P,'cDŽ67˂t׌wz)bnsKU= ѱ#9Ic 81ҧ.h%AZq6 )aNA*N|O8⾻rַޘT]aÓmכIJ\,q,ɽ<+m7.巕[cƑIcl(O H/ws>j}6N}`;]47w7}b>Uu\E|ўP5_D9^HCq-)=CVOѺb:c4n#zQD f݊͵n]h7E:qZ7Uk2u+ GuJhbE^֭\օ|*^BlF֭=՜P,Q= ,v`Th,QޓTa| `PbmƂjge[4 fY@#,COk׿R rKMֿgB`_Lׅ|*ZS纒Z7Y/s$yr?eXes/2Dp%N hrԠ:&S㜘2euaoYaeh c`CBbA8`] \9M-vAT-i2(aM:]%۵,ydi3䜊7e|xYB`U䯁䧪f|b3\rIo5kmV~լi -FU8iN6O.dtUᤁDUt|Ϳ I 4S~b ֹj EA:,C5 )ɲA&qȦ{# YFhe܈r+ ZGnL=ERy$N^dY6@O^A};Ň,6/)Љ5<⨎0.:!<}0EY=E`iAˇ0G9 6X+bLsD$N`9Z`1HhyBFKvng~ a)@y\ Aͱ J.(fs'dD{#8fF(TSDM  !Xa A,*>1LQ#1Gn @GLh@.H 8TiPAnMBF\LpOfBhN՜s]x#ZR rTt&mD+TVhuBCq-)֍@_AVtۣEւyzO/e4; UNAiEnلōj6.u˩F_*.o~ql7k5#}JX_/]9JЗԄD$7qj#bU@,4.,,T$t#.IcA; ¡HTSPIHVc,IXSct*6Zc5vWLH,P#̕"kJ˼R0ƚ2O륄6`)t 0}4^c]*< b|v' {Sqw܌Mva)MM717Y,,p,"ɫʻD-o,?}o]_&Mҝbv{-|ڎW0{^dsP}ϣҗvYbpR4漉47Jcnļ@'B=@=y!>,"4,|LgyR֛") wzzEtBSmZ+1 6vO9|G%弹Jl8R%-dKf_tct~I|} 9"xggCQ0dp^E֠Y \uu{$d.1S=|r)Rv,5@F +QRp )<;Vk$*Hee*%ԚZ㙐( \@p CL$0-K(,TA"¬}pcQ3F:fw$IXЇFn;u[D=Pc˝P\ K&r܋qmqI+?90'Qs_tlz 8o7;y/ xiB$:[~E<38ûqFF6M]xv{Rʹ+\)P&$VѮ[- El*<.fĄcЄh_KiJƫTm2#_Y'Nk.NootXޅ"FDpkm}IAGNp߳Ĩ~;|.wYfzJ?enjLv5\&+os_xWnY[A\6||"(IMマ=W'+OnV3s8kLٶ0*bR+Z[xZz\Z jD*+/b]ng6j񞺒՟'O hV]Y!:Y˲[U;;^>.VM3,Da _PHVV'kC#U#5~.{{A r, "AcOU ]NLb]~iwSn baW} W3}3LKDA3Q3?m Cwq'7d7HaZe9q}t5)@ Яl~D2IghKCS4(K@a /`!*&8QEũTROvkŞ?gapct>[%luQ_=W7z{oA=mbJihfv}0zS@{, ыdF) htRKf f @,,,/BV6 *@b%*TXL}y{5ԟOZGF IDUQsBm  "rUiFJ6e*9ԟD _:?M8'!"#i'!u. cCSf ((bS!HZrRo6'xjl'wx4OO'faS(}%%]all*cW*ٽ~'ee[Q!ҸI2l\mC2Q "H FAp {an&67fa1K31,ZY ZUrD9 b Jlq۔URGW)??5&g1Iphj>^[u^cpj OwJqLCԀo)%H1^v.!wYnqxk5r9hħw+QlFPD!}hsK0@Tp_(ȩ^zJ<#anTC aKЕkwm*'hh4[^ I{~ի24sߧr8cLp*'H ^$Z޳.:M<$*ϣ_>^\e 0fU&8qK[j(2J+Af+X)QKJhi&7-p*Jf # jm1& =x"%cX~AoDy+*`UiP26Q*5JK3E{(A+Je\&@"ysrk~qw]m٫U5:Q/>^5<60!_wo><^*{\M{Ao,VA?>4SԿ}Kcs{N/_Pf:ަITozlFlN4S={\eSmB޹tyQ3`c,@nƣB `re!d:0bOedIBvd$;?Aչ1ֿ[ߠdi *ҏvښY!{!MgД lFeoN#9N -(^pI%ˊ_4j99<JJK3mĨ%`Q?8 }{WFw^oAy@qrQ:4 ھr:ՖB7Ծ6Πu8#^m.3B8pȱN.z^nvBY=LHFb)7/DEI e:a"pyphq+Er4v89-0Õf h(ʼS;A8͖j =xHk GR#T"p&8+2bL Y ]$xB{Mqyn+c& ~tk(ysSP6~'PۦJzX 8]6VSJ8$z&~+ 4~s÷3%Ǔyͺ߻x I)2ܮ'h @jH 8 1ɒr l]wr ;PӇ}s?I>;mxb>AL^xNv}=%ہ0}(,0/cCt"m"M 1$Bex|4v! 㻌EJy>R5yu,A51Seۿ V7zdj^F?yc)Cs &נ5I4 "TQ EJ]%W"Xee(V [0-$9y)T) %oKW"P;7&%hCƝYʼne5,'kʺVJi.;+>f9'0CB0Q+PI.5WjŴeYDXKt\Bq)2^ WƷaPqUU9Q"b n0DHBr]_*Q8SjVԽShѤsHf,qA<_J[! Weby]Yl-V6UO@xaL>~.?<}~\o~6Ctoeenwُ?|}b3w /Ņ08Y7k~ܭml"1P|{SFGr䤉ӿ[IиrH!1l^|o6pV΃Iٯb^}f^-!ώQ G>wH <9؃;Iq-DPɹgky P@7`LLZո{O΂E yEXz'OtdDsU[ٝ9{RQ$vb y}Da %Pf,L6 k͟Bty3q;yϮw3y}*e`ܤ ΢$ղ_` c`L&C)0a׮ym@Ia--:(4IrM -W  o#7_&u7HR8L1v8WWBl榍Ne(<;-@D P:ou "(TTlNlY߱SD*T>`qU !M/vU S4KspeC?Cķfy}h^ѬgNg~wm=m\1ZFaJLdF|<6Rm7H)B_=8x g$Fr@Fkܲōoaȝ0) brJhD1}FÕ.OR cl;{a=' tOׇψ+ŠLd⚮#r#t^+@t_+уYqwG_[v>OM=Bf,~W|_~2k\t0IRj(xS>3]l0%-ؼ@S=ŹJ'p }׶,-GG)hIyQfD}*d =M{l[B H{,"HؖEKgaԸ3(zAHjޕ@; R]5ny% I[23je|ev”*$jD m2G>F'] 3 >}?{p)x?D4DoIw(Bkq&nɤ*q!$DcFR{1HEHQjIt;; ;ch @ Ab@ɠ>D#u 8G812|Oopaqp_.囫GHޔ+:="z NY~ږEp B&} f~ۉ4b]qiY`V߯t+٫=Af~|Wv_9|e;7P i@yUP%@\iX]isG+&ȺrwuBkZ*78@ nY!@w2 MYx3z6ᰟ&@sVԎަ^#dR-6:_Mq+?^9=@9aV藺>!a}6Ȫ7bV c$fF>q*R't&D4#S1$F([VKV~z3+:K~eIߧ-Y<9;osߧ猑m7K ؖoB3pz}q{EԌJʅA 1DQD sxô^ }\>SS*TQO@sFdX f> C%1JDéBk5cmAq"!.qW 1FAv"R,,|~@:_bZs*L]3=~<4ln٫8,΢ӑL T vG#w/XC'Ն5TBMu~o-B3?Y A\! UKC(bH%<8Fkl+"ao__X ;#5DzB%f{ % Ls!Sj31 :7O BTAl Gwa*sGDJ1L`6CT\֚%G*۷yyO*)P{97TR^ jrΨдY |&qV}T}JZ0Tθ:;[a9|?gʭhDGFq#}3 *1'-NeX,6TYk@8( JY-T};hCnGĒqRgQQJTlscTUoJOozlx5sYTl}{=-[ߢ2{;Cahi"K3ճ?M?[ꨳhN0K~8dZ;.q ]oё:t:n$PC˙)jCF}p՘ŬZc@IT)4Lj%& VvL}@̵(Bկ+Ao"kaiVW!RW.ZBX;}+X-DLyzdk >xUp@;;<) ZWvP~Zͱ>v s/gfc^'؇nK-c-C蓢y2ɱhn^&9;@i3Ͼ=;E $5\;Itv*yb@9s<+` FE |;n`۹A*G'iyψ<ۚ玍EUâ!9V6ln9"eα"tpum9?rzNw9ޙ;G9?"^ߥ‡P|{>!hv FGyUfby~ljc':;$Ψ_lq3ogsyY6;R{*Ƅɺ"^HozHojPH BjZWBU͝l|J=Fi[m{Q[tnnspΥR0)uxt]=}F>zڠmWiQoWzcڢ822"o.UbS/"N w"Ai<#%b8XRJ!+A'LJBǧsv=Yi'~^t큸[m5ꄹCoq?JhF$MtR \ vS(G7W?麨ןZU_L#P}cСv((FѸ>e.il)P}Sdp-G$\_8 hxK&//~Mp-䖌GtmO~QÙq>g~*Oi.*˖eCeu5 Fa^z\+=DWfJTYs,8U\*=J7WoR aV9Ko O`k CL&,J˸Yyʢ.t$*EFJ.Tp~4I!*W3*GZO+\yY=J֮A '̿QvNę,SFT9-J%SM.UZZI[t)ҪKE'Ij#})[?~7|w9WV`Q4gEHV- T>fBI9YdQNJٌOY,Ţ˶QAFTE1Js@Ԩ*֚FƕJA1d>AR:*ƺAnZ0VbA xJ)J%$*z#썸 l$եDe{<k $|h@kvz~slC 3r )], ^IgCK;'VV! llt@uf'4+PȒ@= "AoFy4:t#CX|&za5bH;d =M4hI P‚o:O&.sSB)Py)Tq@_!UO'>i1 t{0MCs~uSҖb!zܽxz"(-D衕xB57D~p4iI |[v7N׌/g,ӟ ͤ׃C7-.oG7&hNEO3+SfRRf15i5JuHs&<􇃇:\gnưk|sq zVu:q={jΫ4iϷ/^,ޱ!rx!S\˿fvz:k!-cHj|yrå?R,͌nA6 ɬx=iA5 n<֖+ý:,شCJQn&J0WX.l.ejk_29^#1B*q/L_I=8\b) i^ݻ<;i:omԒ`B MeL%ÇjPkL&V)XnE,ylv@6,(@T s6kI6z s=C+pxB EB?/(F0Lm=IPcƜM$Jq[grX8LL ŽK&4l&@DOt.I0)}l5QځjK5ҋO;?y?ќ&-طadIa*w9/05q:sXݾ~C6uaZYC3|Up!RN&FE bV$s!8ݛ @i~@|N@κE/HS'z25huj $O YfNT.`hzIz9]?embEߚt$)tձ;VN$HRN1UՔk(Ö0%52Uۧ)v1L0W?$nxJ]M-q3CņDOIYikQ!̱]m<{4r_-;z DY(9Q!##b 1nZ౫#(AG (y)"v/ڙϢ@ $~0e<[t=mdQE+b:&_:bG 6}UQ9,l7$q7sgNLL*,p8&rI͔c ++!T38UVNf2Q<$P+m}I'E5Q82Cf*whfh=4sR DXMUHg >^ߖbCv G<2m`4oS&}#QGpNjX?O~,v^NE"%KʑUH&9B2!{bBRez66ҎgtRU)Rx2 L0)hP:_q+?Lf.B»^,?ÿ?OOIЋQH7YU?ǃQ tZ(td Oc4|q_@Fzw:ҚamGc"h=FJ+BhZ`;Hx?"d{ſUhcHXPCaY?z}4c/OnYХ뾻IxmIiepƻLꇁ_._v|ׅJ۫> mFR ';s"cb>#- W+nMGp )Hg޵\DfVSб"! ~0f2cO% rP0]'-fnz-$@O l>J3$^}XZJL2[g8bZ0Ϙ "ɋ0,o]ߠ<"̷E}',!7^B[c#XQy͍P, i?3?&JDeSA@X8n4G̑F3ь5+~ڝm__}ߵvP 8H-6:jr@F_8oʦd ċOITr IʐX\z$8/(=;#L\!+SLcZX`xiJ)Z/BcA8R=ZL@/0 :(G"ppER]L!{'o!gȡC/Фb63qQ15?z4jaFD:|XȺx/S VeTP? zo2xqOxW<࿆75Z;#''Yqش֓ҙyj0 3;'D?#_I_q4_$/7rmlqI]΅U[{?p>E&C) &@lgC塣"2++t6^<}TuT=f=A](Uk%~=KةlMͮ,'3yQ0's?X>ʬn(=jқtSwh0gƪ@ʟVx$<#HF;G0: xPwG$]oDC Ze$0bA>ܙ@LȲ4}̓ǑU~fbڛ̦7F-')&k~5~`@hfÙ57Ła2d-UC,.ۏL'FgvgClR?s/S&{AlrTr_ak x^Gjx=i\[VAo 72ɍ)ע/ɘ0z{s XspꀃZa6‘8MQݤ9O2|8̲E>`i8JYn,{Փ[o9n -G8X͑`\0+)`:Ap HD')`u`RY~fuo~g5Wwq:+ǫg8o `M8f){$)ic#RHHIPNS`XrX͍!LxBbz{THM[)nV`o'jsx2/<ݏ=[ͶFcĮ-2_ZZpLտLտLտL?˩u~[ L\s15fI_}Ǻ.?Ll_ʧMJ1/޾ׂu&N\Yl؟\*Jш~P]5+WJې\DdJc 'n4":cN?:` f-Ѫ֭ E4J8>dǺQ癟uAщ}Gv8v][hDZ6$;(s֍Ht-D'Q/2FukABs)$>ǹc\/Ҡľu;]󔓣YhDZ6$;}-dǺqϢKJ c;jn'oq|L-ʾ[ Lv e99';iEb" 1+lJ :,`K/{1i5(8'Ǭ@$yBR E2lSKXb4/\7Ǧi1d B,h[nزs:nj˲WbyYN`f}3Q~$鯳G9䡱2 eV%#-Fk'˾*-HP :Dz7z4,w܀AJwR﶑ ( AJ-"6ssq-Љ̳̳̳̳8y51)ۛg<  'XMoQ=] -B%J*a.Jhb'tJ#X0Ơ/voW71 싦jbt\41P3q9~;m]"s0j8胋E+M6y_!85$! +5IMs ΁J\$RW4FC? 1J8J@#jOIgSwW4;GR[~Q8(Y0rA8 :T%[KqM(T,cH-RBRE`OKOnzt&7w-{X}.nI_xL_|,)5g5H~\0FZk65!06.P1@0P)VM5!rC a?O'M7!AE+ B#B"UJNHIudrqTOX $I>$BCb ` À8E!@ tj@]SlD(D)I>_6 W8%ZDV¥{Lq+2$k{Z#^$,Bsi%\ j883zoRDT 6ջ\|ؑKm_B75'@*ʍraHco0s8)MLFNbYA{5A֟|j.{ZspsEdXE؀`pZ-1^jO 2dV:i+/?&J"1Ħf#I=?tqoؿ@ڇlKV0[KjEE*nl)8vsi3f*+W@Scӛ>'ܥWp;PYcdQr]9\.]IDM_h}R98&ap՜NXܽZP2Ha?Z"(|7qg5p'gG`F]x}}F֣S9Ҙ[zArF.:4{F&RBJ])H'_O$HEl:o%9h=O=)4/L-= 0QjT/~I_ nc$+b$n,O50r8O!倛ys\+!00*1Œ3wv>?eaŜҠ Ǿ/v۾@0w"gf1 l8=9Z p|?aE](tNK[epdK4ݭo(UYi7A0p^S Z͇:/72yl[ަ} k1!:H-l2_ jLDVgX.rvс\Pj)vG_p4p'QH.+JB PlRXڷ<ߥI›BPH8̾ع[qX}ڇg6RHHR({{ A\7]BC|q=|$vgX޿Ҵ{6Z"Vp$vt DŽt`6 =/'M@՞a1w,aY#*cվ8fyIA)񰗞9N#pҝQz^y A c4?~2Sc*Gd7Oi5-G8f44G & =Lf_o[@Ig&!JUe;ھ) Gr$s=]99ԂPؑ1CTA$ }@mkyd3 AW5"_Uvj$VC h٥V$qGڣ0%vQ5]I0 tbnr7~ϔW| >$M>gB[83=!xX> f9!LV7C.o Vvsjxͫ^ -R}|s쑜W)U3)Ib=Q1GPo(D@\\P+t|`!ll ?ozũo^϶ﵦOLRF][F+^~`O`H<9gfdk$YO7IIDKSm`ErȺtWWWW}{^pUMgzp#sԨ@iScҨ07*LT_<)F Fe+NJw *x=#^i}K9ƎIqfHj| j,"dR~QrSY-OXa5!Ƥ*2d%d@Pq;dz UCCN>*V*?ah&U $yS$<ބZ3tSUֲNќ|F2 N5 ad) z".kN?GʽEU fsD: VQpu]".-tEMg*3$]o%fi93[ '}$5s8&.(G&OYn?C"B9בqz;0@Hm~xUHߤq՞% 0XPk@4-oDX𓎞ꍆ9 feȐY.O8j^II2D ޕfp WИrMۀϙΊ8[TasDv]F==U3(8]g/ s5s5[d9yn:Q(2a{'({QB PXM%w.ՕT Pr(kje0 wz̰hGcp0P| ϘYFG,oLPl)3?n.L<Ϯ OhL`N>E :zkҀ^(g+u fӮp外ݦqӶ, ~MbM9YJET@*v/pP= [By`mC T 8Ȝ?!c )uS= %Gs UR4T OϬ#Ms)gBN0!(|vx~/*\=uሪe&0i'_M|U_sܣrlO[Zw@{Bh6Jڍ!?}1 ] tOCΡ4?C9F/J@tl95K$عpd93oEQ8(h:f|E{dS{ ?# eç8nD=B :M(TmY 5"3 X~4` PC9 )Tb")\y\pxogy,ԄpD9gL)0+tXFڀ)F1#r@pw=snVE~ ڃ||* >~ڶZMhBC#&pGՄrIhu7bȰ1ڙ6rp\MKsr`6pD4B|3B +.īxZN/1ZrRT3iAlo9_Ƀ9ʁ "G}f{{VIHЈRzB &>< Po#孧AV_z( DHIq±(!GR~oh{ 6Ttx΃ J+y,#XEČs *o S[+foc$"}qJ 0`C?VRHW R^)@RzGIڕMSzab⩽9 s9Jg ߛu̜2,Ղ٣نn*„dpxYoJ@09*vD!VGMJ)-N

@7;)l w%#H9SO(ǖw$Se o+<6s&t}:Ϡʼ+WjXusς| U &~ zژˁz$ ` qIV <XxqG%&ڏT^ ›ZM7ƉPK+ިa(nYV[jq`9: ݈a#w(RGaĮd?[g rF[:/[tr/yx89cwwu?+N\3[t&: IW;7S* ZmA?R?y)$m9諳G=C!5` ^WA.X<*QֆJÊOQּ?*%.f(N҇L}?O?#Wqtw!+jk ҏ6z0=a*n_n8](W?K+\K}BYg*J"R9ʙGaŹ}4V 唤wGn_N9}QUoBL۵.A[&r,|ye8^4ӛՓSpϯ=CճJf}DzݟE72ϯpT;'g,WV,H*;;{;:^S߹tzJѺvbZem]8 '[cSVhD~~+]?jf|yx-8w]bɑ,TMϒcPIg4H㴘(.8x. Lrq&$4y ļ`4bq%zoP rl NpR\ujR(o F-'^0s!'ހV{ʒiV72p,=Քq.S7`&z >!fD[ 725C"Kj2_J\R5%v0Qwƽʜ[{70Kj 3EyohtA4~U?:3ldCI()2Ά_?#o>DƔaGSa#}\'T=Y{X! \N6 WaPE&9>]rY6I>sE_vf>m_ņ"Smb!OhD€2>"/#޾|D=l#6bH;#P\C%5z鹲Ni\*Ռ:dT3@xj.~HBWB,\F'OA{Du[[>)J.G|X' u|-qx[ SK^l"!甓áOusx8UPJQBbbiTxʩ}_L9x}D;.hÆ`UO)X_)AyX9ȅ$$c5VV4y:+JQ]B" t5|s W/-5<jO5tBv8'[Q m씑2|U#СtK=䷋o{8z8`ϧ#jQ80E IC.e]n 1ui])F۫mOIvv*'15~e?z1\Obe,@`{X#BzW`ʭ,k.P5\^7P[ZVfWzAKN}LG5uJMo]Em1֧Z֓`, t:2zTPs: (EϺ8 E|-$xS_ OIAQ#aH Da&EGȋmdCPD BA$"gH*N$H<"[PRH;GOiBE?]UU~o2mFo7G*Jo1Ukx=.oFJ-ګ?Fd\MiU(Yſyq)0E.z^XEZ<c 39_;7bx'K$:2XHȑBR iVlve wꂀ f]IQue@Hm+BW5ZeQk"~ag(r6y*dSwme>A0eH/͊JKt7%mt ڭu[U]ݕåyt rQ `A.J5%|MU@.[Oa&2("; ei ) @?ʸד"[vE7V˞Rv.s'Q$*ԟS5d^ADi!uϜpL$JLƬ D2 #HDJJ=9aߟDSM<`u%M(=SD *3C!G验i9W 66*2D~9< >qs8 B 0bSeTpD! T?N"ds$<[5t+p2mR%kvJx&˳omWmmfk 3غL^K؃bX@h88Q `K8?zIb,$-Pl#_OOè朠 X<[cI-N5#W2M $: 6:^۰|vO!Cf5K8/ _95LbC@i; F֬8fSX,Z A"ni7&h5_ռL9ڔCjD-̉XA[w\6U"kd5zMC*(hp>Ա.GI zg(&C R1i|XGjWٛ\3RiϞ:y1V5 `^!)j].oq܋_0Rؒ,b` $JKu?.'>RClvU>Jb4@J2eθJ* FTealwR4 Ĉ0H?ۥ|j1*N!FӃzIOdr{VJ,!Kb0)1OgJ$#zD;/5 `Uj@^,c0 J4G%6ُEDrwB7zƒ) $%Qk5c\L\=kwi?_dXN.j~\c{|;wKh옠6s h/ N$mqr +Rr ?by >GP陣D]qnZK_#[ORz7uݝ6V رyO  KxQKAw%Aw6JZ~QhY&q)5oeAG Ǵs~V=^ Ǘ`jaXf4xo4>01Cs^uyAև7WRk_t0ܻ܅fu2_M?L 2 8y˯+8ɧAAw5Ҡ02?M4(Lt=3( j2J:K` # %sy`z{ƛ;yy7zwXW+K |hR]|2xS ָqBwst4&SHѲm7p?!ET[~zt;}i')brvo_ɨY֖I;pg3V(:]*ĠhޫL< N !ջu?xPp$;"Y/Rm?)p`-(F}WYQB¸ZPN*4 X'ea8AJ/(iiDm_AWm{J`aiNiܨVj2XK eGꌥE2,9[% #B4N+20_zYUcaͼ[)P4]A$.hZD'NN~;(bPa#GT ăc1x h4:jlvRH)TFC4z\j13-Ɲzޅ TY%[ +vp̊LOqFs:?O OZݾ9]:':j =%1\L'qh\:nE˂4Zj#(0pި!u2AAIj" eAW_A Oȹay`P a1[)&{#I%?ںm¾F2F" zq'(0 1544$tb Ʌh6<檊mRꐌЭ|o*Key*"4!qja9[UrnQYZ7O:ۃ{ /ˏ[zSw+ނ?}7 0˛7'@Bޞ|yoN௸fY_7Zr?swWW|ྐྵ%ש1x%]CT#E'rfƧ$qY`^4ScR@$$'FwcP/i(2kungM습$P90ʀh N'Zyl԰ Rd!;ʵlP,`up 4;ӄ:^`'uuqݝYYSQ٭/!KSOV\e3Vq7wgʜC]08CLHY׳e DX3Ct%緫U-KOQb|.:pOR.1V3s{;?:MD>YҲKy z2et"4.OŖ5EF|7"Sڸ7%,Oa%MɒzD6jvc$Xu!m}ٖ"hZ*S9^C_Rxt"K|4liBȔLi",h m44`GShTc|hAD*h$rXMQHOn?0&W(g{Ȕ#4Ll1GCЂkdI6o$Z8qSgOG9_) sѬmnv{6Iܥ3>X9ڦY:3o+x^BP@(|/}gLWhI[i_~zj}_{ > G# V 0Sk6p~Fj0HS4TԹQYl:0,cVY ӊeD8ۭ`8]tQZ&[d,8@yy~n-Ay%F])EHa[,7Ϻ;6粧UBKI@`t\O:eRc|jRտmkfuNQxY*rxB{V'+tG螑 Q.gkA T*"+zysQ(<8K[17'~6W}UbaVWK{ws缡ĸoμ}7t _Oͳ~0߳;k뺪ئ@OK_YmkOosP+{pxoϒ)>K,⳦)޼-וT p49C4U^4:kn|4iE_|~={\+uyV:Ǻ!n'ԣ1U=lu~.fs_D QG\"El%P@'G:5 IUQDg=Q7FΈR JRU)zE* vnQp6)H_Oq+PqIv/Vp1^zi5P{@0=/!Xﶼ$FFTN>JiK&H%A9,^-iWׯ_5Ͳ*¤Q@ZjY "ШS %d'kR?Ԕ;~`b-Tog(ͮ{#0{#4ߑ_sbQ6=3' sH=ͬ w-AR6Y=?؞8 c3 : G .NwV^א1r`۴`UZ]zzG rHN ;VjS!gʚ8_Aic@uìpw+/N 3$AYY hd7hB#$hTW~WUeVȬt z 7E?t|' b%@&%<4[,S~tM0;7\"ͼ Y~!l3 +M55Ac>[Q<Ȧyne s~{d=cX]ʻ;gc4NQs5VRsO1֮iTc7Mn]?Ƴ-mMm`)8 G%=~Yʃ:OSHhp֊\>fzJ9[eX70=9]i[QK6|̖K_mg7[]{p6vCc > /#v-rX<t~fZ>z؝I׮s]w3>M{SԠ2 ܒ'n~ +kkGBr}LqgJabϩ|WG"QDAʓPXHJޱ2zO~"~\ BF[ɗ"dS*0Ȁ`PʔYZ&o8>o^UsA8S5+sI[_#%Z=9A﫭h(cN.|1anq˸J˸J|.#TUFtH8ӂq-)40*T5hK3(_*#arojQe5A_ޫvcTI'T-׌J4{B#F##%즄54|!F\.;4`9T.&5F`(ʎ/*F0|ᅎ4j߬k5 $D:nr+Rh۬~ #5<*c!g%.@ bR 3I3GR[8B 1LDBJ- _=T"d,!Zr*(0 [NF́8ڀ"|#X)ؚTIcREJjlf1Rp1 !%u13 K Ӎq<53oZ pQRgQeE)2r侮)lbwCc48F(r|vTYIgx0 ES a >qnj`'5$*Bs%ҭ,kJʧUyMf%[M7IZ) ~K2&HRDߑwo?l)FC/m(l/wCnrÐXwR7$ۃ˿0cY,1KMp8:f]@fZM0T"x8J~8#XьQ g kƶpL$قc{KZú(4yl@./:(LՑ:0L BC)e S) Z$u2emhH)q(tѱcu H$:&+i5:A##KAc |Ѩ#'Au^i?\m(Ӏ$JbƦ^RX!*!:΢1YT}c5Ԏ$q+B䒫pdYQwH( ՝2;ehIC]2 \>c A`TjʸO`8PbVDV'A[0&(<N(x| }&NH5q bc ,]B" XqY$&Xbd*pɃ[FCr`b*-FLɁX7%1%i(H AxrxЃd;p#c`10\S.`öhLk(yj({qR3o3/n]WO\/?}y^ B琐*X,vDt=w9 "z&"^~AMPBu(=\*M֦C%>XB;'qiy<_+EZ4`eBj=)(6OכrӄRSPcsJ7֝+W׶>=%( 41d E ,imKPN!FTՑ*D6'gh Zr#?oB)j@M_ m !m)DQ""B*z&TpV*DZřɗgjUsUP:%&Z057eQ)/Zh-"K[Ytg'_K- X: P 鐃;zג ceU1/l5nq_݄% 4Xƒ {CnP3-A{*iಘUp&23XgD(R"R(*hLPJ]8hwq.qףǏ0RQ$8tyV؀:u1vҒ0 A;HhL#"_<ـOz&y`P`iDz&q{di*XIN{) .# vd:&<]_.FįqX\8)|,/8 S0lCWCWCWCWy;@#RTYT?]ʱ=iWbg& ļ/~tT7 E^&MI՗K(r-@}w?`zZeRbsDz7:5!>+y5'7Js,mNʝOoԺLvݬ0OgSAլ_߾mM>ȦU1vgL*0f^A{]?s g0KӼۆr_ r3/`@nT 8tqB'%%%r`ڇs5P/S T0cL fL39M^0cqhRF<ɹ5ۙP&ѩؑ[6:9>}T|zWk|-@OOe(f)g`)# -Fշ:aؘпYz=Yv(ךaCz'#5b811 \V/4Դf r%!Nĕn1S禗sMJ3NLYHK~n3uK)C*>3լ<+DEN<ٌJk:ͧ`Ge׶>FgXXqnX#f@6'ˤ^Bz}VC;]mi£eYގoXULNɖ#ܻOb8웿= KyTli8Ǐ1ײ;,ccJ/d)D{Ƨdz_uη m+#IOb)}46 {/ӆGԐ*RRI"etKT2##247Ӊ"/%jK,x FWj#ygcB4hܿ S,ݢ7jM56 Y^sY$ + g0%,/~>˹jϦ }Xvn-FG}Yb \c/k"_+&uwbEHyc(6t/.7'uspy?z'~qpV~U^ ̬Vd BtUoyun[|nBJ(Ӆ/ԮF.cw"y >F6*?HG' R\Ϸnqn; 艳*>: xo%lwDPwMU[aly;Aζ`5bve<{wYbSڍRoͽW&uuPۅ_Uv gq9Q ;O6uvҰSw!Ʒ Kq]߭VynW Tf 0Dc'L鼭Jo0_ ϩbz OiQJ7~C`iS0@%O~_oIW]U3ITOEGMmՉi'mUmJ'+ pSy(7¬!o}g- ߆_L;T_7&eRN}=`'.8*ZmT"m* "8 B6yJhT"FMRYK1F+4Vi:](2"jxkJ{}\)5'_$e/1iu9e5CR0%l)P zUwo:xnlO' g䵧׆%NHMn"=)2U_՝5+HД1f*)߮W9棥lWsBkfï3w2D̷:Mi/t6]{7fξU>?qƧ‡0]+ӽx2)i}j>'0\by戞-D*&(v +-4~T#Q-D2XF߾#)|K^PHۏg< r#)J)`Lc{!k Fw99dp0kz]A2e~rLw 0"C|`T{461_)q0m޸-x8khcV8D[YQ\K__ramGa~l/ڲÏ?q`\a+ƹa^ryGKnr;h-=5}]L'sX}A'b4IĂxdM;Αv&TJ΂=Fȵ}O%{Mʲ^qkXϗ. ZZk̓h+~Z❴YN5#b.ws.5tMCGI0:A,R–MNSFX5|]GХ`$hZvv1hbBńᣵI>Av1hbD] vacŐ sMsO㵋})GkS.Ŕ裵)=< JX)Tsvyo"X^],VLk=a ىzb2D&k$&l#uu1RO}c?lQ4Xi S[0`9(Qr3-xF0XX~(җHd ad}UBjgO,v1Byɼf7b200fe$!`:hc ^i(T̸ ĭ]Mg-&Y7z\㫴Yu?2=gVnWVk"2SxY.{Mq0l@|!J:܇Ϛ-"@?x.9L /9s\v5WWe[%|ÃgYGZ!JQ:ZH*avb Ct5{TŻ׬gM`Jccz)GU}bֲdmNJ`e QX̯Gjs%P`-gō/r/|}Sav~{F*w\qwsxnᐻcg/]r7#TW4UN <1YZ݂aքl'ε})kCEcxJa{vEKA%։c/֙Ӕݢ#U[ r,S\vSTAbPubGuw֓Tִ[|JEV9FT#^_^h7U/[,UX'v}[w@Syk->vCExޟݿn\BbPubGu*ٷ5i|ӊڭ9rRut_c|/BjX Nh9MhMgWnupȑh OaLz RI%b1:cݺ*]4jn5#[r,SD>gvӚ~nqTv, hZ#zm[lUXe-CExJ>n tb1:c:<ϊڭ9r)ڍnPbPubGuW4e5_]Q!G΢Q3e/L)Ɯb̵J`1SbE1E*cfcN1:%>gË1s$P1s`c>3L /y1s͒16;lQĘtx1fp+Řk4s1* ŘUޟcW"Ë1 R9Ř땄1 bc>(- &,)ƜbuJLF /,)Os@cs1* Řy1+ JŘ)Ɯb̵JBseYS9ŘwISQ#O__^e!"gWs{3/n.GGn <FьI0DeL:Y]Ojn>,Ot3ZCM<^.aXfdHKB0b1PgT̸ >k* A)?M9ib/ka4&׷W &abDK'1D&&U*%6>%'u)Bb6V_f?\(9w6~므+4\6(m+ws2_<`qpZ3$5.0##ä7ZX?:aѣ!2=ea"h`SQτ@tٳ !LHb2E9NR)" APd87821a˸9",YyJr/K>f ugL Pj2o‚3nGK3s{2OFG?Gߐ}#f vΣ۴I{c/C^`buOGOf}q<8)u|䷘92/.'nn6|\){0%\69.&̐MPM|=Q-4Qp(-x ;6 D@Zt>𗷷Z~W6|Ub~_nW~+֢Ylhy֧_gOڿ<*ߓr\Mr4Ϧt 2SCDh<xXxWۚ@Y_m-T“7HBa7kr9Q"bk晴1BR8I A' VI$C]S]PJͮޒG ԋLll9=io#Ed6)7h,= N#=|Q,Ikl.:$;5C*wT !O? &nxN.sϷa0{=^]Qg>[6l5)Y߼u12U5`^rr2?Y/ xCeSJ1x6#~>=?=O!9{ `:N!H9y;NW0L0#@ nn6 ,57iqZ .7W;; ޏw rT4 wdQqd\S/X ?-;鿛tl̜X+Ea[}b&C2o 4F֏ߟ?)=0-13կa"TwoOp*#,+wQCj<.hf_B~fO厩?s@HoĴ0R=QC:/%}4nX41BU#Qlǣo7ձ>.MjK-33e??~=kH q@=?vVF/DTګ ȑ[rj%NQÊʢ(aAhF|AJT 9MkTFD>`Db0CCh9T jKڐxUVg T0Y2x wbc|~]D,mo`IJ{8t7#G?^f 6C( PPKs흧 ;|!Sq&Q]xR' n𷸟3!D1q@OMu_fa>%A2!o ~u9JA 64Uhu. CG?|GӟgSHS5/AG/uJ[hMeb+$C# /4'ac\qX y"{"ܒ Mf׊\M`{h0.j(BYn5UI̥X O 0s< uj+'\?_zoc=ǛPVVĭM{(8K i W#zYmKѫ!h'9xck3l~W~Ŝ2kic5ំ?ܬf~0_࿮_=>YT.}:W%7Ϲx{~1>B:mnmTmeOwǛx,SsКMUV:z?G}7J3D}Qg5uk#tk1m4x f{ V:1'PRUL<(aHAmlJg??oN`LKHFr0|`ᆳFږT7|c:IX6.m)ޮ_1-gޛl M4JR-H]l7ݎ'ُB_(Ceoyڴi @I^dQeOGkϕa.d£(< %[N] o[MY~8'IJuD.cKJ0mVjz' OiݗM%;#zV:+ +*T;/d51<5M7ۂ Д2r0JŎ7DjA,eTXeh)*⨖1Ŏ/⠖<y9ibOۦ, ǍSHh#EwWOwN`42NE`0G3lw}vVԩl 90Tr4c[{e<{kXqEBmmh3A _&xE}%L{/V{w56' a:oP~ VE`#KBؘzO 9szaԙB$^Sp#6/MGPIDm}T̀pza?Æ Q(|'JJ$&b֡4\]N P"+i,>se9=dTTyU1fݰ&dA.Mld(Aj6Q;]^bRKSKw1v89V#W1!3EiAS$a96Hr<3fE>,Ƶ.خ^>ɗbbkKMmu̵Q.`Vfvro#jyA!L,6W]l;F;Yoc}v ɨx6騁t2xjAĘuP6(#w>lT6t@K7>M/w @hd Ȕ 1L4sBb&CB \_'!I,5ai@KWJITm4 J869;@ 7Rɶe4ĤgY( ?Y3g)C"Z]N F$#rr TJj7Gv-_9c$3Q>Y]Հl뎨{,3Ռ𫍅5W{&&b{K籽Lg]_eoNOk)|?[rrGHinǍ Q h)"B{W@@Mjxi.O6}iNf#;H8#D98jGBGpԴz!l5=[ ;bQf%I?W;uCuU%&IrΩ*PTsT&wj5BWCej[-"ABiqhwm+u.t ڈ7w~Wdh=$Xؙjq; +=55)64r/1|K60j`- /ϿRȁu `ȃkp>6*w#ljxwQ.nC K>>]ūO#ՒlݫÐkZoOW$BK]H#l5z/+W>6PoHb9fu[X# ԡH~n됄nx $aIM0QprP4 V]iqYVRctuKmWt B{M5V%㫷3#҂3?.GZ*MXLu5԰8ǁڔ˵?}z?CRK#e]qgMWi#X -3g&: 0jCAz2a6mx`뒺70xd LnxY<;ZU06z\CV$t-+Sc7\ׂNu>i E-xigM7YZVPKZItȤ8{?&wf.o',NuENjLPγ8?"š:[ZyU/̒OY9I7dҘY-ra(tFG / @xGɽZJC&gj^ss*j L2c=?|$[^wFێ%%F6"C~i(/r%ǏkK-xeRnyCoL([zi}iZ~G<'dȆHiϭ̂,uAjFU!۸*mvJxgZ GհGTRFr*U5]V+/ڕY)ی2}.CݏU)j;^Ғ4zyFRA38{Up+*AhiIDKs2w_|#K{RG#D7Wբh"XOcTUR[4ˬfGGV/輪O"^"<*ĺGYȃ&*"ئJJ""ҞC)A#昜?ѡM_> ?y7qƕ9MF /^Pi} 0*%+:[ UTٙM#[&2ȶlRoV=~Yl:p<..a~xo:R4]k0NJ4L$\"hB2L2"_˘QgY;*ڢJޏ [&HH!:߃-5'&-k>[X tgi_i&MxߍwC4}ݤ gqݣT )nQe:IcI^Nx0uL~ \,j`htOHtg~yL$yKg-A\u ͗l_EihB/tE&YBB+ =,O΃N&>[v6z],W弅Rb7B 0aa.|\tx<p ZU֌ [9 xͻ\ v2덃{ч%k,uO3xgI^9sYlv /8eZI !xgSRVt~WLõ͎Uɔt|ƬRC4!"ъNJ&P1`s-InJY >XuěHV9VɦI!R U1< p=Äi 2Jc $v8`% {ÚM$T[$1+#%(L h/d E`dd1GK.N\~_H p9(kSO~?Eb_(3" )X[$fJ4h=sQڼT`(EYM_"Ғ5v O*+2 aw6 -o4詠@caLI|Ro!\Ouξw3^OЖi\:|(Q!A3a,yN1ZQ}hs c˞P SpW)81T+F`#NYe4 Ki;W >-˃B| M'ND 1tޤMZDdq~/tIO*HxIG@#hC9LնcbZàI"yP /=8:0yh1 ) FF !_$QcAw`0GxAI}DW(PhD T7AR 6' \IFXyPr%R4)MO ૭R`b;qqIўBXJB % (zT3F^(Dؖ=$0RSʗ@-ګ|xʧhx[d=Z*u):RGqL+^| %$|+lKA 1|O:(8UX =nvu0ed u BHb"p\aꨠDe*HRW]|F;YQ%(P~1 lT!? @hHE). 9Hę:Hr+idOnF,pHH(d 8 b$dMK*+[qd H/dGhCm#rJ4#M_}jWhYk+*5^Z]0ݧc /uxI"zGU׷?_K]*(T(8B9RGu*s]B ɓfLɇ0{򼈀y=f 5c". 7B( tnޏvR%Jgs*`-R~^8SaB^{8`Osu:>D3κ> ]1+֐i29+s[Ijg GdgHЩƘ?ߪVpJg8~XpW;3&[bɀRGfr_^jNJ{_xW. &حw{Mr<( 3' bKSu܏(ZE90sv;i ߬&uQVemޞN1Fzj{qƐKuk7_Z /=G1W{gƬM^ښvɕL3ZadW`l`&Wo0YCp "% dM/~eoSf {wx3ÝM`;Wnw]|$DwGb,r,$d(?ZH>,E5>@&iuHW5ưdnrI0pm5ێ(cIS5͔<)O& *>'$Ò + 7Lv)yɴ,,'s|DTsط䤹aQt09vb&.".V Y|d'K ?'7/x`;qhGxDՂɓm Tʨ %N6.(B@1`0'gՂklky3'RNmnSLl>4)O:*˽2/ż'J'uLܢG2%nϾK. Bɩn<z->[c ~\oosc-;'8LPq{ SY𑲾O:}^+V?M^Y_Y.7B{c{aV墖ԁ4+W'/ExOUvQ)X9TTL4;D2W`E.w7f~0{$8yثJS;f^B(Qq:+Qr3vxYFm=sG)vq2[_t|j{6U/W*ݼo5޴GF+,?oAC-(֐hjԋҐWţ+ N FhVbXTs9| LhC쫊#DJդs^ԣjGOiMbt}G3>ӭ ac1W:',moǤ {-£We* a <,\kV.]׃\K!q\aEҐ?lR3U OB6泛an6{H52!?VRE\3f!ۼ#:ރ_:|rصc +\,zZ/ȯ&(QR>'u %k(F]U*}-%ZJwJ>yIOa:ijLْ( ㈪}>+?9tFW+)$%AP=ͭhywgTfBM%0/ߒ;*]?3'雍z7kPpD .0ZbgL#FKw .7?z4`G&%v gZ6k!R?"ICF&p29^@x(p%RX !g p\̒j!|IghUgUudq0kKm\}$~x߈nq82ݯ ͞k5Fs-М'&B/%{3H+ݴ g1w0\*PCRvAu;N |FgzGRwSHBt[=u..=qշOPm&8ve0mt.N5iIbaL#&Qۍ.\"T#/>#l³;0 -KFN#z>ObO>L& ZD1hD NFU)k"a$= eWxLqU+ε`Q<ga`cM_UKpM>Ty*s;.z $+N<bg H iq*uv*L#e/\+"L鳞=l)>!+@l9yqm"ՋHr7YuƔ?V *-ܑNaEcFeH{$Yd.pCH  :Iqm-L0Rhnѯp1Zaпg=VoZ1\*߷$kN4M0|=~EШ+]u} n2w!B;oy9W1l 4biCLމz6vtMdsRM[R ?B!a5/5wxs r4^vE ǚP;KʬE&Я{,KzF&chKMŦirb)V=MA SsU%P`|4[~`簟j]/|kƥ|_#Z]b;kH[&@OQ6z(OQb;TSіC]M>Kp~CXU)RX} @C l71ɔ#GRYb(DaK>S@zŮw<4wb;Ng␃6M}s*nCNR1 dYTc'/55y~ RQx%aP#GjG1t#~I5cyD4v1[~ / E~2}7i:EzkZk~>ևv(t`SfXmh{:{V a{tԒB 7Nat}05O[0`& Бޣ),CW_l m`Hr sttktr[~>zwwfۇI,}a?gg{_}g7"3dtR׎1ʼ~},cNnJ3xя(W~$[~T{9=yTe*i8狆fޖӫr=MfEH:&g[$9!~Xi}o?-<@L1v}@fNQ] q)]HXJa#AO8\!8R`ķ{DgGk+^Aa+m4a*ݶSaXןz|f'yRT`SDl."o@૬k2 {-E&6w~5U""p1)o8!;I'uk8؉`"7+&ݼye˃r|Q/AE >,EF ttL vt!xklw3鍈q鼲/4_,`%gg4e(S3~Ү]vQEkuVd{1$&Eb@; (5{2F: rחXE)VHK#Ogb&$i+y-u%Im|2IRJ25IR8eH o jyPXͧMMf*'MQ4 aVqD'&8F>VʒR)g2ɝ"?: Hd`R-JΔ5䎘qv @3 ɨ,\I1(thTRLREܛ9 @D -tJ6[ [. Ȇ@mW9Ǹ75g|(UpKr$l2Y$ژLaݶ>kuVI@O[)v$ W@vSD#.e\FJe# -y`"b4Ab] [PjIkmۇ_ A[}Z (^/Y66jw;V ؅k%`0& -iyl`;IkrJAbۻBE%`XW74%>\bk.i+]FѥۛQYBuvH"\Hn˫<m|s^LQJO9/YR0Mg fg?_fw߽v=..>lֿQE7)RLS~/a`Fyo:+XZ޼~pp`w)TDL%9l$#b[cB&Ʈ'/!?L Μ5]xaR'XZ*~AdYwu0^# xgy4zv_KW[WY9sʃ JO¬>*)OFEO8$w+lZ\})?v5b NM)v52Ĵlf];D>@*=7%d>,_fż0=*?>~} 4g.3@stI3ky4 7XrI\YD7R;T~ϝz\#7e4O_ ]ȶɮߚ!uVzo_QvB߅FH>nʻ *ݮNY " g*,CԙWEr&H/uurta^\h鿦G7yV%Yg%̵tr ,E3F:E>;'~a~Ryg%r܈Gے1zV(%p,ޱkMU1;H;|V#NybcpNRϬ]>uڒ:Ըv~M>[5|*>-cz .a9iL߃]O<$-E}w>iN܃܃1 y!xWdBGٟܝRor3\pQ6E P4e G{0SeQZZgp鹋 KV/tg_,XzsL ѹA1GeZ[J8OKuxJcvhgҽmo]1 Җ}Hov[``28ZrE=uFENeJ{rUCr؂d9^F&o6) :TF!pJs n<:It.% .W9;cuCr# -aӟ54ZM6:yItgm3d?'C<䞪6h~h3f"mrl!O*ZQ6-Z-2MV4m&KDaCq F'. yyɂ刲V<@ nO7ׯ()>P`ṕjg>"#e %7xx =tRXXx1U;>~fចluf&)Ĥ>^Y8JZkڷWo8ajm %y}AMJIQj: ׫wKݗ?W8kW^WJa@5I퐮.9XP_ZB}uiqM_7dWiq؃o b@X11aabOGgZ1j+GhSvs1Ԇj]f ʏ37)ONWZ[xZ+ǠG,TN׫ HƾIէk0{hOKY7kߞ,)bä:1t+.k u;___Ԩm':(}޹w.߯ %KMu{]|rrC(qv;4*0κ,7/y+:O!Jy.Bb1a$AQi=IFvT=q~1wʍߌ#Vg~#qb垺G( OYpw[P8|vf~R.WwRmdr")( ƳFH($ w'q%XTzXe:4ܲet"c5N|gAk_^* bƞB*Ͼ@NW d; D-UU$ G'HjZ*ظfu2HSxQZSԁu"%ܸWP5tJ "rm=av50c -v:G7UL1wxTYkTÞHKvDNk}+gAF ojZxΊ1 $v tXDֵ5.uCE}y< Aq< !6Ԯ ?MT?ekz맛`Bְ?1u+8 VgGgaw̤"ʿ- \],>}-'<ukxD>= "@%i%l*Ə`ͤ3]T9Bֆ·EǸGIOܢy[TҤC%]=4i2|X[d,ӳKq;#agHԁRGx,.!HwjtY~dl&sЌ%< T!ߓ1D"OY(>h5.q.vć(1Z^7moC?h.v>(cMy^}B뛣A9k(;o_VQNփbswO: |]YDGҐWE:;=VOiݜcn#T/Vϔz]/+aXfFQ1pLl-i"6vc+N1|Ճ ] 3)`",9R NA reQ L]mu㚅hZKnVz«D_P>x$lT.˷Mn)IL;xTDt<¯Ϟj%؀_ V͔|>Y<Pw`HWkq85("eHĀR=!x!׉ƌ}%mRtjPZ][hb[VuբxUNו4Ҹ:ͺ&uniؠ 랺_KX(# 7Q{{IXœ ΣRa>uŽLQg&h-!lbb_]#eD+@C huH* *_# lb V5^+Zb{kC-J쪋&Py-*Rࠔ" mNҫʵ+b eJoa4\ K~7_Q&e'C$:cr$1[h6`6V'T*l1Ujgw\[6 ˸K=>AX@lye ?V5VNÿ3L@^E\~~v[țZkFCe -oar`<9-7cxGؼMKko7M,eWROQi5pHLHUN!dj-1pB1; y|H`|HeJvLrSgOK6@di> XyV^LRG`7IEb9Im:"^ ȷխmmUp'NݨML*U^\ ϣEψ5#ƟM;9ĊoPT6a]4atUo Z^KQqG| u9Dd'2y4PaDDHVS!=rCgA:IB1 )CgŒڒ KWMZ5hG15O,?Ҷ_unRVZ>^(` #cUT * cgb l=FQ{K: M)T4YHv3$Y4[>zil1>/AJc1rzl^b14䕫::G!Y,(pm NrAzDt ,teqhj䪳bOq/2ؾ(;8 N2OFVqdlzDR&m4> c6$pz@bQ֧y1[uR2I2l]Կ^ ԗ~:zuaxBYݬB?'XR>nl"'RclY3 q,rͰs̒qkN9DG!ҙIl(J;x ^* ;r&/Hijc@O!@&H~"6=^ld2VY+vJ}zKI^{Rû*~Jn UJҲ&uggv2ƪ;ː()Ƕx,.!H*ҺكZ,^XAS2)f,tLQ]ؐם&ASÚۮTdko+_UPֆQ¦*(PMʳ;X7U Rpz,Ȫ*| զwf{^B{_5p.})ɸkMux\ȫ8jV+"1V|6A)fv}iLP)K&󺯽w snWW#<bqTTBu4Hŭv:"~{A5l5&kX.n}nik,˶$\//)ɣdIfk{lNX,N!8V!Ic#3qf<>4䃫hN5mݼ"OD R#hJb8q*2J *SS8w;Sn2Q1XO 3Y$֭ *S2PQtq0A1MPQP.R6M'?ER+ҙ & S¹$JA*@'Bm%t}0OӷÓ3٦E%!~w Fj]EW# bhB(RV =w(SrHTkJ %UF޹Nt~`d/! tH+ w԰"uU5UJ{{H죽pը-|to "g"9`E17ADg +%JT|ATVrC K#wUeߨ܏{DkBĕ{C;NdD!=X4tŀs;ں5e̖[X[*բTThϬU4Ht;",;}$c`? w+6X0^SʍP9\BKLY׸iB*U]ʊ3ck rb ޻m(M*a dHU )ZK f)¼;|]RcFM#MmˇɢlZ3iEj[TNb#Jt}1Cl]pIFE|:))#_HQ6ܾy}UH>\Dy= q&~x2<gjSLaeh2՚fg/*\3wY3wg? ?^1 .Efl&1NAIsџ#v]jUijnhH%EHs h\Lha`vbm($$X$IA#e,S^Wk&%ij_x%rEҖ*v&Ƥ(o:j[|6JԊŔWiv3iX8Ec. FxiݦKYRP ڿ$ۻb[pAVlĻ Ef _ I7?PrJW-+ۅ lU9B1Rа:KT] &*)ښvaJLPkR~rAڊf`x}VQ 5Q䤱6h9BǞ)ERK)?U˕S(I(ț 7"jZPn;PKu_NW;FOl`|-GͿs6%vڴjj _KDLRzcH&G c}NΜr6h):^9Vf}Mы0|+8V @^q< bխDZYŶq˲eY}1eֿ8.?M_ɺ_Z.F3߿t[̍_MP/_WWUկU=Bj[Ǿ0SZ,kܺ\ʑ98h)n+O#-E+ *A ίM^!k }P_#Lb8Q{*vGxwSL=M˪W-릫zNU]uS1v YgZnb`YFT-W% &A4N1 q;Whp _!~<<G !=o*xg~)YC4-xgS K{CQH؞vނЀC1.*I-EҠܲuAaeIrŞ1^˷q2#oNr:2B'g.r)&Aؙ*hdǟ}#&2zs_evyPf=Չko~b?}N]篤ԱNuCaPJXJˋp$;=I2U پHhhٱ/()UP1n'(rN̗$LY:'l]}%`K ET̺*Y,?2-©c/? _ y XY X턣ļgR!Bv@ P 0BTCӏ?>{J8v dw92l]~=w&!b@I!V!B!uo*pyL`.vGPLAhE%T4b LY}2bhoR~ \tQYG-/3:׋]2sַʴo2[7F/߈cFps[ϗr|q2{T*ek bV/!Ŭ|ﯛ }(/[iJ|Y/T{Ґ1::dǺI)XXNmpށ|-]uCC>F8w%[ BT'v6mSln][UN e/I]zy]-7s-կ|z,g&zO>ti=xB:rCKuUs% /UhqVaM!Snmv4j k']%D}' AUA+,G1&(0e>lnEQ)$d![Hp$֘4c~;5&1Q&91֝tPŠkLq"GBb)I}9L{Wr.BVB34wd`|نih<;~OFۋޡz~hJ6104()]@LQAHAHr~&=4,26tGnA6o0Yɝm$Ck17˃8Y&H DhO:⋀GiPz418wq<A%VVWZ,L B5?-FE Q7=}Qx*ջ]Nʩ>bR2dQ]OnbM2ԭhOnx6F]@n)<|*}\ F3ŷ~Q|l=d7Tpuqfd@KFEð"rjJDMid6,L rY|aֈ{x}Kx\2#\2ˁsQv6~= JJ7_1o$]EB5"'zȹy4I${[I /!O261}w[,"lgK"~z[-Óg޾"&LV̓l%UTÚeia1Uwvc%ApXr]8̱9,y*PzHM\ҧ釛O/겫 e_T_RȥyDŽmVuSw23gc g7_4Gb.p8۽sS@ړk Gf/uƊeכyvY/'Utnq$wጩiq ͭ& htE )jL;$4ϟ38bLrF\DRU嬾sn7n)&*ܙ FɩTcZtOۀhF_bEIr@^^2^Hd꼩v˒epx)wX?+nRڷ<  Bj3}3wFFDn!H@I)d(&5\b9 ?푼enUb TPE '.︛|Is;PYՍ O\O4aW?Y<_ÔE4_wxk~8:jc OjX?A;EYQ0t$2Pj$.azWׯm0?ٻ޶eUSuc ľ >l5Jrw?dӯEʉ39 . )ak!ʕܮ=s6.EhYbCA[3 lbhH=iInmRJ clS0_aշIХ2AYJT h7Wl̳t6Ɯ 8 <J[,E{sjwg}|E,f0Okݵ9O"O"O"Ox~m}-Zp9ݛj{fL |S}26UuQ_`) H$Wn2-!gyI\ǓB@JBƑ7$ h'\4@@h-T Gy KYP0j)'%GKޤDKMՀrFH:`ڛ9!:YrVw`BħN`u)GkJ 1 !:a'0¡֪U V|uXh b% * t D`ռ"):Α4{4:gމ vb;b'|'qFDQHj<$ XSk;j (k_XTΩ:\aQx{%:MToG)E@12 URܐ!dn iqKUp`q׷}dqpwm_Qmx <H9=uwRж FK9YAl1.]dHc /OSs!>=#CR(=J8M)B6X+1:+gk'h00_x_?Ξp+IЛ2;IrMrJ}VG~*0\1.^9p _;}`uw r^_ַSO3f/߬iK/qwg?py֛[^ٚCg?;[/7gwްfW[o~m{=^;xW(NW=Ů{O.isqJo$Egޱ~쏇J iL:Ͻ+Sm5g |}|0D(F9a^M80ۋKcVA;39ь~o{ivFsٰfn}:-r9S_x(Vgٟ2]2\DTvo;Iņ#_LɘL߆Iw}P+L/213y.;{R$ݝ~9Mz|/ܸ>h;͸hwpMNG# l&{g籷y7:m^欽omgsw??fs]N~gyex3%> B\ݽIV> sb^,b~}lo>~T^O2'Y`cHջ7p8~_Tҍ~xv9ؚTj*~ OOaLwoakO \z89e}oKvݳ;!_MM{XZwkʌHL+nE܌&'Sqиy -8E;geN_oA!+mE܁Uӿ3vNu1ŵ,Y+D6dU`@Q zs13\K}JS1 ge#T{}L X)aW[,b XpyXp1靀HBrNZ ( RL#q󁍑8f3;9w3P.c2\N,nm6]>]-Z߭-I?^nTeIչsXʤ]g{gmq{BB:+}߾B.QvvTT&59aG{h[iꅠZ &!C64 Pj|l<ݢ 0cR*u`b.SĝmQٮ֯]R&C;wak.}ǽ8{[{;/9KMo_zv}>Nc^f; G;/9Z#^lqg&Fù?z|^r|[sa!s0BsRI)ٻ8+PDDf'[6 c"#a5YzlT]-Յ]C=lVWD\+XvM$kڻaRCٓ}7}y6k SCu l@!zW>*Ւ4=e\.KԒ7RưE2F vNؚL.k[jILkZȝ&dZ}ގ|ȇ|ȇaä! ~[kh>bCrWSZJkBG*;6sl5- % -|ZdIʪ NZ6龂o$ }L$ }LҰ?&iZ`А9Sb)M&S#:^;Ęv;/:q_EA"4e<vDWjIpvיU-ѫLbml .yۄrٸƾybuQ:bTa&c=2ʇ[i΢q*΀of }&I?`@&3IlSaU[a52WoZP @:¶ELך"km~ f#ȪˉL)[UaES@3kĆ5Ƿ?] ĺ=?}b, !dg*2B\H*Γ‡$m O`Fe| )b?Ӗ GE$drЃĴGk28{v𝬂J)fzQU9 9ȭ%%`#!TS}kq*2 )Իd}Jha}rPoNYA6:J$ ş(*8'rcbBgX*z$)%dHiQpAa %4(.iYa2T C_&,M,UbOpt25嗤~#Oo}M|}Cȵ(`C u&/~|щլYE+I-&# WWb:9h4ܽ;\=ռEm'YRZ`\ICTR[Vs^ZW 3rZ@=y׊|wLd_{_ߟ(3eQ'u_ATR7T:Y꺭8]nM|׽j5P_!{( Ăo-dkĂy2;7w[=Y.Nhs} feޜo>Kp= } `m's+tޛ{+w0m౷JS8|+t @ ׏v+t{zBw[9l-?#mDDںsOITD-WGITΓJT NJكJކp$D\@D>+YpNT>D%OdmCrVb;eGX߁aY%&*,i$DsM',tYoN{ ݼ7vL5 s)L/FXP6ꅦo*GzN!/q;y+C.v1rӫe?|~u_/\_wo 6.kGgϏ}ԇ4`9ކ{%ۂ_ի؍!/W^]*`/RٷkZ܍uk)ɚ\n_WoWOyٿO]\n4BO\ }Ť6vHg/=lzUӹ fb?u(jVvR#lsd+-PiTGۘM{4JҒ׹fM߿6Sd)xAѫˏs.o士{XBgH/TGɉr&Z'+U*-ZJکLRy lEaH˥={E.$֤1AlBoQ˂ojYҢ *xi,QbD!yB d+>֖ ga:jBPC++J4o0E^!:oLTۻx]?v/xx񤿤]8Xvͧ/nCt~W_Ld_vK;X:|qQ~Я>ˋxc| `Õw]to߿v~ }eUp^oV/߿ǣXc۹~ڹcsy ˋ]_>NcٱzIȃGm}e#dag-@rZB9,q1Hπe= 7KS \^:N6c2 zGq54֟d{s.QW$Q@|=5K&TP[#ڨ ".Qr졞GƆ>aPuAj}7~bU{OO!{Yr16?^w\iGzvֹlk] $T֌Y4bsf95Q]/IVniQ*am *8wŨO+Rb3/*,g^soofk?~Gvjʜ5jw졽%d^IJw]tW_M0b`OajVϤQc&&bppsC0k);ʭ̪1c_дM^1DNNc tBr[#^h\?y9c !;ԕrpJ^>eȴy<&J pNP Md{v \9eyVÿ"_n]4`,>µU\f2 ( $Okx(^!1Z'8);vG+Iz3Q ?gvsLXd 1Z &Qh5!g'X;PkɑZd&iN҄J9cV{%Q#g>*&֘UΥ_*o.#k*m6`e^.pc$IqͰ '3em|r2U #`üed9dF\02Emfg;Pļ]QdXg94_\3͂Me al170Ty68Ǚ"yc}"L:A*5İ~HNAsäLvvB+zlM/HWO")y[T)Ѣ1=o l+k<ՇUk6o1fWH%lyO*;vwd8FGoSRkvb@N !"@#n YB!ѴM@Fe<"Bhѐ '•Jɔ[@lɬ E=Zǰ? fڙb-;PfF 3NцEwьeN+@a7+Y%>#<0>?b(8#p;PQp$ `b9Ej X^=9,=-RԄ52"5fN 1AP{A!xݝSCڊЊ*!;),5_|ÈC,={wC"u/H^UeW3mkUAg 7fouћzJP$cگK*yƋ7vGƍ:LDb"7J(4Z˹\Dډm= aO{ϔc:.zvbk}qglùk;kKB]>X@A0ܟ;aݿ0]#zw+Vrd!tLAGzq!_;P2,#'n"*]υϜ]wyS7ɟ.]@F=N99F­4C0Fo7C0%[qu7ri{sEUqb6O,oLdlh]`dUm&E'V,K&;%IefFNVN 3|%[>`XlוęvZ]zV6׈b#C_?fH6lE dR91U5ŁˤE@ YilSLBrl%EV$U41+f.f)t UB$Q|G"~~$=x gq6=c`d/!Ilhk̴߯V?6,Vzpi& \BRROv&H eJXx*SKso%I&Ȥp@k iD-~:?֘v4$,;;Ny ɷ_jLUFD*%'9XK$3jJA0QCM=^0I-6L(2fx0#)憤8BȜ碒|e152;!3S-͘u`@d$ Y s'&,ɘq(,8s8 = "i v ;^ޢ1y,& Pst; a 3桝w{k8E @2){/`TMSb xz%Ǿ(!=Zq!$n&Go2Sp+g* gzp&,Rp R(c]L(Eղ] 褤61p&(n+&TZrEIlye|Eo8"UJ2> M$ F9-Hf2" 1u (u#0 =GG_q4q7?%&u/҅!LWfqo"ǃEJ7 8h dT XQ >RXe&EጇXQ-/ \mi ;RLZsM>j-y+ڷ=3w_LV^M]A=B5S+GZv"8b*ƴ;R+0\bYJ0 RdΥ;Cڱ,:+gքQ%5kO4^,ܐeF[,!V \ 5rܶ$5Bux./g9Bkz F3l8!WNi͉e!oIrO8=U Y( y.h7SD#A$vL2j" efH0^>kCA|f[dƱ۽ʡju)JU:F "a$\,3t&Ê! K Eވ1G)A Q*gmTmyT=bCNSDJj\҃N]!UbSIIx?]{F~M5H임y5-[S06X:sKh0E:0\<%ls&ƯQ(,k 0%Q?_ʔiB”`Th=6]=a&f:Zh♷Bۥ^\d%E!^0??[7N0m?޽"/Pv4O@s0% +K)R]`_@ e 6-XdS mӧnd$DqnSI&ؔFy+(9>,l1 ʴMr *Bv`Фyx &:_rڻwExva?tla -/ 3Y%Y&Lbex}i٣sqX!zVQ9K+RlM^[D$,zC 0pnc #JMjIoKN+QhsMq8PR!Ǡa!b5Bж\SycDXA yEvqύC|nsb|m 2SNm-wNS$Z#hI nT+ۮgs1{\hLu^;F?? .*n>0]4&fU32]}%;0` 8mSK̯趽q A܏~kPQJL(B2%)^0X,+ ~'ZUܨޠo=HjPzitz2n%m}k@ ȍK^:Ǡ6 CwdWܐ{32@[#o by*"Ju`HӶa`6lyN˩}8L|HAӊstej* //M ~LEyH=emDL';(ou9ئLaSC_=Y=nPI0鰻օL#G|>QuxAsk3|\D1_Y}BJb[&8?IBQ6W[x[O&OI,":,]йq4x_O֏=8o-#P-r@kA]ODPDXM"At͝iXm9n+kpO7*f7kkS`fwN-a}p]H[Zmnp-eۂ{@0m・mܖa2{ ջLȦ 9*y1xMP |Vma~pw-8ٌ+O>L%>5O(ar#|lӋ>2h\e5i|`woXbe1Mmaz7Xn_|n-nJіzmc/0:{uk`Omn}]]#ToC:1SqqNC&Sʰ҇TO*F.ZF5^Usk[}/`]|~V p)xuT& Ԉ308܆Jo^Fu7H.BncsQ^yi%*hR( āTDUb>n4(:eQx Q\O*>E{<]t>?(_A U^(Bl//KV"xG;{( ^e|5q]vbAEyjIݼv耑6԰*ݡ?kOO_}$v(0h/۪2|F W$h[\xJӛӃukk\ h%-͒@"[dQ-zUCqgEdӥmxZ)M& iʛ& 1j:y /;19 ݴ+ǘդJl*$ebwW]]]U"ouO cYlҷ+VqQU0R(@my)t$⢞ge9nhV;oƣ2ZP䰗nGf,9&T@c 0c,ٱA%og8c~T1hN"!\+h9{Kczh㽹@'q>. jp<0JʺE.ǥ| PEŮͬ\`#\mQ۸ KjLїPI` =F"4b84TccͬNe8[9PRGq2(+C}zMMׄ8/㹝uu8ȯ˨5O%$c-UUN7]؊-z6rƢVLuWQoO9Aɺ#:|4 =͕6ʧ>C,5񼌅Xۆ`F)b|33">|KÅ]\,;?Pޠ}0sOK*zcL!cHRV~롳7n$+T^tf\?['d\|-quaw`UnU,DJiMR`t{ !+lNg2@Rp 6>D)DOaK< [L6lov+u~H; ];z Q<\X+`=F$0" (2ho/[ny>%[~@7kMfC7"gDxaGb@A?c߄Œe<]r[ETnE>Uw&%*m76|]jZL&}7-@HWvb?)EWOQVOֈ-~2^)B4rNh:"V"\BMWW3).sa4%~x,Ofr7~c+3?c@x F2B#׭.z2?ߌy{@Avko@79o#nD|ǰ |kL@2+Nz9m<'i5xrɦ>܋y%%7-=)A^%]O v7tH"áp)@)m۵v f7j߽" <\(@JjAЈ !B!C0(hJiKru?o*]L}_@'~Ik' b;L0@\MsܫX>>GZ ^bω Q2@mz7o_t2\z^ܵ)߀{8BLrdI9tƔO(IdG.$Obg١_'/EB4c!`aG]R$@]5F*TL.C-h`8@0 HȤ" 1<?7S-AoV<񝌵{RPh{KIv449d8fD? Q/4;LJxkJ}BelKi a̪a]P"\D= yD{py 9EQ$|9i<_}pH'R3mԒȅw"T^KBK0̿%;x2 a'#x2-m8i/m[?ͣ'~Ea}poցӄst t1 `qK_kA8oOaB\p+A16<o45u-*y~e$#Ã}` pb8z5 l: .:8P @7 &#V1ߧp4H\@+‰q`I2IAw"*~v}'ϑ {Lnfnn$<<>P'F~Y'/'Wqw/laAW7E3 *ͽS;|c&0"^_x#807\Ef4܂o׮,Ɠ#4LzIɠ,L?6 uo9K ^U?JO?{{}ߟD݀w].T+呆u8WpO)]2+TQ G^iuݗ܅KZ :^$yIf ̓ߥC=7mC⹝ˋwI \$.iF (Ms,coR̅2ML|7MWCu/W_1tq=pI5D udNo |DR|aؿL'3{O ӧatui$OؿJt7}!/k,(=9/tKڴ]akrwngvj2K "Z9 gNGz"ބg;II̔xcϦ{NR #*o:F@ ׫tX|q?'@:u˞ݾ4 ?Zw~g4߯Ix7a쨉Ӻ͆/_YLҫ?Gi燄' ܥ TG_EJQ!M$8yP %g}4A` EmL6g{g6x)d0_4vvRڒG tw4lcwK˼ŧ B+ߺAi+BfX^OWb͟K*2Qabͻ纮H3#VJ˥*=0xǸ 02虌ivd0yX'iB|7jSeMt@pFk%"Y1'x gs-Z-־.Ht:8Ilu\(As޲v+Wy8Y^*0&]Q]A]Q]A Օ`R.\ua`g nw\Fwqo =!܀ʙ']00*9RIqD&ŦmTFuoTUwy\܋ꎈE,cyF$QɧSdٳW7YINk))*2ۮ/J' nJZ\.)O{jmE(QP#>7EV@KCBl0džPU82>Yh5(;A@P&nupiy^.qFl8+c#Sv*')Wմu=SPE2P hY0Rsi0 \jF6څd=e@}IsI<3k2>&*Aw@j7wʅ2j = q֟}n* f`~0мs1' {ipYEvwT}J_3ɄC]IdMEL*d%*%smmzDjJ*`E[~u!/"QÉ}`ٷ9Ͳ͹՛^"8TJ.(]C(V;ƛw AT>.@;[vi֜kׁ>%\E"GxTӢRp* NlW7?m$(pѷm郳p;$EAud4?HS)W_);DԖ6̦︸PrCM.%3X]/ဆXs CfGL5Iנ!|3 c&Aȭ:wVт m-4d֗ii֕]ߒi}9X);V`%߼0o!S vu7rɔ9ɔp,h5)RAT*Vq\e' 70g Q%evd|vzva/~$o G"^KixC}Z0Wԅ+C ΅6 H jUde X1?܏ vѧhAFg8B6#wZ8ؾ$K;SJ?-%×jG|=E Kə{/gVIe)'!X]\ Q㸬gCֶt:Iܾ҅:K̈.M?8Wӛ^ENKuUDk?\>0GGl8q3.6=HdKUs<& R8Q$"k&9$-Ak}&v t r. j(Zb-!!x}W舊etRd2Y5h BȤ-عfmz,g!7*4cERQ8P5`,SF cMj?+c߽Nc搹R)er;TbU/h mtpЙ*]kKJLg/OfH{+4IJϤ%$֋dMb $>YqK*9Caj4Ȇٲ̤Jf ]I"{g$n.R!rwQD9hJiG9jBf6B!jw)))i*;ГAMc6˕K$"H9tiґX1%嚓D'rdidICтKYmҴ|t4]r*/zsrr93[>GjOCؔI;4R(KְJ4/Xd"tz̅-B҆NVvR?P`}O* 9姖XFRk1[LQ6m8^Ux[@H%em@<,L83RWʦ$Hs1dB d3GSq4LN%d<ɆM T")a9ZR4;@9D XbɈ$C0>f4aL@TLRIMﳪ U7=flbzC&nșw:7}2yP]UV:5yG圂?Cv{S+*yM~B7@Ρf'fN G0NLay8LD h|j^uݯRcZ}baю')8弤z%lzãi灟k4\OdTOhTJކQ9tI)[ IqRT-xb*Eޔ'AkjcNB9v~xTŋ!};h?}{ӂљ&wbz}R9րb1t`G˴QGcF+) L,߷蚊q)-rCIq>ygqRk$* ,;F`ׯ;ߥG]}Ī =2nj9y__}Uci0EMNkG8Nv.aӀ^Аxz ǐQ& F;AԮT k#]qjݗ7]`'Ob Y37]kr1,ǧf*qRs%+X$vկq*:?OI!=Iu=1_%Uz(OqEVj8J)N ?{SP"Dm*5WjǺ?KS\Mm>Ԧ; &v'z tzRf*$.T%Z{r;KPR; }ҪtY콂pG1rػUP6 w供0\8bmrqнoe0{_;Jt 9W:f%|ozLdb~n=No!T#TxI&TiPu=Z(IdPOD_PF>P |^iDs=h o)Esvʜ' U-سZ pi B k̨ 1 3t:WJI5pȞJrY!3[ {fJ]_~FpJJs VQnɛG&"13 Cl]=~8'v?';LZGv!?++F#AQ+%.TJ- v(\V]mqk} R8 .? zaf=D5᭪dv{@yY= ^=XKDz]ZT.4,xBۛt5&"ZW O_Gߜώc0Gߝݾ?=|1\87:!Qn7XJpQ`ϽdJвH$@Lm˔Z&2;@G}ʷ¤Jߧ?5tvyQϙH.-~-u: 02HCtVzzG1>M^"f'W4J^$?dlշƢ1W!9ō5 СYqZiH5YC. ,aD(dmg"A#8u d+۩tZ]tQ:d qUڝ-t)cstq>:/a*vrt?؃` 󝬤ȉ*izJ)ZᠤqE@DFI^Vԑ?jJu-eӷNe}?hwUBṞn`-wlN}إһ?4%≪.Ǝe`;e?d̲vfY3QN6C^?B۶)q (O ]rtDj;QƎ}o4Je{kCg[z!y,IR_2jV1j.q;.`e츽 V7=SD-BCe OUd.*> MR7}o4([SEt5I?QAaJ.[6kBl͒{m\RMNkevgo 乽uQPG=P [WoM>=?bSõ~aljï^)>xj߿S%Tr1խڟ<pw5Az{ _e&* l4 عoK#nCjJ%נҳ X/ TZF6rP e!rCЄz[fC?:7߄f#k4״Lp,!21e%btĜ(s yܙb2}V`)"C ]*-|RWnz]ǍKu0%s{ObRKzck(9%VQ"x!s^9H;%Ȣ1-DU;vco;0q$YJ9F*R+@+R&"D $Ҕ'% 2H"AikEB IiD+@*f!=JdEC F31QF:/$ E圵R*hT(v69PkAʧ0) DXFHWv^w5ft(l3ҭ%bIJ$10zxQlR(쒓Z|3@*0CI.l bÙ}Nͭ;}#0lb (JRH?e3MqW)ym_9DI2ψ buXp $M뙒Gag]xMPrF{XFE7&r~f*AN}l" BY?5P̦,XJ` ,q5>ub^ARѓJH_u*5L WZbxk8e{RD劷ACI鉞,FRg)4}71ۇ!#s$ȑfMT Esl<=@(l:n Čɐ-N۠YJDb ,31 1!bGĴfEqЭ=D[%:f\okw-g>X 3K7pP {^Ʃ,)2ĊL,RJP/0!kM8,q 2$% T {ie8$ZAW{‹Z@QR]lצnoR]f8nȩSy])$R܈V`IC̨ ˞]R hƒ6n7=8* 4Q7"aW.1^?K uu,/q3?ǃDo^O`IOykmH/{8jr@.;v\v`@63؎$' jJ("G HflYUu=^&,=Y#l~LEQ%9J BzUj?/r$!\DȔ) &=jPنj{tmsZr+ (<uĨhcݎTr*H%vAc<1:Z]k1ܟVu/Z5BlSE&5 2w,K)rwKd"S)ɺ΁s2g=iU{f bI'ECeu7ŅpP,} >h>~ſ#nߩ3{AbN6)^PpYL c l)pkmFZB5ujn4j6E TY!a:(o|nH?a[~pN]$: =&_]]Ln壯C z./Z?n[_z^#_'Vȳ+cU2Q!'I% d$%.Γ[mZuV6 2I8@Z;Z&Id.NjxHUI'pe64 l-\*XuR5R(\fds7Iϼ,ucZ5$u]8NW_.^rА$ .sR&a ;!tl*Ve)X(:;ۼk|e;"9u }o 4r.H`J*g2.Jt3"fmPAr Z -KFAZaqYT&HJܯ$oqccK-3difWRd $c(R--VR$ t_Hm*wm$ j{5~蓻/O#&e;%oVy@lﻑ.p#6rvP:LsVϧ׫'zbE^\Vd8cfTNWe==r3Ρ琭( NX `|ojњ<}Ag6c Ea$nnȑSptmWp!]Wh(]:OGȉZ6 7-oЀ@m akp)1ԇGGj$h.Ϊw8RL!DLtt֮ gsY_;h<:6:̶VݵOH0-Q-hL {#$dt'Jf; 5^Y4s;n<(ǃC]>>(? z7X*>wCuG?Lj__iXorL$!OO`=o;=1woB=)DV/|ݔA ! kꮮ)j_H"C7E?%ʹɜX:de.dڍ}?e!|t_\AѡHA|mՑE.!Q¼1OsFJq괦{Hg\fx^.Y-8nPnt8KTf@v"?+1Op<؞?gE'kFohD- Q 5aΔ|.1}0^Y֤S9J`SG&s<^7j3gU=sd{sg 0f6$cfg- ?lWc u*.f<'s~yr/n,3b0FXh.}0wEӿw%W_Ie ž2yO?F!xq I0`{H0|u`DvXilo[U=ux*[nyO7oxF5kj[ypZ!$Kމ:];u}HPNo?sg$̠d>K‘gAO/(4du^'gMCހOkJfɛrl,SƊsjYq+6$T5J[\0^'w'W(Ww{^û0d[~uY=e-^d?w?e׏?1 ~mXYBb8~xBnj~>rmR7U:z>/n_:dnջeWǩL“G8|qg YYP#ćx-]j;x:dDMh@__oTt?CX.kEBmϷ|(ĽUD(!.^ƉBƭF{6OO/|cg,83jIeBORMlkg%#hŮ}Iҕ}`8+&oaU3F+nHmR. CWW߭G ;?{u(|~&w{q6L؄ HC0S?4@-%:iLSOϟ`rކpsZ'\ 1lmhsi_ s1i1%<\#I ;-C]lF +U!.Cp[̵O3BFMbg7<1ᒿH<)d쀉) #X̹1p+>(1 MN{>.-iZb WN9qLp0IRPvxq7Kj?k. "5s'>Jߙ1Cr}S-ncf>ngΌ^|y:i쥹{nCU6/ Nekqvo'4SK}71H{gQ(QTqK4xq ~m}l"ScH6kG23I6kZ\Wl\qS5hgJK~ VޗTjMn{FS}wѻ8XGNe}}]Qm n ܗcNHڅ3!k xpVt24dA9J1T%y̎<gy/Ɗ㊐&?>QvyзpG?o Zq($(I$NQ9d@fpI,K2 x'asdDH2312!Lxkz~.JpFKO\J2g#5JVm{Ie޻ƿf8Z5 8ِN [9mD=`A^$!%wvKt0[B2aU 㧻mqs&((@x>y!odga6t/]/jɎi?'V!,-][G8lw8|ppA3GہCAZ8MUWDIEGdk$eZB=s62N w"A 0i$je;(vhʝ9\| Tws-0ZPD3LJH(@kX-ZEw\fDbS&OVVeƤ^m 5(^-}/T]N&p؀@}_dr9f'̼efڑMG0rZ滽 E[.T C̆J0M[DŽuTxrs3Mf¿hlr[Xw+tBIr~eٖх0E#!$F$j;`}Ҹp?61|b.>^꽐nF}CiA1݌>f{.t$@7Iߪ WQ)B֮_x*Y ]%RXwIf7 'ۮS.뚒@DW a؈rFZ+ _LgejÌFE3OPk+5w]w9x+ iД7L1n*puڿ=~?gI}:M7ߜj'./ܫЯuj1Yٙ!R*ǐRY"rFKyɇ3{;yIglMN8ʥ׈:I??,R0;0;D /$a6?MgcL?ƠO'uK3EQ|3]F0@SrJ:X BdAj4\B N "̬j47c۪#crgG5DG,#i4LK. A !偹ڔ!<AA"M^ Lqܡ2%Y[oWn$0qݫ2CGiܝdYC>_ kt?Iysky`+ /Idػi9CZ_/ ګ7kZL?gMӗK$nw  :gBbhLr!/ϋtC@*?SPI.ߓ?w{_b Yn θƕӹ> x׾TM):33J\2'~JٟB^4V|'_v#3 _fD)#UkyqK9GdlwaQ_)TgaZ_GmʞtNe&YqҺ2cae kl?匰v԰>078Lekc`!0x~\9R t(u"i M!YЂJa4蜪~Wzy`UoIjHmцT`<Uee,$)jÖm<e>@sVnh'۽ns̚6.( &5ỲB( hϩG% Y w_?܊hK/<0FǓuwGW_=sȧr>I9?S DV!j'GQ}Hn&|g6VA)p U\ B'F%gʥU 3btUaeexӒ6۫_?ݗv*H#өoMu&[}m<*|0& 5~rUS8[,g: |(GJUaj- ZAh㕍QT`Dք1PjrYjR9Z uy :ǶIk6N _ ~ -VDGtN(%}G# d,Y"nqFh ?_4>!FYh܄frQ*o ˪ c,m~CrF}ɮ Z ~qC+NvuBV/Y>xexwvo{+% F6*"p]OQk1ttql4ʨ%4ophVș'0S dUpIΡyqМ9A *8*qA( Z?1D LFfDJAVmJϝ4Ϊ3wlqQ͍@`(gVwesՍc4g"I,8SfT>͂dn"aYZ"78ni|-p`Q+El%?SJCַ5iX| iVoZ3 {j? C0x-12u 3*aktվ/V |[GІlb^{0g5ۛ[3݊,4LC ~aFW\ļza%F2 k(8;Nj`8OP> X8JGYQu}Sqy:N=Oǩf61BM4urs!ښךAP A@Ą>la+|ɶLxH{n sXy#SM<6|E"8GLp] ^]{)Yj phsmӡy:9o45R\DWQN*A֢{RXK5~0J7Pc,H/?>jPN[rR[ / T)N Zh)cAYZ@U K_8oi䆰)KXNY4NޛFn6of\Zojމl:U NE逺6u`&ZY!f=o=Z:o,FLzI2ǕUI8Z7AJY㚃hA:A(WP+GCIiLA vvs,sl`ib+:Qy.cQQb(b&RKG!O13 4’S1҈ X%#О;E#CuKE"(H +5xdNUVB'HTWbv]WL.s8,1Z) 5-PR L+7N.KCI{ne,!$t"99Ձ$G}ڈT:8AYD e.G=J!E{Aב |\t&:`p[r (AZy%Rz0\}TZVb U;m#(ZMB ~؉ #A H8q0 үDwюCm-d@5X菺NP>!?eٜ\\/I->9:nkfjNO7z|čG o':Qk.P\NлB$ ҧ 6ҨQv$I V#P# g%5/FaĩWU()Tt'nnVܽ/Y]myK]2RC6j|hj'sq  D94+ =*H8QDUR<IgEJu0Z* .LOn<`JV„i||H͞'/_kmJf/(aqYξ&O ABXg=T|Ϋܾs O8!yw¯u2o"3y۰%pyOdtk29j' 2g 5jX6(k31l* sA򻳭l A=E5Io[},:~6y{(iW}A~Cs>_mw4 {r# fQ#Yb1J0l0%mdVxћެ !oւm_>>9re'%z%:"@zxݖ,Xzo¼T 7 ջ/`k_ ÝGBŸf?{s\O{QԬqt<7G=Gw9:zp=pTr`eG<1d1OaVh;MA=1|˺1ByeWM/Z#tCH"/uaزq$,fx~%Кy٤ f=r;GaB ͽBOH_p`?w=f73gO'cI&`rtwg>֓1uއlzBGgW29@o(a2ZJp3_.]Ph3-vY`m ԝWRnk8}s&GA^SkNK9b՝WXn j-][yV@JI스o&pM80u;UE.gW29J3#!s.pp2tRԌ3 F㡮ലfB]{zfQ2MMʈj)ђ到4D""cr2T@Ôubzk%f'ֶ^K*Z 4U **Nytq@ ZhN)FuԠCAV<!BEX͸؍ \ 8D& r)rK"#H^)#UƊJpa3$0 *'!A00L"F=2pb$ʭӤ` ܸ_'ߜ&+C94B N@ C(N 8V;:`Tp2dsDjf*0ln%`ܝ\ JR@kZa2",Fd,g_@uYZ:ͣ#H*3do A/ⴚ\\&(hs[,l/pwQלFRSuXu廹:#Dݒ?{WFr /PbXZpٳaɗF:KLRu&3ldQÞ鮪KK+w\u5j4G-Ž7jjT[!Ia ] /Z;1,EPJjE)aV|0-kT B(x) kfzkl1=Rp 0#>raC iu{̇'iYɀ\+ث8`ecCRa|nni\J*n0UƢ%;>Bq~43omrٗOn/B-T~{ǫw??EVgQ/|''XZl(]|knJ 񋢧?%Xw>̺ `tP nKy}QhzM6~uy)D1a10~( hgj`fxyiayCRzXk{+"uJw"{wTU $U2`VY@UUFP=XUaz܈l7~ LvoߪVz;rBSx^6zbX9XeY=T{t] C+=\9T{.FKuB"_t{gmxzn'?wtdqx|+i{'+o$ DU eG4k|)J4ODw5ZIND}EB=e P;_ޥ{v`V ݹ- Ag#Lr:@ 1!"XI  VQ$+e !tzr;u|%Ӟp7~ OR=mH:WI{®LjLj$)&/d1b@#A(C DI@g 5ĤJ-p?LLiaԻڐގV?旮NwiWX'.cHgbc*5a:h)p >o贈 v._OXl){f*;!Cј-ܺvK[S 﨣wvkcѽǔ +iqƓL* JHb*d{.dxSSԇ%VG B4@])E5W*[4xҲ?G 9P+5{s0jڡM05ŜF`}Ӗ>>|/A9+acEƼG;oH!q˒8L)ܤָKG UݸWGjqe;R<=9JI};aJ vd;?Jmhq~΃ Ċ6 =RE?cR2?7\ ЖU<͎J0C|ӹO_V%B˔*+V{|3Xx,9YMZ?f ݍwUK%TMΪUy;r՝Θ/ÛyBQccaVj~=7j=fhYYon)|mbEL_V.gx:ڮgmWzj9m(ajW.sarXlbyoV5AtNf,Y!òkқUsiJ0i>[MhڸNR%!L zq^՚u\{Dn5!On>6EhG;[[=;kWSeIEDln]I[{K廷m1ѶEBIG{D)H`§m̭蝰իIp+1ۉ?M;(JZ aUySVo].F?Rlř`r|~vN=ZaTt}L)7 [9T <1 NNX̋  c}/wφ|_.Dm SHũE;rXJՑ^E:(PFDgVhac#4R'JS\ LKZ 0a=c;VgOJ RC-(3C(sG1AZD&(ڰӅ́..oܻcn ɨcԔ2BvN+Ji}D]ǰdY+M qLqϓ1fJ?ys3p^Noa)kft" dRNsRퟓj\Ty f3C ;s0q\ %xYmHqD?לQ7/V9 | *Uץ{~a5bkL)7/{oTpq"{sL`<2wt\3#/.^毾LcJ /t¶r; R竴м`dJg]&_r|Ӊh[nzsAqTthC&q4 )7QEq$h~s$4tD= N|Umof#HSoc*HIqŌBLV >jaD@ c"e bDo©N/~ժ![RLijrRs9[#r 8]Y5̞W@`o1a kqqK ⱅa^XgF1EQe{ut\r6c*[MC`##1q ^п"3O+ܔ/׿8ʖw1T iG;gKQ~;lJDv_{VXOϊa?k]\Hm U]Ya8;+-3F>}JΪ%8Y< 12 (R]J{kAQ8"`|$R7z'0bZQk'E1b[?䲻bL ?BAVokƃy"kc<Y7$h+`Z rYӽY.F[e3J"D^WR387&3H$gP߫\?TXTq>_;u-.j*cTp&"cueR?#tWϳƽk9;e-ũ-i&;@>uFI$% ѳ u2ǸfH`AaP( & QrN4b"Oy !`X`DIp4G:XhO5VwT=#ӸJI=9`@IJ&L1.1H'(^c,)nXfKf)NO0kYB$ &Kpp^:S’ɓAQ ! 62f񆂩[} i=IX#L tsUVLҲP~$1gNaM$|X-PƊ蹏pu>*.jl`"`j [-'!Z J'Gڸ[Xxs0MI+,erm@ ~ , ,AZ{z:M;9+=x rFH)L~ >+#^r ]@ѐ(8V0rr)^D;xQ"bR.,B'<qրJ`Q,g` uE4&$_O6OV>>ً_So^_|e:݀l8J'y$*,80D2X ohq,ڼ5yv" 0av0{ձXÇw{f:) j_S'gE, :7}`1>q<8ˉJG 8 xҼɘEtg'$(Pa$%3X]$< C΀}7E X$T;9&1&&9`a,<3!r&%$X =&z4BL=~=NZ#lbRBզdwv&yu?,vC#_zMА`ISr'}z\]>pA\V _7i "S COzs17,G-’ iN4,I={ )ͩVzrk_ړC``!"+Pv!hmqpH|=HqN9bQ.]F5 6$V FԺdgԜ[˭ʂ/ qIx@hqQ59b@q gۘ&DZFJDwPaTrMO^]_6Z(^}~opV&F@yc tzs=rwuQMJb~/͢{Zf~2ߞ~$}\TIR-ؒRZuDKzlq QFrκ;@~h>Ez/1ūNG@0=w~Q:a+xpL LC:M&ؤR[ޱ|AB34ƿh=7W菮zT%bZ=q*=މ8U$ږYoԉ-lԉШgeUReŵXW%6RN\QA"h&ave|8[Y}nNQ74Sr; ~jǡ"&cC\-c{,~cR#Lj0y'%ΫREˌs.y^tYp6//]̂kZº~\h-fv-EL&yiC[ ?#\86cfǎ9[Mu>Ḿrp>:mW?sPG%qpjp@X!FR#F8VeF^RU)Chw{8қZv~yw6/"e͸\[Jmv7#yLY= 2V +Aw Q isk&/7#6O0*bK,twJk ֆg8B\n3J逃]†p=H"QWMR҈=С,Q)" 8"^(Z{Bm:*$ޤ ͑ZhNNIO⦦}%H!wG￞ rB[R(MF.(R7_i㞠^BOJ#ki9}0jID\E~V??-+OYRښ _8s2-Lؑ~|z?wo:E )JM`_fBNfaƐ&'ktld#^cVHH^:4^@# aĬM"!v{R %d{U>U 0,(RnYLE B:HY5m<ӪM6HjL(C.k!Nj[]?)+GI$Z^4ϊ5(TkݍuĒ):b+M5+xpu-f#K)Z4T &yp` ;n͎[Lhi[+S ~s^M~XjIn<~Kc-"W+l:ײ(8_(IeiA/&kΔ֎Fc+͌T^`- bhRHN5ߜ)4S^KIfz4N~KIé=SBBcIstqPI:m%~_nfБyW|h`-Y(H$V" ȗX 'tgYB',^9S\#'׬K93dc0R*.5g|7\䵿/y0ʢȂSҖ8Q`>M0~{rn«mxky6(zvx Sfxp?Q}V\Un+`Pic/MQ!E* fA0Jʉ]3'rch˸ 8ktP E]st/r@{k jy3!Ԉ{qyF Biw~dtqDHdVz/18h62-9ߋP"O _jbj*S% 8!׳)ʭVJ/Njm-𓫩MIմu#RF/rpX˰‘놵dK za$U|qH}A,i0h%v)X׹~a}(0nyO_M BR MܸBa `&8Xpǣ}zyRfb)s{>;GnұÇwFi;Q)? m7'}MOFPΌ1=gz+ǫkl"-4mBbn۹6ټWt6;U`]/5Si@a$KLb,1UgޕF#'rIc! )lZFdm(Ks^cm[IPcd@ |_@k hMbI,5Ъ*0'V`'iivX*BJE0oG5i e)pn[+0F Ҟo.K!u$$Uqr\9X{sp5 P5 PTXK6%&ЍR2 0:J!M)RAܓlr2gtHBk+xKCx"&y91C0QՏ`1% F D1p^:n DFi6 8Y?rҏƷ7p`ޢ{Nf"Įol&13dLf25HDV]3;pRm߾}Ugӳy|Q|Q"(CY0[Yl""2 vr82#lc,NZ˔oڋo{Û4GOOMNU>|dFHݴnh}\,5>fKèY[ Bo TÌ[^h)*K/?. AK&Lc{T]f /PHBE*,w\84 ;DӁQB”8fY*yG >-0LjZvTI+%/9tFP0 pg)L -U0 !bqR @[VJ˺](N$Le㗥8R jF\n:8 s;#.<ʈCqPjxPUŨRDQq 94XwQΈjmRzF`6yg#f( ^ӃӆW9&Y15 Ô%TxJRHiajM)n)btMnT]JL1_t~E5peځ:kc0s`2 4Pʧu\ #{E2!r!]CȁRjşbcM =aD rZA!\W0xOY%5񧌠ZthQm6dńɁfGlzo`{6"b!ʂ׉7DR3X _l^;VhirW+dŰ${"&ًj}eQƤÌ蠉Y7B{@c &,{ Hq.R38LZ/eW"6 l Zm=yK YIU+.g䥔X`sxBe`C3G,R TYP:: ȃ_~38橒_ǠR@ 0U01g:/uF iKCm R<9.3*S6쯮H8R  {MѸ +T鴁PKamgT n?qSeJoO.dgc2wsd:d~Xj//ٵw/vqw,"xhq(>K޾|\^`B8M:rtv᡹1t")36 _z=Sr}3[2lz;cP% z_00UגGfץyG fn&|}I$4rA)+vJG`Ƌ?_zﺗj+چ-+.-!NB6&Q5b{Ыwj׆$jPa0j لbnA׽S" ^>BJ`!A@ۻdJ@v2 ~FkW2z|H!`d@.UIti_sKjݕ' =Jc?="LrJC_rK9ERH[k֪nE'Co~nqq#Y%̎TtZ%㏾Z`IL-΋E- #O H^rTC5ꗇսۋN jsoe[Z߱5\Yꈮ!xB onxψM㗾y9oVeutH*ѫhؙx~.b J}.gv̷9JzZ_h#S%l(d0OD]ކ$VCUVn}m#1m9IW`;z1AY=O~ V'!Bn\\|x@uٽ ҌKui}H#6ݻAS_If36\EwחX.]Nz*ߴǿe CLvؼ8b.N9W9ɼ}@+9@Sҹ/ ig]MvmD~nnDdMCvT קC5m Y. 4Ν- JBX /x.ؼ; {+'Ew溸wMО^̊+3;Mɛ //X8M)K8EMy㹐 䳓+$ݮicZt9<Ȉah񢿝[x2Bp0aS+lwDӥZ,K;$.:⪍C Z9tt4tI!4ߝY^?L t>;X!zKH yG{]Rz|E/uN};RLV.%{I{In wfiܣ*xUX2__ݚk-k}a'DJF&vFyTU*uS˗@e/1Jیj#oLfttbQHӽG8 8F=֋FUct͙Ο=϶^]0*Ie]n(ӽGQa]fi)w5mnHyGewC9r1SwIho# )tM; Nupz=)8Dp0m}SH6B_z=)8p<)8<wʎNVu ȇ3P Y4pz[pt+z9)Z0N0IΏ}t qVƛ Ҋ*/UIx(_qSrqtd#u5ؙŊ0C/u.!$"qJ f_nfЗiJ~(%1kK C(D0OI.SI ƩCT8Ġ*U*%Nilel{ O>\nZ *SFY3`m,t$k"iP 7-9#IԒcNG!♋2Ls 6L<{8Bq5ճz=oj//Fe\N+] oƲ+qsZѾ5灓6Fh\.-5,YK!QD~;3Xlu46.@dC헋y}oKND8_ݝ(;򜵸P!ABeDJYCEh+AE5}Ll3__@QcY'&=ȤO`˦z'eH\X`ߘyd,Gj#xK\5 6`Hjx졜ib%{~{Hm^Z+ ࠒ䌐JA QE8([qرݱ@Taؓsd4BD’9rЎܰBl>j0_vܾ=mftld>^x=oD> iF(| `0!2Rd `(08 eD0 Cl6o|KBB%SHt06 騡vFgP"LF§~O{uobt g^t ߖ 9dx-fbwn_Z,2M۩xw;J>zG:2Hӥ]>ۥ]>gd F8 s`PVQ@ F"f0*T7~LZ[q>B:E;j5l8n@M<8l7kuΨ脅;^&AD;KHt9*pվ~#M-y[}w̤xc!G"TgŇmAtҕU|p;*>84vҕQV<5+YVp+]Umkg*"cd F\$W_ѬN+4B&Sۡ (!,M9gʾ|(0P 봕bqi/k>wcDE:"C,`"zjJ!*#)iTL+(JhGƀ!= 얻vs""rZO7z`"G)ݍN_/‹eͰUS*qZ@.,uQKODzp5UU9-B&@MxAMc{ +yQwMU .5%RV k]q/p4z<HX7p{=aC0ft8sf;!8O0!&kˣ< Ϧg'6kg `6uQֲlEڹ|a$%m?{;cDH =Hu<\MFYeLTZ1a/<"0 "{\Tr,8εәg$+ttYS}ybxĕym>\ P_`?(E2 x&GBv:5zQmc8j$'Gk!Ɂ].@pN:I*{Zh`,wgT $;-^P%,TqC@ (? ¯9%x[RT Q"2tL!Y $ɺM MPv?~#2nRR:e13xBT:;(fOTѵv{9\w̼aLAD5fJǷhf>z]ґy!`Ƅf*JGU( [c(fǹt2K5sIJ7e-I)%ez' jLݿô6ЋY|k^nT˗g'b7rҫRNz9LPo8"%Vu 0y A:ܹYE\9)[rU^lpְO |`mC^B"K@| %s⠫?{DI8=kmg="J+\aG-ǘNC(%:wm&ICwC?pS:їH{k=2)|^]N"j7MwA&T;dp5ɽZfnʙKv5cuYs`e PZI5riI9PrN]OʥdtR.G4f9Oٟ?!.Y>"1{B.3'62+BQtSap9 8aHl7|DC_("#Ʊ2 u,lI_֔)5.?'Wг NL̘g1B9V5[jVNI(0 Q}ɑu)&ήi;DXznB۱0O B|}\F/.vj4& I#PsC5f("B" aEh!a ऀd>wV=8ٳ[MJaœdR88+!K* l϶!+JտyO~-'l^T!6uQ1)7 ׍I|3o1Cfv?͚jS cnDagˁX_H(ǣ Lb- ӳl3]~}k4`gwmoM]2N1FOިL| mQ֊ XB_L"_@ T!Q!¸R"Q>3q~oq*nƄu aWG.8'ͅdI&o҉f3&i9Xj/;>ݘ_%aٻ_y$RH-iG敷#զ\%(jvd6M} L[󏆦iڡhJWӀEF_i/͂u 3"S$'eZ>K]&G_vt7|ֳwNOݟ^Ucnq͐JD?=.Luz bv3KYqG ߒǭY1m3̙{#߹^O.~9٨mt5^5q}r$-JU7 ο{oLCu:J⟥)\6]4.`ޥJ}QYGo/]ظjbR]v#f{ԈˮD3Q_ooŞ<Эg}DS '.^Qť=[KBG;:~7kM?\z92ӷҖSzÇ3N$z}7{bwKdEKuƃTtU2{=QZ]2ɡI-4]\Mgvÿ] Z%k|j~/6Xc -dlOKv !֢ƕ}xҲʱR*[c,30q``u$԰)vl rmGj9[~̰-+&YR%zkǏG9N<81, ZL`Vhb,Gdʲ!k= w%e/)(MHDٳ. Ljpr9gPgYRV]r%ۓ,,s: 3JV%S@g$H!L|C./;Zɢ!C%#YB/cbOw  r}81DiIJ$%(R\8#N`$_U׭//P2R&Ce۳$Ѻ?L/gmNW{k+Le3E7 lH"51w9itȤDxtV Qt0Bcom 䥤,Ki֞"GTCЪEHa2CD!L$A uA)؇jYXK$& D5+ȈGP_}8ӭFޱF7,i7< ̆@w6̮هI._bOLؗwQt82vx=S:m}#\f~Nt QG]Ӧ愴Ͼw֠P&wpw6 EwgA$/uLՂk( ~;U7\AR L;zD7w^ԸQ0"rF}d#nڈkmwb؟h@)|S(RԶʃ HdVx(BKYvuUE[yѴ"\cvsapi9k9] %k rf}#0GڈrݷnO ;E2!5ˬ$4sQCU[]2  x9Tmp*Jߘ%҆dH C(ISLy2%%ɓu?+L#~n^mڒS})e}LK#:J<j+?>Q1Ͻ oH #O^M.©SJ4 ۃN#8Ӻ/k.YD-l:^ݢmzLi)1pp3q-Z%tNMna}F%MB?m@ D6"x d/H \[< ϋZ)N Ev 5oYUbw.?#;{W]:֘j,p 8Z=sh3礝NnڎUvwyo~iwuƴ # 5mߵ.>sQlMKs(T$bt[Ȩc_6s ]r"pL 1oB@*%¨egΡM}G)D})g) ¼'du&H2m)UxfSJd]+MsM|ʥRwă`t/jXSxz@]"Qи"!uyVH͎5`R}S1H*a.n4h.N|ۅ.mє%C[3=4o*Jfr~Ro辽g:$egSf"Ob}XW6y0pq16 t :LY@{ROiҢקX* 4z@EǞҍk;O5΍8;N %+~\ +3Ӈr.TBNˣ<"C:W#L O}m5Ԟr\&Pv3OG\`7>d A;Zߎ0}p*oT,cKPS/\mV<6&7LЦ:8ߟ۩ >_de۞/EԨyL=?1cM-w9fܴHp)2{q7MhdZ εj&cnG%LMIq[nU5hY{Gp#+J&͉LB_UuBn' Yf"঳og< Tyq£zrmɹPD.u |/rw0ᾶ-1++|dIQ*Cht ? uz*(^}%R J},T%IZi IUqR@ҎٱwOpbpDNڱwJuNgqIJmkUYMdsH]Rvp4D$!qv9~NkHL`J(O9_1#@}MI&єA0?/6!//}zE*H㊐xLoF&_8 oMG ;4D@3<ڨ&C֦Bmz WW(8 ԏJE#]z*Iفnk$ʠ?[IHGN%SI p ,d?kp3qa`g*k~ttCɝm Nb;L7)`8?(Isрyr-Mӧ|sQ@V:0!w]w0܏.XH;AD6 n]:}otRNQ藻uy7D.bOv>M.F~UxEF˭2.~1?p,E(Gݯd`ީqQa,fc+ ;먡* n6(8,0Y ƫbyA1ʀ Eny +W Rd~^¢.j.#AkT?~bh!4Y|M|_N_^ޝmܔ.Iv48ј8s\Y۠9Z뽵P`|p6(F-(OsY^*Oۺ ۺ-)]:- |?Cngrgu0Ph"+iJǻq"$F:G+wp!1*c,ZB8pɊBA`z^Z2dɺwo^% K>Y>|Wгv % I _`DiE\jLR9cS0SPɜ26BXM@|3Kz! [ f1*u\:RjۭE9_ sVv`nցMQCLss87вR̐;& %$dZY w\h2̣8eA") +hR4qW,X CQe>tydmvp3jN>d"8~HT6FQ<2&(`@i|8-$U,0VQ8ki *8wa9vN'LY͋q2rMfQ!PEnM rw@"(6Q2ehB0R6Tb+PK]`4hԶG }rQdSfӊL1(Z:$F@n)"m) B!80n"R|+%F fkӋ%7BT}Hmy"%=tr*5qyN +kfvSQ5ߌj@ uWK5W$!늝>Y7 Oys oOqT. ͋zaO ଫ7V_*Qt`go.?M) *?\cZV,OB7btI.2^~<~?~}T>Z==wQ,5sʍuÍc2OtKGvmDJ0#޵aa5.Gt4KcYųjqGMp V^Ý0 ck6Mf[!0)Voɴ˝3k%S.Ǎ{ZS`jf6VвQǠ]E8D"jժk- WbMVRK:/ah7b6ly~DϘWp;f5\_IjVgbAp58 oݮLGWWI ՝wOT-@jǙC3{ΗY"y^Ɨ @yVT>gTCK՜C,\ٻme,{ֻ|?p$iQ4m-"|JF2J Ie+>Ǒ:~M'q]Xkһ'UiE@҅ Ui )[YAV iEDzܲ,0!{B~EP ޳f=T={.jz k{t0Yx3qM$kEE9um5zLْf P٢hʖaz(;30@t~F}尓ꈱ}q v6U vb[ XC GMG0|\+.G]{CaG%L#QSSR!=f~WGKIk_E&l 0dTy'ИNRIAeW w޴ot\DЈ|E*"`3MKlC?Zff>pUihٞk/1A ";AD> "Yɓ=#_F+I2wI YWg2O@&Z* 1]Ko:/޼20Ҵ{۷Eś{ ~ɓy#VU5*DU*o*1\25SgMBbm0rWD*g-mטЊנqQ}M8$&Z1(ʍ1ofX>C/6GyF&hjScC pbԿ?WtwgЮ*t撞_ͻE<'Т䶣? qBag XEb[W4+)vf|mlͨ[i4Onƹ%˞ YuTdG3VIGY5Ui \42$5mUHhs|(H)z8D7^Xw|L0[v-qq $f *rN}e%zҫRcL#Q7/D\z}sIDʱ+Zۑ.{Ryِfd l]iOT>9$p鋄G;l3،7'1eu}R.3Q{/?MKU7'#rf!;|rX˲ Q76<*n۪XaC璑Bc PQ4w[y , il0\IҤѹp2p`$FKt1,1S4(@;I8R#EGRrtqOX#z{[ .,]4؂,K--HHYh5}K N s+/A rxJt>vD0MLocɎ0?/Pop<[Pp[^]~U*3Ay+Ӣ%}i6s+14g;)EG5#쬪xC؏W,zg;sVJlkϵRޞc칶iɪ/D쮓& h4(1*@)׾Y~ͯ&r+N3Mt_GoK5k_5ǖ)Gn̦9]Ga޲LOsC`\aiC578v-amd%_!;j[q9ΏR -Bμ9}Wg:R 9P*FT5{9 v~u㔮4e`rՉc҉Sɛs/: \/Fc# iՒz9"=/Eq&S Z@!AXv½49P-,qbcK"NR?Z.QyIB#;^Ap+9zdW=^4;uQS>r+nT1Cضמ˲)XoMwe8}c4WjIٸkY+v|˦/C^F5/h|?pjgjl]A]X^DnUPTֆ$kwsudlkn|f?]l6~”eb vex}1 壳v.G^ |۳u5 37I|.hD?jv w:R0HET#e8C&&HYT3Ƒ$TDdDx)VD!ogO-nSgξuYS㰓Cu4+D }ft^މfOg(ۋ_!tG/1/\<O/24~zͻ<7/./<_~]dX|/_??ۻgj\4y:Kפ;2PY;Z%&]q/>|}>'W-Yz{A5ӞP4.lx41VG\_qt@ťe|jfWM0 ln?!PܚQhfO!Znumbp+JWFAf ոr ^,R)+6^-۵ڄ[_ԁ?C %˗}:K2Y71}߽6n?{p \|ϿEabF?& @T+Uw>[`^w܅A]ݗۋ'OBk^aJ5qg'[7ߤo3?e9{]zM>ozS0!nh0> 2v(:KSxoÛ`T쟋Wc : z_?9oߦ@Qe*'SxիL7ZXCZΈm\g؇dX#i;tnLyek벆LS3n|7v$c: ap`$ʰĄ]aP@9|g ҅?[Lbfgָ̞qf5gps `YSeLO]1 ,<9 gQrs+\mV0ZҨ\ xՑX")`WdmI)X-Ň8Hs.c,ƊwB(%aIlpYzZrB8 0'(HaFIy$B)38F20q1!Bs[]s:X곖EjbHj@U&i!nC(&Gy9&4Ar!k;~$pfm#xZ'K]졔1u_q <Ot|CB9B8.0所oo+sOL|$ AiG,TyrKԈMUuC\8_v))&0I]r{s-rԶ堯l;7= rh Cwˠ.2ysePߖ8oJl|1Ztp1'D0Eb^fJNFWd6~Z3~nS߼ fޱFCPߒՃi< ӺiMh V{?9#"4ѢnwXaaźf- XL)< YXz|`)&˝u!$S(B+ҧ@i)p2N҅}#iɥdH2'U!ըۏͨX&0B @$&C"EA@4!:*2L  QX G)2ӆ_D~P\Oe~zԌg}; NeVY5sDN7^ĴpS/} a6 {O^⨽nxUΌnM|82/~2aFIJ?zK){littПɷSM-5?>j SS>T`X/mbJ; ^4:oyP'* 1kQIX2X5R(xF#PÁDɩ*$(U&V(FH0S I wmmz r6Y`d!V,Zr,jF693& X9]UuecT *ARbt!zP--[;g+:>1Y<+ KY M%\DdX 8aA'LAG$>F#ѓ'o' "#deBn̝C*dIܜ^0f7GQ-E o71R[@Ysv޳?ʣ%_!X2F- XvF *:|a<ޕa01*Iwro O#Hߵ Vnȣ.A5,n{M'^X.D=E?je:CJ]v񷗽XGeǛkVr߽hqG97,^(DBNFФ$ DK)B2@Q9ـj4ZmkpᲃKDn@BKBȘ7|-f_YyIyŻ %jfJ0Q'QԱhq ѸPʩQ#k|bb 3]ZT>avL7`6a"tQxVSp8i NǞbo?f:*SgVgQ\ )y3G|-?8&~>`pПSFMtٮ9] RJ XeZ P!I\+;];(#ˈ/8F) ǛS9T9rdlW`<`Z{/@R2x)4P8%@(Ck6I7K+h)PKFqp!Y5{X΁a HgZw/+³MZ\eՙd d?\//̉Da^(K~iz 2^Y,ׯD/zv|? "PZilI}3I'ٶס{˱ಞi<76Vy|um~`xaw6+f1w+O鍹T)_E,bҰct\C7dfs_4xx m!LPD⎟VBG7?|-<hg;BQ;$ 90*_\U/`8Ӟyp,*s0$xᤎ޸DI J8@fӡxk:Q{˱֔S^RݍjBB<>VA\J*? e^n}̉g;TQdqAHT^ nz_@FI2!QT" S9)rsTDyŨ'$mN@ -E1ۑg4&hZ?;WNW/W:lk"4^:U{~6(GR 3^py#M)G r5_QU ;Fv鬬WKP^ [V!/&?{V:$[)v:J50fgUJ)@n@X}'*ܴ9f Cm N[RR).F=ׯw!T "c,ҘUFw_'(º T_?OBZYZ}y+pU.@ \K6ȫx;ڲԾǟET~ee-e) `4C-b+X峯joƎ;uePKA/unp)Vp>ŧyN/O6+̚9NQڷҢC.kP"k&_C[Yrg&Lڥ41^V'ˌxͬf?͌qc$i)]HD Q}؎LPaxZ'[UN;U4s%ss>6p6w %kԈ̺oy(ņ{]~-VDaugߤ+ɎrvDv;؋#%G.z%?t'm-ZޗܵɉgW Ұ>>;o7D=[8tߍY~FrCN^.d$Yi < U1IGVO}ɺZ%Ga سOG65<9/cV},̊feoHUs]p.حݏjPNt3B鑗aR-?/kMy?43u('SQFsC8FJZC0.$45}ړ'HEDx NP*MXx%V\bSJl(--+yd2q˨W@\j!QrrWSJxJЋAiTbGcQทA%\fqsФVͲܩBԥS ңCҕq]dx<[idkBxwOan@*p#;'un"6#Vq}עj G@*I6@}c>bR-VT2XQ{K/6 "^tس©p=`BL[u’O'@o&Rt%WDEY32~ ELٜp5(lo3Aˣ+Ӯ"ϖT:ԯlL泭ԳI "Ysple䙓 wP6wJ7JH|AD}dAz^QӶ7s|5+m傂Zk?I9_'OUY2!S@T>ŏ"0crz@|N2rAHTIl~)Hv1m/7o*x< KLT1EqA"(VJIR˃H1Cqv[qRՄl"E yGO b # &d)%K!r,rV{LCDH\4WfqZk"|)#6Ch9)P%-bgvӌFG//.څ^Ak^TT7946i-q,C{8a->/1nݕ #4[)kCb} D,2Dz>\ހI zW܌*¸Z׾3ـ |tEh"DSqv33#c.?J\rNY02sKKp+Tb[z8RiVf ܿݖ't3x[0|5O-=ek(]|yyXN5s5tzPC<.Bɰnj-2a;B_|!mYuϴB]6xH+*A ?E;|=:Cޏ+q$G/ N5o2 ̃=X vbQٺVU3=R2%Pj0:` ~J3 3yiv}yۯ7կ_V~Ŗ1iH+{auIn ײiT-Ry܅vfF-h8__|>W8qu~궫UVzt֋qHkT(vcn2meMY#؃(rZ*\tism^>yk"kgu:$ BZ|* î#=_ t}"Ă"&L*'S%HH)d92'XBM\'_Ģ>2x^mA_Ӈ9|9ܻ.3F]{Pymԍs.Q'?~r@w%r=~p-loӗg Yt ɭaZ`."%p:lLn>'%/᧧}z:*({U3\l#T lUn}ݗ $'r8؍͔@Fim+qb ML/ lXlA!SPQ`'۴ƻ -#Z„PQFTy3YǕwέ"3CJ~qJͩf#'K/i ).a;K).S7_piNpH(M,%\<2"vyy*`7 |Q *ƿGyjPzcXWT rjo0XMCij ahަr;8"z32E_yŨ T."-t9yG0PS2ʵ r } IPA.ã'>hcPYDi'%/ '_=~ ar(C/Gf2*IX(Uu%XrZl?Dttevͨ@( ҍAxUUl!¸ɏz%/ߏуC=h'+/JvF(#˾CIv*;cc/nY 8^䒩1%ZD_\p t8!(h2i9aLBΧ+K{'-1Vt4)], wLrAئLnaR#&61eN;`;ilWKh>o  c H7 HVF i9IS\K4'U; 1N VGW`=1](ZLiUF%Vیh6) /و30dmThpi$^1//Y m ynvLE&Ta:|_t}]=Vء1.RZR%v #QB6_ۑs ue(&9]a:UӍXT7k|SڱCr=2\۝A{/8{8pӀ,zv[h^& fz28JyXJ'XHcy L4r/x½Rؠ hn)w{5!g󑺒ݧĽ;53r&st-[tRe1x_p*q92U~2(N-#:*֬d&zj5VV JƱL0!(ԁKVSw$d )1kObFrlNOw,Rma\  A" (J)TP&R*mhҐȭQ4yЎ@hSiZNy ^2ТJN*ٴF題UZ9z r5aX笵z8o;8Tf\U-KZ{jUK eT̘MU-}ŚgZR2#y@Vs(L 7[XN8ZiXIQռ`XH@Khjc7fb)%IbY?Q0`ʂZ/Bi{pF*I9/%`H1ӯ$f e%wRKԁP}Fj\.#'eGMMWP/P' (ܐCqj 8 jn^.>_E ws{-?)K1 TI{K[њf#qxcTr4ocHWy_=)ЩxsKT!0*1og]WtIІ7O# yzom}el*3^c8|*cQDx# @ m gfG( )23J ap.Z22-kͫ]'ûNw jxU"ėeɸ8ϴ X*^z](rNYkkMחlÒ1}|Uý 2/Vqð^ JI6)#6oOJMN$󻰩vb"K? -anօ+|ͼv!1zWR BSVoI9t|/=*84QieK_2 eHE5VEE$^ pk!h/ȇU~ZJv@QDT3-1y 6JC}:AGFvu: c:c`*[l݅n#}W8׎c !tZ6r"W{ζe儩9e8RQ 7$VP;{*@zl$'4EqJ9|RP^"tNve={E/0%H]eKL_4ȿ+ـ$Q[8\] ta܀xOr:pc(VwcNDR>u$C-EZʸ GVU 40\4~lԵo2ܜ°psC@:iXa8oɯETç dr~ TLƫr] e?o?qn-}BiQ!8W8bPk!  ]6Nѭon @ftb{$sTJ'vڞG;BAɍU#5x{W T ζm %X? xZ{*3V7n:TMxip4?KIfW?7†`jaV(<ڄNuQ%u LՌj2/! *.ZMO2>8[zZy=b,b@,%u Ǵ;a)L6fuy<-dTPumT\HQϔi]*&b09h(mWERئ$$ր=?WluzFATf~1Le,8w\m\)b<@FS_h&U$8dNrDSQT- 4&.bt%ԙR@Y:b@jU>jEɐV8֧YaӬxFvE 4cyVa3C(avkP#R qz! 9XuOҞNJ'=c*#XggVJ Jef+ 04޺ ʁkdk:E׫iia* O{hW/4EۤG ]|_^19ǗwvdvޠpcHUE9)x=䷻_w\5h/MBG #*DTBP%!'/Eb0u, 3%[b"\T 8\TuNτIZp&vR|(P[Bq\^JOJ}T M t2tU}3L%37IE%Ӽugw`ZBexu_nJflF1$2urgu5lVP.!߹4s2a n'Ӂ0`"9 H*{y!xOO&e)CnqdvI~Ю}u .kTire5nT -hx9ƧUѕY/FO9m4*zJr>ttE0Ը;v0',U!iuHL e'&K'[KTxZϿJk!"L]^Nv4z3%qn\GEqnpNht#F*8wɽUy+ФQ[[ݖk|az#y}П64?{סsn\>X^|-;6@"@(DDc(%~RAH91h $r)QH|0Q&h%2R;McL|=ol(XLeZ)b4(u"ȁZh1 ʱ ӞV8y6']`ñ)t"4*A6Lvͱqs֌h5/Ui-e5yuDհA߿_}ANR3xZ JZJ uD䎘RV G+Qa` >E`2r頇{Pw4mV#iji4SBM322GӒ΀+ P/K"h4*5yzybqŀVG++X[Pi(@1Z7Pn%Dzցo 4Bz.M;$珻r26ӎ{ZMѐnӗg\.odě07)Ip)msRO?嗀KүAUHyvUvD|+-3JV]wStU*'()jQ˲דN%ܝ3zZEi!?:.>h>O>'k W~Y _/i'H]|][ۗJ +f0>ś}F5j'43VU`Di1`vN MIbOa7}{B=z׷7G />zE&`^6X8i#_sb J>Lkk 7$uOpp1Os,զ]f4S3'd*9nc+Ωw a 44\Ѿbmg)pԯOuQ.~d61R\ +]5 ` a}YXEIꈋ9:*&筇̀K+ԝqc"bw2@>٢@/4M$Wnw}%H#{Рq,jyCtUa~ &(W1im_Ҽ1Ǵ$f2M$8K&;ZLw1QiH_Լpf|fQݫtNd.C/|Icw6iHaTƣi K$ 71^ sD1&dh9iTf?2d9D2m)+ sMA ֽ"/T|\ ŹPPo0کSp+NJa-(;+ʹh[ht{!EfeZ8 n+))}T)\y"5P3풚鋠v%$NUS>]2v`"J2ҋ;!tw1ҶĘp-XE"x>odlKR6qʃ nY{AyTHޛJpel[8y^$ Ǹ"%yca*Q=ph J|>4 2o|3 R֥ͧtd ۑDup+VC%׉(A2M#`̒BdҊ(DIJ۷v}&E-= }@pY~9=_hprDtk w) XѢwMEp_O 7q+O8:D888AĜ"TpAjJ2mG\sɥڠU D~9w{1A)Av OdCS<ɬZePi*v7y-s].Ly*3 dπ G m25Gi3L6ufiIdgHˮxdc Evc_Ӏ~^Ig _%= U맻Oj$H|C}Xu'b}2wl҄6Z`g׿OU6^E:t܊'I#<|qre<]}V~[lوqS#8vHeu냓Y=mt>2HuWhiTJY__`p4cwNpI蓞U/8>`8H3[H{=á)TfkI.ipL@8 tcT)8itP0Y+ıs9jc_c>8W?|b(Fx:#ˑp,)LfH(r?)TPT&"7T Y^xʋBH|Ѕ rȜXOD^P; cg=4ı Bj|okCΔJ;V+ pʭLYdy0DT %X#JU * rQVpÌ&@Qy, c +`Gі@ɿ|x oa$ Y9m쇋i?n~8]D}mFPis:P5bNS TJB CUl$n'd8Ÿ̓UXF1ӏ22(V2 ˤ7,'yCJ6p^߽!= JK?ksӷkr horڥI7-@mJ_[PdR+;_Rd(U_-*K6wp_f޿?ĿxKFKuM4|?<@?%.cX2x,x{3 ( [ - [Vya GɕKKxy2[|OKߒ}ͮZO}Z|$@]ܢIGarŲk(umoޢՁ(mA ʙ5 0aZ*xͣm};Yj5FȠT^VAnsP' g7kf9qI;X}b\Kޝg R*!8ý We 42tQn^;]%[떺ƻ^\КRQbfiEtM_5ܗˬ6y&@,ZfRj}AlCw+%tPy҆hN hTNq7}f_֓V;!MaApϨ̐&9{7 VވJŇ?BQ*j'?-K}+h}dJƾ 嬫î sM˗Jx-$m$m\BYj [3>)ޙU/M+6r1Hr"eHD?99I-I7M9zO bxdeY'􈕖5f>N}N0U@TT6 ( }:Ђ̜ 4\9!s9x<+!m _iϑfIho)!a2}JC'B+yމ3`e7z %wkt!Sv@OzE@_/"OУ^>tc$==FZ`wWSՏ`*C{!{ % Oj[I- !RV[!JDRrf"=emu+ĚA6!]/Ǻ@8yLmݡM01&}P[,VUz]&<ՙ f>!;/ ,3T\`ieJp,@A4xÄqԘ‡<~H_)2R喚 B{[s Pd;k6p8)=rWJ[Vs`3E EP ǰkUw>5+$4g,)DP8ҜrYp ?DHqDjcP  7 1@lp 0T,LRSȸm',5k-{b8ML,ck/w~/c e D,KF7s P0Bӥ.(gQE+D!Jx >d^ӎk+p^08Ra6G:cR.~nџJ|uNcJSy yzUL|N1"%;9M@;ku%,*Oof7%>CwN%ypt.Kw5nLVOcz R]ny^oyB0T.0=%V[DEt܊#i;0>Ǧa>Nnm*P9hW;Ŗ Lt~JF??8ko][IJ =i|#Mw~@hHZ\6I{1Eɛ7=2yӣ&غU}K{8ۓAK$'B$LS}Xz:^56؅vpHΏ:@Acz)8#8aJ/ou=Ggp0nB܅]4_}x\};=S  )P^hgLA9` \Q@]}@KPigt岸s(#[xu^MF99 O&%/o*<pV)$1_!9̓"I&*6h"Ά&R+0`kFF˿ߝE"`64Εt>sXӔg)rʭc6W^)J<<%q"YaNeZAFr!6^Z,F fM`ƣo"\.ۛXMb\;1Ze~ߵՙJ[/Ox(.^>>\~Vj0_>=N/㫀^ݼ;DafџIt(Q1wL|L/S-ySsNX?N>jl(4p-GWLsI(HlSCqx(GEPqF_P-jÌ[≮?~Hlma kÄaBYqz fT9$*`eJižyPѱDIٝ-q;)p \<_ꠅzA@hh) }BQ4(xF|FUQ8R[b{x.aI9}AX.-" 8}A;TpL юGxN4RZ J֭ %Uhc WhPЌh$[rJؖZ}Jj AcHH N: h @&_M Im(Ġ΋ 0ro !XN|:`pQH@CЂsFQcI!(sNPYA#8E;*ZYME\4*/(ZB{rPe{c)h.=ʨb@jJ,_k#[G "&0Y"n(m*~BUfZsl-gRT y'gU2cpApBX-vL($nSh [ T눝4WR;9];PHJAeFkj=UUh1u;`l|)k*'8Qxdsz0|L7;{3o~+7w}7yyjyA.l\O¢N>'JH;y3YQR Pr?)PB=BڭɊRPOkGF@gHOd_#. s{ $] JlU(^MڔኇgaMKL98Nĕ" _e zuO=G!v.Wzny ľmlq7ƛ- uRDmdzll$t6V/2CLJuj}pVs F޵6r#" g#?3`qvOgݗ M-Gg3 OQ-ۭ-vZ/fbUK,I߭5jm_}kOHhyEb!']ߺ?G[ZgbTԞժ!T178;pQ!]}?pthc(:=}jpج1: <(wBɟ92x?հz.7&q]g>eaBH w Cܤ0y0[c,sLml0A$A%\3lxoa5og8X-kz' 3܉SPB)^Fx]l;J.(29>l)JL:TQ1~k6XrGLۘk/Y c=!GlU$3X:W: cUq)v'gT TJu ϸ$C[.*MǸ/>̓za2ǃLK[Y!8+RYLkҧ~wԀ o (#>,7' 2;`(d*5::)dky<[u'¸=W?[;%@[>#lS/C.Ņoy)_)]wF׈0ߐYufRo( UER)yVY-aQvYy!&z~4QW4?/{: S?yp$Ffwh2?u?xwBᅭ1Wj5s!R)/`߆Yg'䒫癲,sly b8\OeydZD̥'hyd_Mr"$S;Ϯcj7u@햊AIv;\75<Ѳڭ 9qM)F;ⶮh!-2:FvE)kQgqnuHȉhL1Ku4t얈A1qJg]\jH{iDKjѽe#zUZ˚]:׿_] s0 {-~uA >BM/}#5tjdsc|`g??l.w "fT BRq=ZaWĚ㺵ـXYKGE϶v%ѵ;z p5~2.I 8iY,JYB^)Z`8Mw,#eq9! MqJ c*>eF,=Ja$@8\F9μkUժ9NXev[kV@A\Bhz2Ǘ˸^yT#*uIAE@'f{2 kR+%崶T`FxX4UD/.$y1O&͒, TSBIvBOrj[?s;nAÝ>t6x0r 2o^ޭP -ctr{J." ?OM7e嬕n]Cۻ%t[RuS& jtRmEOrnEORo2kmݞ(٭ 8-/_k9?mQQj4_= !Usl=}nOη 4; -p_}tLF]pTr"(jm?},{5Gd[j^mfzW`!U]k䜘":ߍiNvuk`/ Cu\r1eq.Րr1vRwtۯ }뢪9cM=*D6zC(&Zؽ/s*|+FQ*˔F pڣQpaŴA&Tcd0bP&=u*E@<>6MaH*Y 9qM)QrִAR1(%:cTn b&nɑe[ r""Sc y!sϟ:V؂ojyO8 !uOެgY ]czȒhSb=iJ6C_fH?bA(MlO '1$IS z]:+Rҵʄ%ScLl юn0R(Ō)ތi7B _iYRQNz%zN! Q[>h:PSB0Yj=Иfglϵ< B˼ws-ZB_VU?)rMwazzBպK]şp}󯿗"(Ņ-ctTo n)Łmi%ϦBY]E\Pw#Gww{-"բK Zkgum)kvi^ƢeY*' :9u֛6R |1MQ-AU5 l7Q?E#9bqZQ8uKz 56ζj [g+vh!ɺ[şr_ieޔ팣O+OުɁ1}c޶·"ծ Bڕ;rJn?=]!Ts,ZTq1xda rOa"<&d5|y-L ΀&CqSg01)BCLw%F& =Yt 3wsuwPόr00Fn 'zD="<\zwB$W|d)`Kӻ+leFY.";\D~ߗKqpoTqKyV^ƈIk&vgTS;c諴'We=-*z9!j7WUUQQ}@ .J`Ry7K-Q UdF Pe&3x^ʌ ь^(N $_-4j*B|Kig{=|)[ZI[8$dbk:R~yt9] -pD5 xp:鮖T(h#t,n&JW eJCn%uǛZx# bCw^?%>U'zHw{͟#Jd23PYf l`FPTX}2oX]5ǁ Cؐ`q )b4rBPԜj0h1>T7Z牺uraU \Z:n%>w*2bgWT (q2 F(+>D aV]I9Vw#D7YM8xmBudS5)Zt%TCq$8{"Jr N)L^&Imv8v5/]0}Ԁqcfvq1LJODBzFb/p|̶f?T |So7p<.Rl أedڨ1wUK3#%Q3g0|c}} ږ#R-2{1^Zնn,y%_ )v2yrWѠ?N`uڑ(Ow7&Y;˛eQ5;KRzJA8K))), e":PcJ)O&U\<&ع,>)(03FhVfqe2Ƃ :TPdkUЂW^`Opb2JvƂ: &99%^JRH[:,UGRw$WRelǔ IT 2Ëb71=}ML,yS?;OO"mxO9 p6k˧{w 8GB0OW?~L^1ku*HZuɽӹTߍvuŃlXyjzݿb&\ }Oep{u `5aGM >ϸC6dIj1:e+i8O޸%6xiQ@T%GQ8e;ÐY Waԋ aRZ!Bb]-q[9ts[z1[BKfK(R eGZ2ÖcLc'$B?6b@@{;âl%C [`r`cണs,M% :k'ш;x2$L&E&GmY ff1Pfj/l8*2#h 0!!%'Q(,Ii8 dy2 )2NkOtdu!* $ض q5Uȑ!cnieyF#` M6eJeyieQXVRPBHd]F 8nxR.REKػו&ɩ Qv^~FLzҜ .K.PeyBq;ᔆtUrhIɡJJ\=3GsrWƐޟ:_^.zߞxv\Kٯ]*y]ė OLޏs9JF}hm>r,k%-9W}=~2-? #<93LG#o2ע8^|؉/C??{Fr 'g_ !Iu6y@鋬JԒ׻R2b¡]q8SUuuuwWf ~s#0Dy ёk+yսn1^2gT5 5zKTUo/mqe:ÈگK:jԙ"ΐ`q)գֳ"j+zVҵ ?yxp|fu?/?VQo$]rE-m`lmltrq.Xpv,2A.B 󅢨YWڟ#$F 3/7[5DH?Bp{Wcc1&!mOjF7ScsXRDv* eLp)TfmE/ցW`Wי hr7uXDOt|Z$=TsFitkdnuSՇ8b_ƚjxNQL3NT=Z;NK+9k׬}' nVٝb:8ٍSD%1u}X'ԁ_#UsӾ_eQO &j|k N:|cT wj@͹BݰP!~X9͑V_#Ba1pIyVZ*-`:! ?5m(*蘴?0bʵTPJXPGLE7%ZO{QEYP1"mTϰV St+t@r'R1iD"*}`OpA2'LZ"cCZt$*'at8J#F.5Q+co 7xI`;=]c\"Φ>(y~c=.s#4UC_*jHGi v6cC-iɳmȣ>t&$:hX) [ Met7o>ds Aݗ!bg s3ԅs|D(f]:٠T\\W/ AyKBS\bRCqJg!%HhIQ&Ʋ5<0E1e!(.UX %q,4vum%t)IPot;OcIK|x!:{}tc}y\/K=0= *յqYZx1'5V}(ORq zw51rK$BC;BwBQo! rjRLQGA D yX=\Ԃm3CISgM}xP@lnKHLjl$xZy0qXzeΘlUF@9Ѥb*pTYUkcZ+A{(FAa!I]C8U%'̍ AӀ#e:K`r՚Iyl̀ӌv(#e*p Qfx3k`Qe\@Vnsnް"GԒ<1N A`3l=Q(#cKZ=2vi. Ejﮛv4:|hj#:.oDA={:}'wU)Kω 7('v&]Y iuHےt5o ָҜg5SxƐ!g;72ۇdƊ$o9hL5w䓇-WU.3u4ͳ~NCxCYKuP#*s6.{JxJ*gh/ =3TGycbrU[PhT#W_}-˗2f )b I0)ޕ5RdK֥Gd[2BL8hqvd1`Zp6.w'SLhg-&b^wW:77nA|{]63%5J)Rpi4v9be"D3)!`z#MFH- *id-p9B%ZS9抈ROZx7Z .{eLN Jn/K]ᰔ" jS2EJdܛE?86fV[J@|!)<:}V"0G|69gw_Z0 BF(C)RxS}'U:F:(:h,w$L+O9ہ M8G~w8c"\M6cvA@0vG,8ݡD_~,}Fa:M򸃛dxekvK+{Xt YUi!˔aOn ǵp2 e/sG1u֮t 2M5GF̿=r~fv8v<ύ˶6a1(~ 7wGl> c0VOG@//QGÝ4 н髇jTb.igȼ<#{?ZjjvQ?=.`8*#'oy pMEG b.VtC|ŷ/羶yyY@+.%yuHa #$'XRRI }\PY+E5>8TgZ9& )=H,C.7Nzn=VTփ@j~mʶ$ 7k WBk#5G.fQR !긤9MQ%'8bDlSz n:L#LLRZ c.@)1̙)roSkZ`:)wXK 2 l,aR@2N+]nJ^I]JX{N>c``^z\?gyFEV4q9Z,0|՟d杻:F\1⺎u9ZbIFs >;J& 3 Q4;=CsBi=EUF0Û*r5}2_!MyUO~35~V͹d0N^ܟIAi!vsf-!!= h62ɋ$n=d9սD+1l2 /dim}3to瀟w7v}HkE:fnwJ$݇n,wEYTO3i/OC /8W#Fs8]C \0԰R'(3%=?}ͺJG~ҫe\7?b,rv+nF!O)Ć܃!* zYւN`|HsB#9AodP$T19GJm"s:^P0mFQis0rKun4Q흇g=w"Y觿l|wf6n";%B?>QZof@z8J3F.ߧ ]7\ wm_a^{ުJL3ަ\ z%dA]e!*$!k-(!Hz'273kSTGNH}+J|:Q BWwj Nл:M@p4q DqI6<*3~e~ @2OӾ (1#$ Et,:ZGhڻûLY63E'TRPU '%I{Fi3aqH@}!;^g ޚo}X jԠ"M*5$`HLՋ@ß ;>5$TVR V'ܲX9e@B+ɵfz\8q8vSq[*9iē4Q&6AMfhiJKpeiT:F0L(Fp.  fNm!IX w-n 'u[ j0XWbn dZm9sXWӵj|ʬ&tpoMrܓ`ޥz,|_ǓJ8WԢ[?n*“jł\uL=uaEQ 錤vߙ&䭯ŔN7PŝhO?R@m?/!,r,.R%`t&ϩYxSIlH-c!3FD"d LdyNR`q;ܽ8Sa'ň8ZB$\WɊ]?Huv# ԫhEla[M |PՐych/R}' |~ޛ ,I[&m1N)QǺZLyŨZ\PLTuyqS+|YI9$oOu8sRA(it*ujleu&LmT~b " JxBQT35I=}WiX]K;ʳ*r6pyi |V %fs6p SFHPj)(1Q ަٺ=GwwΧ7^rusbڳ ,}Ç׷}|Շ۷o.ʎzn2gu >|xb2tQW|ZL&hzM'{Tu :~n8 D\g&]~Y~[\% :a[=Lor[mP5Ǿu0sFP" yɔ3ߡG6h{pn±;R%1\Vhsq :$ )XnI|8r|<<8ߋcYSd>/ۭonݺvo7҆'$Ig)ձ5T)8Pp+bDXͣ8N$7&NqSfda *9IaF vd tliO`rosfktf⤍5/և5pՏn5M"7gOιΫnrGf.}^u>&xONa?W;Z,%(<:;.. z;cVѕ|ڼuLghAi}.HF.Azx5gt܁G{`ϛGh ׏0wG|A]O}E0/@SNS1DIp8QR0a.:z9XQ;+Jh.X ŴU%粢·tIxi>6a\.7wonTxur>G_^zcayǽC$|w"WX{m_r 򥂱`R±X'Z'IfMd=2!LyT< VU0kc[1,%g$|@qU@㹄 GgNK\>zn|UHq~>xB82/CSv dn0rኺYE@ *m:NB-Tʸ⹌1<ү ד^MٌfI:%LIuTOobIYeOL!Q֓)E_r =ӻm\>= TFXTj)T]|y c;>H~St70\Tm& A#$Rm]KNnMuj;i8l~,Od iuIcX_Fߵ#6&K5A}_w_85vUwس ;u; Ltoč'7N&n5qO yt 5%ՎX)G]Sm% @h}޷XFj{ ugI&ifE82/j,)MfE)NBzEی/u,k|`6 [UX 1TBNH)Ф֘a< "`4s Eg)IBN{q4%)Ƌ  o>wYŸ4kqVj4u{m@wȞ'Yquz ;iSw?è3ZI9rzf s9'b%zwڨrsyS4rȞ@W,9<9?zi  r7&KLť|O7{*kJBQSj_ PH+[w"@_Dk!@)%9Ӄ3 :. rAiA DwT0B$CD%\y R8S2EG6CQĢ=-XIBf4~Xx /"Ay|AGD?~x7{kҭ`+ x7Ąt]w>ySxB/>O3˵B¾ǻIj k4qDg8m'"F(g=^ZpJr*8WUjW K3*4c܄@$]_@m;syCA%(h:mMc X+xqՇ5Qk57X$8FHA FkmEx1PO\4E~nALPics 0ZTT8Y QjDL3/Sz&+@+5h\}2-6fVn]ׇ>:Æ9_Kd3ж**o辇+\l"1|VpY%fwYP#Xf2;,8|KܧYYY~qEA YWcpy:ϗlp)VЙ)ߞp0EU֚U\)r a&"Uіæ 5F(0?&^PܦmYcƬ-? W aq2Vt<ÓmEfcTT04DW.푷R#&vםLVG8WVxJYIb1FQY5۴;Yy=5O]M‰:weqo_/ wڠ5QDټ1wƼ_s)E/<<]-bo0% t 3m6܋ΥdibB1oZsG  #wuLy D#T9ӂ-Ի}~ԃfv!e 6w`^m$<(fIP;[5gjM A!jGQp n}ga[1>ޝѼX渓&G2]?xnP|K/ Ldi U ބuhÍ՗4}XC=,̄K&;L ف*"F˂zqA=fB).{VIXT5Ap Ȁ7l5B5VVmrVtz556,w}\L(z JY{*kNDZY%J1:28ABkXN@x';S;s}vġeZbO BkJ9%rw_e7aa:qGS_a 8dd!!`ZX #x2_(jkOTQ;f+|M#9+o SF1ndq0+tz''r7Jphu!>#6*NɝN8Y:<# w<;>X77EgWO-<[4apL,ٿ7Q#NHKBaO ckSVHa)_57hi6AZ0R^ x2 oS[k:?ٲ*RDCD7%RhBCCqj1N 1Ʃ!Ƹ;$P* 0ɀшt_-.'F9/ 0aȞPRPфŷa_yҒkF|Q#;3ƩoƩoܭnGxDRτ"D%`c̬5ߦ)P*C;ϲcBE۱tXK_VԮ[i:Vp2y.iOvn vBR#ȉ3 q*^=D': Ĺ#l!ri:_1M!)WI MK-Ԕ.5gXidcSQ2;owC5W2 +p gN笩PVJ0beTho$]̰eGQ_T8=ZJ Hm=i)\0Zu>(ݨ෸iFnRm|e+ B/*'W<$ҹ(*n3 j4XJpїeyw eCدiRPS/$[Yy V% ֒ 1CE<ƍ!rryc[VO{rvl Wr-ěx%1cdR;w9fCDI&Q\)L$+fżC*hb#ҁP*B@]6Y@$ӳeY5TsJ3B[(15Kh-ntE& qɆj$L>nJᄐ0ә_|s$twaTC,lU8X˰ Yg~H D:7 L];=hlp "Ka#|!gf91عsZ͆;V]6_˪bᇋ@M5 _Fs۫_'*9bw|܌;O.=w3\>_'ŵK~p1|h\;d h0Nr;SBͭ-zsL&SC2]1 ͡^ZJUbL13gQnjeT{<Rt_'6֑941幾u [?sD:l‚BTj *ŏbYGlT∤a&'EtEa ^6[ PTpF@NPT:xٝtJT1|QP-4UD7kQA!8@ 8I ) < c숑\ڙF1Fb!Rk ) ÔRA0QXԅbɥOr_WgU-䥥U4k[*~,E xZ1A*Q9=$% ďIPh3PADb]PA ՑN6 S"a1(TOdrgTl8lx%`DGtH6y9KQ^ѝ[4, $6icܧ0Ͳhz4 8Xmtfz? AGʠrL-g7$d^ bpvR"0(d쏝'!|XcA3ng`߱' *ؼ]&TJZ+8-jkR.PڞjJ9Q],="URY\/ۓ%kй?YJN_CMzD5Aў.Aݸ;h7iMFm*xzI(YoF{5. u𽻺Lɴmn^ivSNmRiLVCB^T'POߣv+ GtJx<ZejEH82=ZHy:9T B룻uS<_@>[BMg )~t7i? 8N4,rNqB՝:/ͨX7d.o*hڡv]^Dضm0A)v9.Uw;%ebSRu'(%.XGݸ3ڏaHwmmW\~93ST+IˤkRӕ{KˊO($H$5V?666լc|~U.NxU1<7Y)|C8HS5Nܗ[w-׃^'?miOdt.DN29ݚGMi\ Agt$6{{I I/M 뜳cX=_mC)GwS("qlW0DZb2[l;_,9's{L;Ń( WU)Ma9|;9h %w> @%wUΜyG_.U8=|ryo>-oV7su%Ul|V(%A/NU0|)Pߑ1;Bkv7'G~Xng_8 F*(MWeal @oU;&(eyq(QpoSNpʇ(X;`GUZQoi)Np>OU{y~;&OqjxHOQ8IOƄ_Rz++շ٢p}w)pB]PHemڒu$UE .Ծc@ Қ3(,sb#Qq90c(2,PJneᛞݘb1{,G/I|Ll=vY(pĔ^cP)"-Amv Jc4WzǣvK)㇎ᇠ1^}BLS.Pg ot]4V<\w#{u7nyk{-Q%D1rΆ@%9 .*N U\Zrp߳|n eI9<* ˙V\iW06-/rʩUtٱ B<ӞaLe;fӃ8Q7"Hϙ=ldk c}Cˏ4 [ϵR#e/\ÕILy9p Xϰ׍hC Qhޥ3I޻O8:@E{}1F"tv@QGIj0px$gQʖ֋7[?aHz ^Y)[s%(.)#Ĩ3v8\mvgF%lRlmlyk (<\JarVV QۤT^&PIpm".B`|HԶ s`o^ߛ*wȒQDK'e"qp\0vto'avn 燏yv bRI5n# OJGټeDNk>˩QMێT$ ;☠(ՐdN3SHBH3NJO&ń,-na|ޮL@`SC:-؏r i(8q0 j8A}ݘo48=G+NGS~/ (@aCif/> |P ۺLp2pbYtaaJ;;;E1S|vʔxAXIғe Nҿ}\tXZ7s*^~u/ - =׷WOj~5WϯiW1p%g(t)H:|ūDw_WKx Czmx$Ly_840yU߀:F-2Wt^G/1=<<`GOSvpxMveNr?{u{07F өw~cHypm[lQ yXx,WFܿ9G A򷣀rrᜳ>|`.={H3 @yUH$vsJ %J2K``+ FI3 1$DRjykC7Kbʂ SaR}@콨/}iUqaP_T7_.VU/W%ꓪǧbS=>O冝$~|lRY십s&oyA̷y5$rۦ܌ ϗpzϪ+U-*WݯʻhJM~jKR1:hƛ+N9̻%!)٧ !29_fd-W/\A S٣ ā6nߧ7 QHM˂` ń-?Fab _YZ'4 %@KxYMwDkeB+ ~zKӷS%-Yw‹C;2ٺ$8NM)GRhnQ.P%)5witW#JhߣQe,'͔Es? CȻlR7w20;:SQL[wx>7b81X2N!/AtaӃYߌ 3'03l #wAlyp}Ō>6VeYV{D_g_ c:=Y\ `wkV 4Cw P*+&bl,%͊_M@U }~x)oObݿlksnXQ [pĠ* /{w{) 4Ei͋ - í:zuPWHs.\!uw8+ӃYߌWH ( EA+LJ E ӄ_)VJUh*_QSh(p`墳@= 4bY[ J`A5ǎ PTI1<tO39XzǮcU?fVh+ xǬ:w@qB.`?f̶8 1eM8 šbw;ٱmaʯ<(7)r&'(7kRUܖgR:dO9 QLP:gZj%@*Ԅ{퐤"kdR ۃp wZ%Rqi(2V!6T(B%ttTUҜvP٭JZrVc!IХUbBJ ,,0nfDs'睌̭ΤR rs+k(c=OQMQfNG2wcX?RN´rdU޷ !P2W睤SK0oR/owz1 Sj,^_z'  ae> T k !  )T6cpQPx_{[cTU<䘥ZN-P3.*3LER}55'gK.Ÿ0S6 16͗ Ǯb;{K;xi%e~Pp}>N>T֐W-HK)4LA;JR,ᛣK'[blR-$yB}zpxrD >6$үSG2sO.8-uIݖV y_9b4Z̼C1Ȼf} -Rxa2~z%{= @ t?}n' h$e j{ N÷/N]7!;g 7?"OfӋ}?W8Nad*Vov3xL_0%5|lXZ<jUSU9F?$G\?pN6dZ7ǔMtMdS&1`m*ki$1X%'}@"3gw/0\N><|Lyo>-_ۋ{]X)6ͽΖM~|2)mmnjnx1Rrcor }Cn;NY?.GV9upq?{տ(Z:Iͯf7WQݝ$Jix,;~vj+i~?a:ޟ@vBXDQ B# '+J/v"})3orQr_' cbHN:^Yp@2]9}IKBqQcc[턉 GDkQl2%di}zBbixURGUQ$=UJ3E-C fGN^QEI:Dy xzUW 8Fyuޞk>/g dCnD]r;EvS&IOn^{ k3x (X~5@B>:I֮LY-wWgnW AqB9oON:{?gEx|i| }=`B@ k;կ%DyYJt*hEC+Dv@Za Z7f>3y[!X$FZZp.dAe~XJ0 `4 ]BTZ1"(RBJ!Jk J f޺ ާ2RZ]Yoǖ+^.f2k_!q2y!2'(pɵo>$T5{a3RĖج>SY>\drk$\9 SfH UZh9!< JG89$0.QkVHn*ST[=P[Q[Cty<]Cw;wI4ga>>d%Ҫ=z훏wە$RvDW>PzhEݗv\Q%=|z&H(L#D:$>~'T/\b4DR]A2LP3T4S]|z`@Rm0 WQ!R>pkWum֜9wWc]w(kr8š۳lhHy:RV1iøaEpp"8xDc5Zz?6 5b>c$c-232(5ә,2*I8 `?>=aS\0 ueP@cc@ 0g %jw*Ŧc֚,xDZ48Pbΰ̓ uuSӧIo=x<8ϱe)K#FyfP!yqrC8FRVG P"\eX}Oe8b; WըVObyjOŢ{fyoc9]=zD ; ~ Og&>.no7W) +~?BJR3Y-,\+]}$#AbTDbO3w_>6ƱQ*P$T5&KNr^kxa%.T͟,"I`r^:ch뜓AFbWVjM2c*+U"E&l2Fs܂?`t qo鎭#qQ̀5emj/++ @X 3N\6*yu@0fX3u1[y5.ĢC'9)8 ȯאKI欹b-pZ96Y S\dн\G-om^| N9܀ x͸<|]v\y8I2)PŢY0Ld]7Xo۹x8|Lc#4ɢ)v$;JX3,vȆ{Y4G0BV #Ü8s9)E-Gwy ODpVgHB,OW+/]{=OcFP :46S {.5]8͓tZqJ {3y6 Zg >qXR;Π v.FKn b0>"5PWPUhToTQFPEP#qJeT9N 8ﰾVt[6HrIpٗi-aC ~:J2b_2r7\Iۅma)y. p\@#NkŘ!!UJr! BwT4trZ2ʖӛZ~" JQ2Rh`ԛ|Pz7ǣJYP-u! X2N:oEf%Jy9LW"Df_=uUjt6n,OTk{tDo31 8+r6Ύm kI!G;~'2Hc)#'ɕЪIA@jAPnFAu ,o<Tq:OҮU@EhC¶jrJ[ݵ|C +\´Y@ 0=USP5=lY 88 K6f[%HSgJXϵ&R)l\k9 ]3[oSFubN3xPE%k-5P*L) -] iN۫4mn҇hs@jInPY˔(Uӹ/$tBS/`k0]bFcQ]Jv{tb5blmT[m E]7T{ǒ 'kwHеŏOۍثG;W` P"%m}NӚMRNPkNBeRAHS# LR.dMv4su] }=Bϔ] \] v[E:=j]cE#H\2#J\B,=~IU Q[mWmnm}] #(l[dLeׄ*iɛ\Q5Iш#! T _O7Ċ sOiHB>%eԪ87έrxVY1*HGDINFV N[ӭ@X0 "(]|"BGDENG & d` p$ ɿfʣ}{񘄑F! #$j+&[Ȓ#&= BV41t&^z'[xq֌!.kh_B}D%s6DlgS{8lK2{XwYmW7a*!嬢-E=x(TZҘdi5 E0LNaX>ݹh('Vl(jmϾ';"42#+ORtǡUƪax>@dCtr*ҩUDnIC큠8O^#`-|-UP!&X2/|٥ҚM-yI9mEL$9Ar1qb/mY՗NZs* x]Ri+9Р0S[ڨV"I{)T8T19i es)p&b%n.)('j/5WRI#K-&{(<ݻ~G&{g(Fy+|74f^pX1 *nja+pVE9rUOǟ?ؙf5ڹ*gP*H:6ZuiJԡgDa~;4Dji0;" eWIjq*%=L'_u">^D|̗|]~h`+efP ӯ6z)ɝd]p鷍DJ4VHE/u1@.Q/p䉻Ъ'I )'f.xuxN2OB"Q1 nB}a]>垠eA򉍁,õto6pkXGRjCǨRRpU; jaooRi:|?_NHNjoy.z1q1A3%JS SۙWRIBDwW Kz7MHS!ax_B_{{E8dTёI*8SUq_ؾ921H+?"6 ijT)KZ%JbUrı*E[hZsQ k`m=8ocM$[/(?GtUkVŞVBUK.鳵 D*X@y)~(uy]U' !0g%ST6=v_f X7m`_O+Tl}>P.OC39aO9g SE(JiaEQźs@_3]!c?aɺLZe>S(/ 7 9QvӏCf0d??8˯,8W-B8RkD\BzOGބq gJ:Siy k0<%S%r>R񣃺/|Z3Ķۊ4}%l}=l@JNvS{R>vMvrv⸈ BB 8 XFV C&VV 06H̘g+F<+, {kB͸FffgxfVz%Rd3Q-^L82C{<ޙL#MYK>,wؠ jT f-|5sq]Ch䯗9d 'y͝q_3k=ȵE.jpke͇[O.c;.<\ iL?6 lR{XBv0w'rڷ{ؒ?1r1,A~o8Ɇv]lw1^ V$q\]<P ;N%fh(FCU*m䷒.ۉ%.my%aM4sхjօ1vVQf84; # 6c%h&u hay|ȲTtg}W^ޗ^3~Q^rbf8l޵q#2=l˼e@O86Y9ٶݢN[I=7h&kMOwWbXM|: RѮJ]ʈVxn{h<hLnTǂmqząVSF¨ʤ[ELQ[djfO8mipYg; <"-tI4#)8r bҮA'-$d"E&9^)i`#kVDAoX)FhtX$T0~8P{0*^Uh"R^am⿂Zw㮽/5;j8$lش&ԚQBi\qbq)gXhLf[t:oL6e—n ] TmzJRrp%j͆I;T`*w*Z 0ꉿ F\ -b ʁ7\ߤC-[{QVϛ~b(BHQ9@sh)Y^4JqԠT}4ϋ >)aSjL^|*seTT<\i"{bqwwӻlI0gOf3wvsOpnY=1gcK}mXJ_譿y~}/F:rEQYe w!1W'L5>:'U.XZ#7F6ލ1N[S BL3bplJ[ޭ 9rmbSU> ݚbb:M݆AJqpޭ9CwkBDTMލS]л5 t>#ƻ AJ[j@ֆַ)(zlU_EDѻs.S%B ti -4yK knװ&V;ǕQjx9uS 8W? < ^ y޻i ?Zop>=zia=vVӧuF8J#XOw~]:BU!lud жYF'm M x(\T![\e}e.U.x#\v|拽{syY=?Df//aRt&\,5=(AMME82*DҜHG2Q҂i[3lDQTIlTD/_@@Wͳa[ *;m(65Y~" 1CnI 1bZe]F3[`˫Wam+㠋C bEQ."0UB qBZ!Oo]Ǩד!v!Si-‡k U|E8Gn0SSL8o-@QFU۶(w.'ujάQL3\+HJ)j gR)2pP ]/CGN M.2:48\Y1RVVkm:q6$#f0 dԂsLyh&=!3eA{;熀@(yGr20޲DH-ꆃs+TCQ: wֻP~ ̞grc>8Rv }F8/jQQ x!i.'raKQyE%z(l\hIx'0a}!3 B+A/xԌjf8ĦBU-ge-URgJdY! &)R"Dj bMV87I8y*FƔs +̺-g1 )$~6Cq}PhYs+E dFBRa +Wp" [dN ZHbӪ D˜iih:&f88'R-ҌK8`"᠕HptΌl 7DSWD}>;u6tEx>95DHY;SOt:z/C`e~LJRm$?>V.x~r[#,&#_?Nbe0?_^^\:4)&-߹N߹Yk K6!)8=~{rueg8PB]d?]7S7̆0o/>,W$1x.,a84N/<*-r3 lC 4%?ԗbs%>ehIҎw {E Zb.۴`#ďg#ǜeIdȬᗺkbUInV#unj>^Np1<\“A W9z'pQ@ͽ/wqPDɥ>¸5Z]NlCweOBUޝp^"x"DdI;(RDeMo0qi-EhWna@N}0V*l7O(3ia9Ӫv. |͑d65a``Ɩ9eN>}4<^;Z%T!;j1Ժ:?;mQφrdF%W![@-kR5O1&BI:#loj HТ>?.#RP<4+T$ayP8Ԛd3LiUpաVmO#]mLKxC;-azQ,9.WrQ_d&nç 5y+_}UQ˳ʞgy~=WxԲ{Ꮗ>Ue]._O2Zvw截>Ґ/ $? ZE`#IGNMhkfSY0W6HAO9g7Dt)!A)HOZ9:WnoNyݸ3z7OIQf,=J0crcBCOp=$7y0w!zFHShP Ƞ\^GPڴ~\$/ :ޟ8%IAi;8oG\H⧴m&wSzJ z ^JT΅8- Y&ܐi(tO/RɫrϏ/zy$KNx?}ԙcϿ:׳Y\'=7۩P+ 85(Td^(U3_0ZgR _1n챕3{]4 :+{Sq9Oy|o/.#&_"_dRNEoCgˋ=3I| 4O̳<̳Ϊl;dw7jP. zI]ZMٌJ 2 ַ9}~*~Kv"'ك=M>Mrq?"&wnqu>q'eQ>s*z<?|>D3 50+ 1dHt|{, Hc'ad^//%T* vM!yyuF$N,M,GT2 ce2[+6%-bY=xUJ}5 1f@ئ.B2) ^XE !o7y|nVUW96cQNK5c-u)3h eJL=c-El>тQSPOC"SDT1c+<-@8QFS/uR*DKZVi>BB2jd` f` fU&Lyr0qfikkC43h.tAkWo6}PEӇ׷`(G<Œ5zofofzs]&1R&(Z"Re9APߙCǘ!yS]q52mƆ P媠eu(0io(+%5U%mX~`F ghK,+R'eLpɭ$BNC# L#<:MCc"hklH\ɸp(+W(Ys}G؞\0z9Tσ*ҧO7_)rjLTxѷ.c1 h#?8E}|X9iވ<Pdj=Ald=Zt[ci=8Q`@0XK# r%IesRK B+g9[k -])E0fvg;e>.~u$_gbūK<~LJ[R, ?>|Z`^ݝ+? ~{u’e7y拕&~DΧ)Mʟ !o˟w.|c~f?ﯯz/.ټ&Qz;2%r18)eƑ\d|5>RIvknw&7s/S"#$93٭߯A2%(Kbu?twR 1ri!!XjTD]u.hn0KХAE!96Nb!A煓yA*DC0wTP& ]Yʠ)<%N8\hƅF{)]"ءQ-R0hP`s =# O(K&iKTJ-T4[^&i1Ur@0wWoD8-Ʋ,y{ՂfJXbd_XGuIc<Ҋ[ƀ}͵ u/EDF hjH&2Dyq Ie ) lq{)) U##GNV4DVg{_ӯS6防8I"t\g$S eBf*sbtZRK ́g+ )=jO2vh.(G2271A?PI9JʋۣѪwP!Pҙy3`ޣ$IƳQw:0|g]Bqy -K)W<pz!;-g+4IFMBw?,̮nW8]0VuσC߷ }M2PN-''#o&ջ5]C|CI*!Nwj\< аPL)AI+w>5 3POO%ƶIn.4cp>3 "vnzQ)O4A 3x]liL@11n#T0Cy RmMuʴ7|Q7À3D։pF'x^*1˛7=6T;؊9Pj' b+&œ*#toPs.,EwF xtuA+{q*ȭ@u C9fK̼3u2pM$Uܿ(jݿ ^rꋂ3GֻќBf5@a 10v5k6:Kj.p~JCɔ !{)0A`e2]i8uD˙,)UCٞ#73 З5r&}[;9`Z9JHu9C1Ѝ:P豢-[u/5t0NwL)j~R#:&_)JO#P?>-= o?SbaTwT* ryon1 RQؾiz\mS#W@MEjmffB5lJF-BwFhơ\"V&7 d(t(@# κ/1l=nj\wZ^`g PqH8+ڛYAL(0+Q8k8r*e+@)i-WH[A H?BB@O]:CO4[%i;(:!-kk}z:ՂFuhN54>ZP&AWN:*R~~[pߩz3% *{l(\18u("z`ՄE_?? ʥޡ34lI7#n햅c)z^~Y~}ʖRT-e]X/눾ςWjwN|-ԽV}-=Ŷ׭_UXSבE4I*mj-6ԌAb#:cngn^xvBB~"$Sޮ+b(yvÕ2pqQ✓Wj\ B"F|%ןiR5,4mx{}xX,(*t?+>Im<$h+-@w?}%-w~:$jR;lGxS98o: T 0_|Wãgsժ|`A~wrFl$~/V]:f+eD[j3G,#SBeUݝenjK~9 9J[,'jQh>12j )LobY3zk1oӸ9?7;c`s9 ^]]h3$'^SHrGw._$g_ E&vn[ݣk-{wmcYh$5,Grӗx5GP9馋[KSb4mmLW 7j3ݵ:G@W sSW9DTJܟ|JQr٣LS؈ ? Zh} *Zf]θ61k  㫜̲€Iƥd9pJ E L8uf©,&eCe6蕨n p'71  Z/˅%n,|9h̓WM#݇l͗lqsl ,*f3ɓ7\uG ak ӹ=7Jek eI (Q3Ђi B_`Q \\l$gg6[Ԫ"t(Vܦf 8{HVBԤi@ڑfM:sP) 1] C8ּqCB}"W.٥Gݽ܂1h.nr׍|`U=S'$$D|wtyIweGP"OiUmg|vm_6"ZwBfoGbQӚPv] f ׷X۠7' p0?`8#J~ugP»ۡQ̛y0ONӧa2dNpSRxm78{C J{ܖwgPu16bQٴ{qVb\nk ,a~@FW OÛJ)RChS(Ww~,^|~~$~. k!֡7Pm-9 -^s f^B~ L{|pnVdL>' ŭ㐼;zTBT zF@tp%nӴ>1 :&W(c({BRϾYRv2=՘s6 *󔓔P0ҋ%qe.=փ+, F, U$3J ,/W>b'*7P ,|_{6Ig&GvW["L%Ae,NJ \P{("'Vd+rCsg3AEԆӺ'n9j4e 9 !G+͕!OC ˌ6idf\4cN) ^jE5/e%yv{\0괥[CW/ sJPαz4?ɰj#SM)\EZ4ձ>}_dJ첱z m{O ]6ROX4PbH!+Pggn9xВCE\,%\9n0ILr2KNV˲ɭNJ9ɗrz_ ǰrhje4MxQkӉƐTn.1>.co_CzU,柯ywd$S e.,FPP :7^RR5vL4' bn. Rjpj}; N;xeY. *Cy07^I]늢fEw2KNXL RB?VU3rƽGHF7Ook^M:K˅u3:NJN ͽG/q?iP(X jN&k/[>@#4K_D CM8f4e.j]ra}s81,i1nm[ "02~2Ѣ Q]+n[[U_愮=rF1r%Ӳ *O?OP7BWrrc_-k Pt=꥖1ohjb򡅹c-= .OX)X/d]ܤ w>a&Qq 37o4:Pv&Eb1Y`J49ZT@Muރ 3+QSfͪ|3tq[DĽ-ݦJ[+s'0U-Ղ'r\*]lx1CZe&C2#1ǽhOW{M|u++ zE$=ʾjDa*{/%ԆmSmK#/sRŒ:x;=_ЕSD?w++\kx=Tn=Ks`Vq˅2#i dfY\E*,b^iGqL&l24f^ K Ĝt!'혡9΂RIyNk,Em!F%"CZ2ʀȼ_]?/NC|Ɇ#~{$.o鱹~w}py{?o >&zKzX>r?I>3ot_a6>aw38j?n/.":HK.,O>0iߜ|v7wՍ>ฦQ,parX/ Ld\wqM Me)Ӆ`_dD53ghUFJ/(ʕD()r (l磣5:>{ y61 /`ɣēqAq 9Oߑl;k9b _QESNh;xȱxP11lۧ-md5uՓ8mBTlhL$7cmD:L;Eg|q-Mwdmd 7.RDMQ*[.s2E,J ,r5) -%:@7…;2,5kE_MLC~YjdCR5bPy^T}>P쏖92 `!_nS( L]4Z(k>0/@2GiGޯeh̃~I&:ʹ9QlLTGcyq3uNir2:/8L%L%Di<<^)9its% &\8*,pzІ#oUP __E Kdjj/ +T_>ʲ~T߯qumE'㲔O\sst=;$9~!GKg_qY ҅vޙgfBegػ-Wͻ7M}oKּ͇M΍tch]Y踳I+Y9?gxȿ(Άf1YF^6 ٗ&9@ke!`,y͹NF);gYPhh40G0F=9gVel)z1(f-\yOv׋1oC S߰ا迗h mX6QX6 /䗴V(lŠ"}6MA:\-?vfjehMu#40֊A鴾w;v&޼[sGk6U.+@;8^F(q8mQ1"7q-E U@dM.h׆4(5=SKq\1_$3ꯘlWe\wż⸦QTL݇eZT %NXZVlud5yH`8q<+t$AMwS9Izp-OϧすJ"i܀l &&iuJHJ c5y dhnYg9 {Y0Dn(<zrJY0y`. gaH_B3A̓6 ;2"B UO`\ʜ蘍'(Q\7"(L )A[x c = |5nyVCA>6^Ac +W?uOr ږCi랫iaCxICcnS?Ari%A`:"IQxV@?]x+|"2gx"g sT?$X5*#:G#pl>]a)SKZSDVRfhU+NxZ\jK,}h$ D&ڣ(h‚ShbaԒIAE)  @"92mΊr8p)#@Sew ]@Mlu;X 32䒵gl1{iBxwAc*Ek.7 M=u;dhom]N?'Q Cr0mńZ- +=D&m5gWlY,C^c?_g-s(gsm݋0$T-Ad6l#$m{6ٗw>`_|in^4XՄSsmKq\3pZl{snňxI5ʙSi\{m-n٢Sf1_=k'Y2}!UO;7?[vSFכx=;{|^,?*~-WS:zwd˶'EGuWkE/[7_lvE R7ֳ1%X9ܔD7nA6Ň s@ »bc:ϨzXC:m{rOB޸o֞x7w tBQǻ:Йw n1,䍛hM9KbC-8n[(>:ޭBλjn1,䍛hcDjZ>.*8}re'"t.PNtC+.\2tBuС[em!&oi}dMv+[}{>\/ktUA /85=x;vRi}e5mֳ :9ZȎkE!4^#Xb1TDj T2lGh|/Dv PSUp&{WY pև^"KC})c1H'W@ÐIhbyD~!4 "~$ȗo}2H0yCH]ۡ`[굹7'Q h+#>W%8H@t\@c^`(!u3BC5ք|J*B01&L/\xl"?"@w&'|8E=X+~:vH]ה*?xpAm\.[s c2"-yk>pW; p贵|-Y6jۮC/׿Dss5^8Vz [|-lH !Y-dR&;R>.9Pq]qlJ(zu::"SշJB!!ܰ@1B@tXd4As3hjN+Ŕ1[)/ӆzx؊OLȌK <\X#[%ͅU A&T((Iqh--5*JmB[݇ LbpZŕȪ9~+jOY%l0FL^]LBeTXR\$9L%M@hf Ԭ'aWrjԌ|MA"SjU?Y$ #<5DsKdx9J4 5MF*ԔpWd ѹY*iJIF 0!9KJ\r0K$y)F0KR( hiTI$KčntIFP0d;@F'JD|8Rcy3jq8Cnۗ?qmnSRhxٲ޻O45=޻==yO[׃D͞x\]3@lk z! gsW݊ssIocˤuo+΍Zq(;tDp@4 NmP!l.F<]*-2ZX5뀣HDCm@꯳E@&kvn _K(@9-H.Ap0E`sa~?l_P2*;IUwD j=E{lk]/xVAb~zNgY[l w3+菒2]}˄J2d{FqtÎIXZ#?I,X[Ε$>N׋L\y KJx_iJh*e"#&%/SI2B 6?]+pWu00Q0RfY qM3y3bDYih@?m"m#6!HGֶxw1|ZBߥ6^=ɟ7Q{{H؇_~\\+o3V^&f3u3߾AWXv=N%olc !Ood oaVfueNbP[ L#LytM:YAv#U9 UBo%XкW3ʍfxsmZ3T?JJ-7WT#J@145U7ŤԠ 02ER*:]5Z,[zS)05R*'%IsJEB$ bJ18eb$(x##$N)%-Sԑ-/E]' oI%b8vW473bL_[l,<$O_ꄨ r㤝WgPNoz ;g KXFc.z edU!;ln_6+WbUO/O~%ޏ+{lӕ=peϧ1cz-_Ts׭(mYڪ:zhu]ݴU:X4xU={ ."/NV ˎ3s5ޠb0 ~26cښP{l2rxj6R):qvTS\q*ːL>~ܯ1TfnI@SҔ\%Bds]2H!Y)Z iDB ,beE:>F _t%Y*eAJӢ2o"א&R$k)ReQ2E2'ys𚐅^j_K,@z ֶ˾džz+9$$+&٫/秼 ߏY[ٖ{~qkuw_~1{b߯='!0+ |5+g:upgXkʯi6_gP FQJVOdBTYNRr E- mjieH)JEV$e rȀy՞"u?ڴy T2k@S}d?k"ـ&>(C4';ԭ=(zgFZGBmH5:2lptv' GXJ0(\3 N&-ĴԴ3 2d5C9BTOf q҆%Qd%27"\kvmi#$IH.YFM!i !l95 Se )ɨkVAh7/d{Ӧm>Y5 'UOZV@]?+?hU{-yPW|8ɇ׊ql"Hg8{>u[St{bY%x iy~:2\vr 5y|C\|1Xs!C+ 9bqosqs'orƯ-`<`'^Z (nLv7ϚMvLVӶNvVQ/\M/Ƕkz;:Psd烄*-A@>àWU *恨4:qI1e%COJKU A&T(@C,rq,3|M<{aI;l,$DR{R5bsbjSgbqUMep,v9eOwNIA~\%˪x8MijqƐx؜c̯FwWB݃mP9^B]NCľrV>H1C/ X9sFIσ%޻VƈⳫ'.EjZ`wȮy@3 3emٽ)7c;=y1y?|ǓXf}6' 9o\|Ymgl$/<0?*~-WYT:UTh']nWP7NVϋGtJkE/˟k/6+4٘?ܫrCB޸ٔs@m 'MSƆB11gn-~DwT=[ y&dS{7%[(>:ޭWP <[ y&bSOx70>wq E3^bX7 ʸF3 R0b:e1c>s$(ll]&J-s1:fW8ÜZ1O#9 1O99&OȘ'yBԑ <t#aD 0&U],v50/[hQо0\xc!\zUٍCN #9 oS،'PWi'C(纭R:kYķ&tVd*>BKgȹ PiǢK!KKRw'r5cJfU*$aTϊVpI@è+MyޯRB|^Ybи*8<6m%@c!@ 쫬B[ui=$_b9c-A֗M*{ d/NV` Pn_. ,MꈽGyߡ6),'՟ ?Kԍ'Q7Nn|Z=ZfCP1c5Wy2AQ"eqa̵0 AJbsb|q^`f\-z b'z3XE#V?9Ul^[B ,Ul!'8x>-^=\wo/xVS_9 Q~"bXzd)(Ғ.K?ה5RPE|O%,9$p2)0(%@a$&V2 lhIG`ۋ6x }{43j<^>31NeSUR-/@F^yӌҟzA"Őv{n!ϓdi` a_2Đ(mȰTg!,ްAC ,}TPtzeKGl%ɰ<~cv~ bCΨ\ Xf _cRfWO.ZNwh%*gg>h &iW^>nG*x1S';2=3uj4!*:T(e=pod?ubrMs /nmZ;5m2 쾲rn} &.'Z޸#.i۱IS@A< ZIMH'tpe *c߫MF9Z5gO7^P.EZ5v?f-^5ڲC^-5^<|'/6_|kzlt;nMwvcKƽPuLI ?^ٱXq_=wuK{1f%Nh0'/WM20_%N4VϑH/KCAˡ+mW@1tr$S bx g߲:M+L'jI)qͥ45D0/u:5.(  fHrE|P8fF+*^4\sJÚx-Os\KBlk +78_K%jnX7:{ڢܯ $o?Ǹ[SU&'˛Wsxog1&lRI|B@Բ$xy|.uC%b9kqEE',(T lHgoUp@-SU`KL\]gBP|)ԅ8B^2I.+ QjjB\ E3 q)SPL5hK A+ IAgsɄIMa0/r-8@ePeM+~㰲#J45c &09S(OFv +kfH7 0bf>@FCZCb(0m *<>)qEg7^roVn\$~3Z\V&p%>jWhF幵.vȑ5TЇϟ_xfGG~8Wv @l? }wamlħؒ~ ?,ZyXdb;h[q! ۿALѥD@a^,ӏ}Ʊ\p%8~`"KJWBX<aDŪkZRU-P}QyLT:Gn t)%Ofj=DKYH0#wA8C4rЧR03gTsn-29^?Ǭ νV0[6p-Dr u5;ͨdoIG !7, ِk\kgoQ nQ04Ԙ\=p$ Z+FRĠ0+v,B@ʄ}Z,-H@}_k7e着G,Qѕ2 r qrZL$<[c2 U,ZhX>HAmHޣ2ˈV[t f.0{ TM*Ѡ~7?F+ZG @ mbrmy逶b~.t!S>ϫEul7LqK$ ZCڠ&C 䂘,JX&GQ-l0cIVw ?&hF+ẐK3>c5(]Ko|7gap7c!(nk1*Ù7Ft^ÍH+2=I)ӓ4ie[BS +Ln)7d_^u;:C nȾ 8ʠ:x&\nR!(ǀV܃sG&Zr`gmqi#+=o%aX7o-!WNa |q)LbY+Q nC- "1_Hm:t[ D֨U>-ۺ6-#b.}Q]1VGx^ɾǭsl5oL"P3&Օ:6yª3oƼY̛u<I}voon4;q|BouIfY`ު*iCҐ$ >Įס'|B8hzaj&z9ֹsݝ)Dzݝ(Rh$;nNu[w--j腆j&$E` M~A&ETPq|o͒I$.Z w_G+2^ @Qx!mCߌ&C4xY=ij,G/IU*IWq*!ZIؕ^S`*"S04D | ^ V [eBl quM>"FRb7M]>La(2I53Ps -ecAƒق`ϐ]M.}~$C{4I;ZNw+m )SUX2VDA^BZB,Uz:^[‹\84sfF=eʾ"l/p;68k9ptͨNlJ>:˞'ŤB吐^qhRͫsm0e's -<3qYeg-fo57`ϑ}^>&SO3h fP%$d9]B+lc#c0} H& Jh_$;9LH|6GwsI^$uu)/4i} WO_BS@X`7|m[^QPbꠢd2K'3^:/!L˖tWւ@bq<Ѡ;SP1c5Wy2A@ .rQF\ c_AJbsbJð#&G/Zĝ5NøY~0@IsE !éFk`}*$?l$ޮӍ(@\ݏ57k]="`YafuDܛe}J}|mFE6.߫] :DqErmKB'vᔄcN&ybs(`O^:gR+ 6F~ +|{{\F.vK,>70%%xeb=;zYVV؃NӬ:#w5.b; WOzblյ f\U_kYFA+uW?ZFP&i/f"(U?5wռa.hT &%jv nMHȉ29һ-LGw'bӇOe{zA&tH2%*+0I ?HLt9Ġ=cRB[oyVi $%4!}H}Kp/,9T jI}^?,Hը-☰6?4= şm g52CP1G!Dd_*9Yj.$lNO%'k;X7%B-mtɛ^P2x6&<=q6P@~"3C??9zU0mn'aTjd4\ >cxƕéS`ðJ9 'dK jZN*SR*`8Y #T,+l3l0T(3V"TH %C .W!ٻHn pdX`^>$ }!1>wKZE_|3QK3#==+ƮV,X,롃QERG T0PX#>:3[A H/jr~N/g.xR;,ע{v4؃,P5[ <i4 \dቊ4zǩc=dLֈ,wLo% ђ8cZe)mS_TWhXϐ_g@LV} U21.__fVX}ԆE(o=L[-B3IHG"x7Ôf1akͷ]1_(ơ>6n˧Ggv2&قSZs>C>jB!&N-:dZ-:ʹSU)ֲaG;L$N[i+{WQϚ:t9@ ؟8N\zFKRau9MGƅVq"`= ];;if?$.py]r-rֆ4y\t+c1j Q3C5 1J'ܜL `DA_IuGYS  *Җ+,og珊S2g 4.$+pI3y2?f$PrE_>=YƋNN9%"mr^b5(,GXcAG˩|{ 5r% ؅m6ǀGJm\SyZs¡2,i!N:B$-X-{YF7;z{ۆòx$wyHd 3'n?!LI_DKGh$*0LyqEdu0( 4J+כh*ȮR6O*!HQԮe #5lΑ>:W%(l ߶[6f=#XPɪDЊ̤@s#)`,j'Å8˨ :٢ZxT֯RIӲmiT@@qtcI+!au!8Y{Qbpku&,u¸)/هW/u Ҽ)z|RK f=ް_ףcLr*b \d|{K^w=pPRB%1=ѧT @ALfrȩu(EHo{2zug08o{(f x-oLEC>|TGg8cv2((EutۺDuȡ:ڰP4 >Nݢ:FdZQ-(ŤidږGGFg@ͧ§Nn<8:6(ysz՛4}b14 pw]b}ևI*X=B K:qkE:q}:r9y3(ԪSk-TTEDɴ6 +)E&RPKXe%zaX†R.p'>&gg~+$}APՓQq/ԏYEv=-%R-7U F;\PQ`u.&LiHAPPK`7kRk^@5U/*A-YՋ3ՒdD,;M~u$]QP)EHm酹4}Irm:]< 32^dܗ.^"rLϾ!曛OIMMf0aFQb we+ i] \Nj#/I>/׿5>f,&<4dʶ4T*L \P0I͸x D2jn֛4ݹc_Džt"͗1ڎNt qKUQt㭤rs{w:pI+ُ9+ iҚ.A7ntQvŵc騽sLqpvgvEJSulh7lGVAԎFx8ڭ y"Z%SJOTF}ma%KmOx[ J3Iw<<< 8UyF\6rÙ}!3sKľck3H';qokZb"ӄxwl/B4ibRot!˓q8c2A' JկMh/1{7l'hAfz7~1ɰ+ov=Ȝ4 O箶[\dM5CmjޛO RJ %ȱgHjKoy}=6HKWs~xL(Z6h!B̑3w;te.gH^?d.5utvvY gg=\$,9ݾ-~J?dsFt~<:>IէpuuFж`,xcX7VK]PSJ^71G?+nnZF945xm%˳gD?,IL}^߇W;2[''yLMs@ĴKd}okf[48m7/sCR ̝KDcPkfD6x!JC1o@iz#"z-wh$/Wݨl )V1383_=xgR+uh3Ϙ$!K1 f,DiS,0-ESǜi56h)AP"/V6XcIo`u< ,4Ȑ}sL^)wW/W?gC,7Ihxen??:\ٓ'YA;Y}y^T .(Z[L./V{Ӳ!BxQ;@& )Jq^q+,%D 6@tЀya9k 6 q:A+L? _]}Ns:д^3gԵo1قzuϏ,oƸJ&͛?% #~(m7+ؿ[z>3B-o~eU;_Vpwyn-5QJ@?Q8W!dGK◮qҴɥ奟SV(*HmȚ> -!-!ܝ.CC]ſ}ywrvt:&ɦxēEs"~vf;Lcű2,7i*lA߸8E㕯j03Y"iW"UO.MKp9=vP3"P=M[`2|Es)cmPw_VOt\c C5,KanEX {%(]Ƌx-詏WjJ'M[˧MXwViqkcnk-X(iX]Ϗ]QNΜpV #1:GKO %1Oi\BKh%Z 7WG^eGcWA#hjH#&hMYTH0';0b7 C~ aB;jj:B -*v\2zQ"d3(,"c -.ae3y ՂٝjD:Uuk7r{3(8 GQaЩeAQd :dc{$ix!dJbwq>]DE:J"-"fҊX8R)$‍0#(lyȉJn ֗  xͦƎe[2vZ)ssw^[DfMިAltcZ Jj ֎ѦD Z+6^9za9>$䕋hLu=2=n9ы)hZ JDv6mIi#nhVP!!\D傋>{Qv3)/MQ^OA@#FOb-(9 kxy{M/3΂TI&U D3e!}\Tl~k:i;؊dNp׆X~s}/~-4=+?9'\̭vyTXQ0k zw->_c5ʖb5b+"n<2q~4mM@[uNlSn S쪈8*P;@U-ujVs>*N!d ZI4AX-1^} {-^kuZ5ȩ7oKK_+>+-Y)P~ʩMg+(OaŲ+ +aWa "v3e)Uڽ_Wy]uuis]5+=1-SZ*(๱ỤܢVI[XX\ivJw\k+XTVµ. (b 6kw&Z ӫTR#S@ z%Z!ՔB"Bh/^O9Z1& (pWb;%Oʃ5iX)zҒJ,(q/mOK㱮K-(z(4&)ԂB3fa[(BBA3X4'̀y Jx9qp`iG6΀ vh MP j:֎Ga* \( =Un\)ԲiQ?ݨvj#A:="^yS UpqrXpO9whUtܦW9arGB},1@-2TjKr %P$햢z8Zh (UkD=XQmh6KHetT듙 h&K~66;=O=s4<& A uZFTK7W9!H9M1r8u-sɻwGB>#mVd ;ecںi\ĆcCC[dVO03#~ N-+d)6@؛FyWO&keP=YTx.e R01E:[7&Ġ{={4iS~A;- -Z9`P3\h*\ybUB{ːEʼnAU%yqWFjTXK"kY$ NJ cI*4Q\E, i85A KJR3jch1x@# wDT$J$;C)hJ'J' ȩR](h奈BT I$Bv;251ZwD]~o۩Ͷv dS⠌mNzL4S䶠{q ^/׎ipK|L}?;~L$@^%;>Oe˦DRvR93 Qbw/,cAl+6K'M2xh .PڒȄb[n,\Pɻel7&ཤ'L|Mb6D+˸P;iX}vMS~fU`iI뎂4ղyeVhFheyW0ԓf2魙 ?Kbb\k <Un&81:{ p?ST%puqG.;gB#{Y1O0 AIFE1%cr~Jr!̸w&m@"yR9Nz@t®Y e#bE%>E$ua?ZpR͋fǃ&wwN]3L r>Xcpt]rDg/4I w*0 ')thy? Jmp^opO5(?hEoYU]!+dkc"DC0qEVDhL&c Y.1ܝz9 m֮lyzqFt}T"} O) p_#2oߪT6.Mg۩/f?Zl k(,Գ=/ǵ.RgӭMbFYןb VF[DDV]A;\zh:TT==mX&#&ޜem^~H$1 ##HBD`I8.NPdyr@2IƵ$O=.]WbW_lp0 nK2Zb6UB2ok$5o*%D_+sEhJjOa(͋6kl JZQS)&6eA Y WS%i'OVYkk*tZ+D`&&4$3E:H'4R^y.-Ȣ2c)6&tZ?^0+MJo4q!^،V\LW(5ARf.7N3mӉ^JRXۤsЂ%WwDQ:~ZWQ. bDQONqX+IMafs-'+j>Leɮ7WL1]~~KS!|됙lH `jPok"$"(Tr[{*P&1)Lr[? Ӱ1A,OlrC69vBJ;5f8[dJ+i`3VֱF}OָF^T~pTI%3[Z`Ε^^Bwc}37~K BZA? 4d!'KmZ׶ )e ̚r&(4KefSgaSDh 8Pda/d/i/9LƇvӑ7>2  $PR+2 ĥR{?:;.P7$~p[pO@*J ސԭP8έeT }5Ɏ"BjL?g:& Ҙ(v}zAP)g|̀k=O&IvWO)ycu,vY%05ZU|s{9GP,,4D!UD hZn/ a.Ï\b:#։,4Xa1!0f*%0CkTa:S>5ƚ>a4 6rs)}$q/(nNTl50/qﻦg&^nܝL7r ܰ)jU!/8þ {uҎ/=E`u8,r1{.Q qB#Hc[nvی ./NNl,V%Vq82hŖ m 6<ѱڂmCOF5(cH e$icJIJ(Tu\@3yʦ^yPԕU$ Ћ4vbJ_ P,vLheL>ӵ) !Qօ@j3i[eV}!{ i gz*AX~0+վc>[wy(;\hEE)*rEpEQ$GokLPT1̾5_Y˚8 .HӮ oFvf"Wq{ BQB8Fe{YdBJih&y3USɆ >l1 ͰV"X'/t$Ii[ aD8Q$Mu2B(Vę lDf!̇VTiZյB[jdMboB?5 ㆹ =QI_T~1hVC`~ =p/kida=_ow? k|]`*AwA_L\ake]{H$e]p|kkJDa1";INiKe.}!(ĹqcC f t!B-"GVjHj2d"eY$Z[(51(I V81DYHh:O7+ `[ޯlkܲl4+*FMsZ D (d&b $2tF[ 7jC :1'\RŒ#``th#vnh&+fW4" }2B#P1X /c*dȉ Qgk8@X^22j>vN0!_lcQBq"@t. b+Y/AA 4gG ۠]r*f"kT4#1G8.ފZ)Z`LTf5 +/SNMIC,qKbBa_]/%($,`E\\B: 'ei2 Efsg*8D2"N*tkoT cL\{s;W9t05vIsU46~dߦIlhz#<`#wp GԌ:DZjAHhEPi,MJPұFhAC;3ۋRpFAC:*V/9ߣ>?ޜ}ޚqN0?$׽>ux s(x + Mdwش?㎃ 8M.AG/{jF9L }Apx,q@)M&oA [ޗxpz5(\=@eѓtcpv]9펻 ºPQIGA :K@ :t:,HN`XB0gOPo K`X^M^= ޜ'G/[;A4Ma}8@3ݽ wҦ<C~`}僣'`38dA.&$|7v{"SoilTӏMF8Zտ1v=S ^<|O_?'xݢXӳ1Wd'໫_oj旪Dkuū$pq30hɦ |PD[*\S˷n<Z\9})93۹Vȟ A3O7˜S4t%sU^#PPhD_恞d3? +ך~a?/VMMKX]%z 79ewxe:9;7]Vpewj+wfοLg'~?y4ޮ~J]ǞdvvNħbbMBӆOnS:&>8FuΗl |tr(LgOtogWGi7]vfn1iף|֞7\q&Ѩ<0_(<^Χ xƾu.Gñ<6RytoL_,MH0a(V F{ ǽ QA5 Kc7ue{|=rS46t4Kȟd/IxFן$s5l\4 ;R])硽5B?빋|>>y&!8{p9_MP.czTa'4[q݃HkCt>'?9yD/N~*#9g7Tಽ+XكV9IyNSL+yoCsό\(rqʀ M;zfc2:Ȗ_r[IBN#΋8 痫=ܔb%2i\:v:-$%[^RC~\]a^I=XN4b'Qs_ #Q-\6SLru:se<7qW=!ńѲ>eInE+&8ib}ߧ @ѱTlA76֠Bv{tm_\]&}z")tYm2MLʸ| AqC!c)K'ZV! 4"noV FrZ1e+L9d&(1)\-GYHuf"ZbmC\??xjmW\q Mgo>P +As+g/DA>*(4d'A Ydr=#NֳDz`*YGVr*V䢞$$ebJ̈́vǐz/J!*YZc%8g\%T D0$ao .~3hI՞v4s0]68*K\vl]m*Z,j9>͎WnI'Օ ۤ$߲4A[}9vVI_ILRlX746&ҩ~6 A3 gl͗'[KFc-d駣b鬭K e.r!R#{K5g*sCjg5FjE6]cP ܍Z{iuQvֺTf ܻɷ(zC sd߀3 1k5̚?fYǬc1k5̚?fYǬc1k5̚?f!kE5$ys}U'BymrR" *ɳBOW7ad6PmL x D7i+Dcaζ*Ʉ5`5jC]QYڪ!B`5&"00gFLLQ{yH/U̱X-Ax!ۥtBYڇhKkIP9gu6I8ȴ@hJDXivH_Sy•$ ˒E)Qez|R)LŐ sXÚj~P>]ټP3X4"%_"8BtFq* rdRU|!;ѺV+ YM"E%0x ^22qSs.(k=@|]<\%Rbʢk<*L E Hā:2EL[tW q^r=VCb,{ e@b#(W1CLDL&YI} 6TkU3a (Z3R*yֶK3I]/r;06h! 9( 1u6,!'p+MmМPёl:Gsg ^eNn; 1W0!^/Y$vc\FAZn9:rNH6M*I+``[b!aBQ3NȼiRlI0yBrLӋq6 YDIDYEy4Ylqɤb#h, C$Ra舸Nj>!f&RQf4dpc@z6wx rUO;yX=1<"4>jC9)H`ʑY Hhx"j  I  'RE,w^dLv=w4#yii ZIªoUkVC? {20-dO꺔+-P,ؗL?XݳguOP`TaJTaF'%Zɘ4‘/kxvw0QYLVY1 yXIb0l89UQdȂiO^G&4R #5Y$$>,H&Fګ) 1C"Հ#Pj[-kg CA02, OieD%]kea g/9S OpߣClne۴(ė-eE*htw{}9n4X $GU ,`Wz*ZK D.)5Lja$ũϏD}-R+<tr&ɢlUm'RtF&Ȇ+yPȃ^ ,#`si,RkFf/mQg;g%d̸%1g[/ >3ݹ~^]_Gwu`~_S^Q+u`<鯍/`&}/ uFޜ90 2gGvv8·ܵ@ rщb1ŀ2k8@ jd9Rq|ez21R)ҵQx'ZduEuq}&(J=&2;H"PhMg9r 5=5F}BљJOoFyԷ@Wy]~yY 7˃~4n[y,d`+ rA)ڎ3x38Wx5;NV8[=g21ǩj[=O\qܠ*4 _{ %R1]?V*sˢ٩27c9Or5n PrWP\積[/k#vjk_U8+߿ca,X!}~jV [jeuhƋ%HvӻG%ܵ{5ZUtf^]^5 Sk~RǢ*`ug9B!3Q4\ENK-Ɛn|\'JB#vL"01st@Yd(BSt Z`9C gθ;K("z ψINbREtВ)Q ay.nYcnh?~uDy!#kxq*y"KG͇@x*),ӋT}@EDkc0e#KNz #Jp%de~3`sHM[5E3wоO f'L5Rѣ aGn c|^2LcˎYo}:(JaJ֮D!xCqT-Ux1J&GkE е0Q$VbS|*dLM 5ϕ/c ꫏4[cy1Fj(M 5O+{;6=Dڠ0怜uc3MƾȜh̉ɇZ-⻉~ +wVO٥{E[hV9Gx.=3r^oRX|xr'od*V ـqZ廻'q/>˗Ne.^<{]h6hܧ U6^M=i!jg* zGg,gA4>En/zuJ.=o 8ogyxzeݼmu^gݔfg=.ޚ,,o&lדWgx^e9GEoƀuqg-*zwwtS?TɣF:pɏD.GN={qrD8}^}UJ@^4䧃GiEWY_Ӡt\-^ Ӏ?/?y˕SOOIy&Il;MyzH<;| :=xN?~_i'>N|ɇhq[sF sv'vZ-n9c :8_aGg';m`f"zʟUq.k-8=չFOri5\Τ@=+7^7UEveEO=wH163{4z距xꝆvihwڝv-NC L.WlPuusڝNST;UN[*v͜ei\]-*9i'U&sB*Œ`rJ%.oVV1 j;/] B]>|Wԕ=̇p4=yss<[!vN k gf=7 5efIq9 s\D7c9ҿm" r͡t^f&+3G932 f6)RhjᦓwiMnŨ|O .8yo\+a}aҭ}]>"mޟF:n4#ɘ޿(v֍O~W/c:Qkl'MK5UZuբ;wN]zѝTs^De\5"%VLt R>gފD0h3 CV?@V9T :ќ/͕քIMsCJha8 ΁tWa5gjmH.Ro8jyF7=z0Dvlw淑Nʽi\m7סUL }g;KRp#AO; {uܰb>y=fm CeHPr>'->jE"P@ EFx"%6 h!lKe.ۛbΟ{OP{5=?}z .,Zߴvp/vemDn $ZUb+A9m R$g/W+>Y kkAw&\բ J#ۗzb[JZr=2C]iDeDް!wJ&T\\&Wa#֠jsːciB< %7,I.0%|ruCq.,d ZD"D癃2ERGĤj$v9 <86l%d͎ h",#QQ1AMIcЮD $'E  y3ګ ,P8[!c?\iEE6!+j8Ik?NK9<37 m[Xc/7|8ކ4NvBܺ6vI2`J`NjB*)>7-Rs2pt|6$t%+ŭ$);8 B({(q ն;,I"WNkm-u7NE )+InLp:iA3]b9FNf`!Rͪ8[d&LNV.]&=c.2K(.oAMѦl4X& Z.NkZ\ɹюvo4dhm7r~ܫ6`E<7Dm5Z͵خM⇷1YpBV 靉ӷ%$^ -פ₆:8YmAhCDkD>'lREuU(m#!GF{ j\Ai8YZGN%=eKH]:Jr{pTåT@B--Hүu$_bרpm/XhҼoYlrA`I4U04%f]q% .4Wh!*[9Cb: 2kIKiqPK,eR0`X@Z>V]]bQF<d<$o6P 4e-R$ݱمxsd=:+@_.ċ_6ZŅrD+D'ƊD)(G;a0Vb$MOv&1p+Z7P:s,.>څ5ѽZcz؂=k$SwHΜZ YRIв}Yk)B K6ErE$5Q-DXQvw ǥV"Xv||vMk8b,x}zFtۨ_u^. CtTCu+LĒY$Z 6hl{*oreǯ:-su8\ `&V iI 7jMx6Yu1_csO蕯A@) T&- 4`<)jK/ lت "f54?%V>=ФY5^4wF\cjtť6ΓiNDj˝DOߞ.>ΙmĻ3D#21D5:rWlY0O;ۑXuLRCN5ñTGwød`==D0zPxtvR}w]4Ǒ?>8;;={8yI>_ϊY9^vo!{.ne۹i2"{Hݾ DzAИKHlQ OL#ztBP[ӱ;^G+)#QEđL W+(Ms qPB2&csmIЀ1yp|&^k $G>\'C#(m7ލR3ݩ!a?^:*JU7F|<3)FqfuҒFa_U(2O&e M#Eq-pl!ŨeTHEy ({||m֐8~jE3FʆoZ؃oKsAXz'QnW&c%)D%*aE"oRaC n$dEv }~g;c]t;_.+]LD_x f==[C`]( "WIRŷ7z|\V#/`b)QHlpn\`{mK/[{pm:/Qv;q0?U՝TB^ZS:WwgCru$n]=K<@?6xf79'@H(a0D #EC#ıvx6댈 aOCZq}t_ڪI_Puo^`V TbFI3럅ylʃoRq\ȜC9IQ"ɭ.F*iN94!FI"D'Hx`C`4h+3ăB&`$0z_ P13XW0 vCkX!QP$/PvGÑZKVx pO 5y8l- c,גڀVVr<Ѐ$ ( ɽpJ@po{H 4 !0u4\8%ӛ;{0""GӮ6h~_ qǍL!Ӕra1Po-Q8kR6p@yҺ@ .wQ 95Uc +>zJp1h>ild} Pr%~`z= 0tRGc)%DsN^)|oYDB#!QeHGhkkj`8EK! S7h`b ld2 !oC<zb#H!N6P~ztQ du0&$PP uFIJ')IN\!zÂhdqS;dA-KFI Ct @L:d``"Oy ,dM/ *pP "b`6~T4 {"XÐ?,bt` Nȴ"@)b`h^E tLs[;[U5/@KU-*ՠUg@}[ ZqA)\ ZR<*hUAKՠUU-Ul-ߪlWLߪxP ZqqՠESz\Vuӈ Q W\ Ӟ D3L#?c|;U qn""׍s/sYf8-ذtzz*w_zgp`Z" (Ꮥ),"m lN"X;Ty{/:3l?e~X^&?ӿ=>yydef>+:Qg痽_Z- gOw32n:Kh Lks5jZ^d237~}Ni[:_/Q/ϲ,d Y..&ҌُY`?)#oNN_|wp|CoOɔ$¹<:0m,t&UڝiL5/ZMg%].k+ɰMƋ'w^62;;p6>~H %g#>.>9|; )%Fo͘OdW0?aG7oO?:>9Ξ$^/OĿ?mm(ͅeG/ߜ::|}79.wO /J>x:{Ggod\Nփ lD Q~ ONP?)#.7d \tIMS*Q s# Ȟ>TԷ7 ɮj/;+jP"0ï`NgٻhKhcYsdY;9)n.C聁yVlӑyIKK|vH>Zu VBtǽ~tx^-ziW?9OB{E~eU3P\3/_.GoOtW\#die9i[F{ "{{0H[Uzh)ѧA[]n7^7{ I|eiFWi2ܩhO ?9cxƿm.A+Z- \+^sRƴ es!n[wFEJwSh 1 30DT㛳^7cpU>lbY]zYZ:`\7j腴4+/٧_\00kN?]x%zˏwf+ 9yzib/>'x=.{p^|u= .G+vx`MOV Dhb_\t;w{}،h.?M^>k N > 9*w"ne4Ӹk{~_z1mخX00E]DZtu:f^,)f)O1KyYʧc(:ϫ%Qe8Fi%IID>9,O>7?Vчj6Ұڹf;a76o:{,LPP2wvI6{Y<8(hrŦ~hzs[CcH>|XDR |h,uvWP8uDC޳uߙa.?yr)WRS AE 5 A|@7$:|R(Jr$H-,rʱX /Et:nodU+M=E"F8% `51 9 $XylQ0dlrL6&P*8).ET6LR2J­mA ,þ,S,TlU@EUj2% Ze /`luXΕ$7)לԁae|K."|Ğ*N&HXHhtB"aYP/,a?#N.:t{OBЀU:MO !x$YTM3vͪH5g.|d]C Qu.pncP_2ؒWV3-Bz=Ld1Gm 1v饡jPoyiP;P饡j ᥡ*QHnwmX**'VWERuh# ڄ[lշcU)ߘٝNjNH(oCUuPr򘥮U=C叕PxǖclǮ]<㫋G'eU$<\oHy<"y }n9 /`w[3:4kG2SmFfY*:ĴU|F]Gˉ}"DaRQ*':޲Jny-rJ}hXv4e)R xU hVLꙗJyK8_EEǽo*U&tXT'wjyuQqUX/QU.Ҽ!]k:TʹnjZoutA9!IHQ|h,ZO>nF}oSU )n"e 5gϛAж[SA[p*z#l0`/LN[O(8@0r.0SfϞ7ʫΧߛ8F QX+k@U31&a gzBϕŧ\+-,1F،zm(@.l()_(ʛOSxuAr+Li6~`h:y;'xsX T@bbpp"FF8r8"좍)pŨe.fͩ"!2.i`g@KgK}˭gT2C7?~pIoy9F;+%==O٧KGLpun$=tiQUI) Bt+зĔ ɷhd?:2YƷSEBR~ n(B3] V8_*jF>u?4V_F)n h'-/9H|;H~d+`ݛ?hH+5HW;/^ ܷ?\OƸׇ VvEz ./ACnn@~`W14ȧ bՏmo]kX.Y"P*/{> Oˋð P̥ v;#DM5ٺ#:iQќ*}OVjdh4}>)Oe4z4>6I3,I_Gva3߶m׻m pDSVYQ]Z`DvtΊ:S!S mL:g<7)cΗ5Y0TYGy(9j/\FK휕bΕmd[sV; Gk[t1wշ4-wϹM6>Zy8éV7s.n-]<6u w /Q&`{:cxLňSBl:+PBB+#L]wm"~3[MAy1^ҩơ;%u9^vfdC_ax¥ɫ%y[2z]r,͚v}d(\ŋcLSx)MȒNŻd㙝Y~'%y`|k.^B$VJCQN)=nr7CnUa TcviO[兮rS=B[kUuRAS#f*ʄZ)K%m?=rF&q6%~hU,ZYJrإ`p )Qh8~ttk[|kE(q&:KPdgD+LbA6"b+̚J`}q]ާe$c>U293U*c_4%W+vV;7-5xIBf H>+ |1Hu0 @{XłE$Zi>K0V[#  2iM}N)be=+ N9d*/P_HJaBueVG 8meA*{r2)8:QɅuaKAq"2rkص/.1E,R$55G55_\rtTDu"#MԘZJ[; jH8RITYY3)YOѧt )Y(BR(T^iJ#{Hk5ŚWVGrOibZ^pPw=8Q#]/1Ü{~w!/\"k.+ ;nDV89k=P`I"7gXxBJ*6 %ׇ  Ϣ$R.PxETʨنk cSrO pv0G7D͔\sM74p[`r@wW?$ZA#bv'ZO:+BGuūj;3n,FoFՌhw4-z2qZ:4(! JbkWr暋4ұ*Jm7SU Qgdfu5eCLf4|"( A"amm x$H/Vi@Rp幚lkBA]Ӏ$Hy L 2X7f:Z|=KRNjXnRt! % Ū^9"U1 f?;QLv%JX&ߌzY!{Dge> u f̛vߵhvӹa]Jf!DRk\W$Ib4"J4&֊%ckO e 9N.0!2A7~| ٿߞ7FddLuEƞ_^>|33_gUxA82# j b㓬)38WtHoXr=  _Ƽz,ځ.3ޜ~0WlgJ]޿ѫ૛iz]pn]`[}mκtsϙn|*ga7Ht&ɿ2%[siQkĶxշ9!}*]83kt8WްdZ4nF>wngyOvy&7E4/=G7~dZ|Msك1gvz7-k-O `9m|=rC49긙:neOrNGsM޺yfy7hus:yӁ C6Rv sesXh́ya{7o?ju-=q=q/[ݠ} 9yzi59^z]2Ӂ?8wMv\8JKmؗԟ3huuнtnwNa Zl?ye|kd6IL(dX$N""TE *h&gAhx^pţ5|b D}O`\%OE,q1|u>qbp$B!86ԒGHZjTZ]:^j\W1%X s ۽_|.s4_smӻ m]?)]xTL(Z2S 0,Z.bXTݠJi,⚿x7wb)ssIɾl=-tc1 f {F{Mo9ɩΊ\{>hznTY Eƨ+P#Fg5YS#kL(torgTka,`>I}4(uc#XX,VS'&%.%ˁl+[(cI@E %GTAڀ@@JHlMYf_~\q3N/Q{}ć 2L=n; ! )$ô>h-0B! <M`d,J#D.&!\R22Hb*RbP},dm@PV<[5S4 Pn vJтJ%b@oT$%F.;d$ JD)*T؂ TcC@" kKh *_ͪ FԱT1qRXShGaA涌O{+5)X0n>þ&J<)H("$1MEx DM $R# \C 26^RQ,؈i (Sxm*?ޟd .x-\08We>* 5U9L:ex^"Rhwx R5+M7ed+ي4'# ʯ )Ӿw^VHX&Cy{r1lHqz(Ɗ^W2׵ŽŽd"סsQ\.u:~#_uMDJ_.Uo&U-Uo2Ury UWIrxX0_~oCP.tPj#~] x AV*]JN9עڽكzH&VW ⭢+ۖGNJ|ghs(bXƾ?1m lErA+!\ >Z#f0cqsBpq! ]#r9:9C +&b3!10BZX%MbՔ#82.Ndd):ֱЉ(爪%$ 9$KhB ze^h,hb "Ң je݌BcDWqń( u WZy4ꪶ FKVAOaM@⹄ۘE\*HP)iR:%Srdz%암l[+齒Eaܒ$JB#HPql$qT-9'7旜rĞi _@x=yaƟV’KЋLӭkZ?>Ɵuo[+zcPVmiS]O_T~k/h@mk8Z伵?Zd/ce,q"dZP#B';el3X>Q L ^{B~of>ͨa>`) _ ~pX+CH \8tъ\-%k59JN(b9BH!qn806:48v-BT}S A Uw[j0BP}r,?5@'^PƀԼNMzGfKX@wC69̜Y )ELh֘ ly"b n)>{[2Н|0{zR#]1aOAfw'ןFP\Ii*"J|a 5Wf=w?a慜)# 0bh{~grR~֋w??>#*_hDEJ I`/bԫ#Lpc_v_u;VUB^!g+!J։*HBAuel 9v *Tg*rxy9ww $bcY pJ6qNd1vEj#b @n$VR`J lC_}^1Nc۽KjPU߁ 7fٯ&pPyފO̸ӽßrv'OW8tӨƅmTnij\ɆaԸMwM71cw_BxEمzZKǕ` &SwzwwwY5f?/[|},^k{^|zw_<S=/09[#!Q~~%#xMϋ/3[|},7Ք:Ai\H%ωac/3Ixe곋/K{^'pZR!.P0> g2KFeeb@7Ҍcdel,ɸ|%u=/>8lUV(V@7O2RU4Ҋ#/KP+Ig xY .jxa -},^)QW<}n%#x GL0`X&iљ:RxJ*)F°ZS! 44DGﻻ#G)Ɍ1Y_;/*_ϺycX*#_iRdN&ijuR&ʒNon37jf}k./<;2x%l_T L-EP^m9bv̦ܪH|W&ssc"t$cD,EJxھ.H3-8f)PF SZ#*Q0iH\71当>0!XjLAݒsP=x%'aN(' 9cx=ڃX܃o_مUK{-KYkJY*UiEsM,̘۰LZiUUUكr/~ك0 q'mtZM\U@ 4VuP7ܧ˝4bR+Bbڧ6TsܧׁS5ɯ[r8)TVaG?cyUr*yǦmxԑ/g1yWv!Z9!'rsgZZE߷2P*_!X"!$918szc8АWt_$?PmZaŤOU-ꍈmpq*y-3s!Ô48Zcr[>.uoWi;U^*5v5k Ӣњ?3 Qs~CfCn.NM/|ԣ5!:xFŌ>#}a7e. &ٛsZ3Sl~j΁06yC)1.HoN橌LWؗt/~f8"1AžFPBwAQݿk0y2s0VąJZB+!̋^I,=P[A;w,NR8|j Ky8)97nUTW,xy1/s"eNļ'b G nuW[b3.ٌS(3O;.f`0REwdDG"4C\ 0#2~$\Pn 2`uhc'q)B:dxZ&jcFO I-ɣ/f\pMĔ&ȹ ^#94Y牴~Iƈ#8Y'gJey+`SB@N)ޚCp7W^(g*/|v_@SNFrt5Z7#Gi&2J$8U w!/y!g oۧ4f4vR]v$ix鈓?AnDKWTpz'Nj YS+`Z@*9L:F2:)U&Rjz9pTɗGe $˚v=6${.u j5 k* Beq}(\a^ nJvڛs9NJsZ}VvbY8k=|J$ܓpDJؑ=ǫUNCN@--s#B1-jY:-i +l 6lGaLgAvz!UrO{ЬVWbBC$Yb| 8t~~IFq8=-3p|ZѴAWY>=yx#T٣(cRGb.<<נH] OVd:Ce?|- _1 .3+Tcdh*ޛ/v$Hƕ@D܀dS'o%QP) $N)ur&jʃJirK#Zs _H$tsZK_׹(Zkr5ZnVO~rvޡ܉4, _&*v d2FAyHD)w걆$֔qg;l`TdX@$$%qr(ndJ>(-zru`S3.\X!W) wi4 ^*/u/)%%$eaÐ9w$IHJ)zH68&ge`ֳU`l\SؘĺuJr6|1dL=QRm% Ey^ ._&[S?,>n3M8-Ä"J2ށk1tNsR9;\rl?y$7Z$DSkLMV;*58@ ̺NaF%F*Fz6/(뜎1j9ewIW*B9_VVڶWHv}*j /aDeԴ! J>8Y{>$N )X.WHNnCV39#%t[=YCQ21V^$Jp/ RջMT,?J6y)6jEuΎ:-0 2,c7S8SzYВq$BcQ ,U*&T.Fx{9nf[n57qr^ "`aℰ!#8eFDYYsE\o܁AxZqti* Zi啈U!HbEX &BDx,-)jNmT p#ز|sZֻͧpd=[1җQ\5+Y`!>+%90EHj(qO [ h8?/K~a__G{*亅Y܎D A|z'0=&K0;/vPf!ه?ۿx 3VcMn3TPR;.3D$fŁ!U@NR@ap\>WMn]" lF8@ 8]7JjŲ/je+,zZħ~z1W,pBcF0&{Щy5@2̘` ¯j  Z>MÍ<6IΕ¡/7wI<J]t( `#F8T`[0BJECLng,*VP_D>*E@`4U^ΐc1 ' K(&&$ӗB̰fO+$sU~403aINraflLΞ$qȻR2Z^ʻ/ ken\D]=UL`"7kEw_^qf4zrfI ceҍY:'/%{2kc=Y| x:PiyRĶeB91LfC/>,~Ca:+!OVHѴ {Ad 7yX |BKcʼnȇ"\)j:=%qz*&"[[nűzEu"K./W,%զVn vl.<!9*.虄`IBl[Szs,_05hk}Xf.^_-,8Zp\Nd&uwXG+p6`EeۯFz͘AWSԵb+6egKstxx>n!pizQux޲keUfr 0gfâ֬q iݕ:jp:w-鈋=s>&ƻUݳX- ~;|rG/{;͹g^}U"/<}_o_ߘ3qk6ab:L=ke;ַ˻#hƜوF5?]E 6M OI.J~cΐhc)|{f aOn[as[ĪhOq&=جlfǶlcE,53,KPȕI(֤Rk2?T2rf ipvhaiZtZ>ҏv-HŽh1ÄGu)pMpGLɲ┐0&rr먆ssģN2i!8%84u@M 8v)#9&֣b0%8\4O 2՘\5;Z=~RjU Uq uWGb(=k,]~&êjZ <13hA{gmAq1 +khCj,c+uBڶWX[Nх91: );AY/B4/'qr"m-%rZ]Ƿ[>\ӧߗj[U q?}{8i53y#yM,] piDl| #lQr('ozoض},q5WJmisEzdSc|?a#)#o0F=]uWa.K2B'w DĜ 7sMnA p]q䈯#aͼ/j.FBj)~OAK;*P7S<*!ޝ1K&C5IV!mY#>tTzsƽy < /¿ixґ@r 130l` !f0x\1G?jPt{o2L561a#SlMI44ZtV_9 \kľ X5| \vO:"p ߘ5z Inڞ!FnW# !;1g-POY|vOo9CkղF1i[j 1J5h,`"MmL d#4Ĝ&c1!ʢj{yt&oW^xj8lkwZ]_\ӕkX$+kiŸl]BLY"~r?`?F2lXJ83NZ|ܫhi F7g%ۗqA'g곃VKcۗqEƚ:,W~ځ9Hsݾ4ztH#4N_g v5y/=Q(]Q? H"|]ձo.~6=}Y(X]x.(B-|/KHz2k}hdWlA8q|;CP ΅d@BTKg9Fs( ]w j0:wւoaJ8[pA iD`(n疈*bOa\Y"C^#iJJ;61"lGD*&[u$ĽPZ1*jA"]Vn9.[C^^q^J3[$˅|ul!N{R@'E^GDE)L|d&{“;O' !R/d@${,rVsM0n>U>`՚On_ pn75suÇv~qU`%W ޷l@+>[.zA?w%Ig !ٕA}0@9b \/+ }Y1DZvUB%k"ǹeT;s`(Ourc WpH[>QiQ0z%'|2Z ` {uP}uW >' : AYbuf滨 !>,2M& $}g$%c!F&TbQ* -%m}]؄撷 461kYW_mYQ2"NUG=n~ۗ>H+%,#'cNr DrZdUnW 6=,c_r 5O`]nxgTa6&`v⧫߾۝9C&nJC{m! v|9z  aYJiaw_}>ŭzwq懀>14s2kpG]` t G7 mA4_ ̊5XͬhТQ͒CpjJ<mnMܲƅL%jW)k476ﴳgQ>73BYB<˜x/ז~>21 S$^fH0ucC? C(9;{]>&2)RtKS-p9$=gQ-Q$mV_smm![x-vrͥOAԲ ޮI ǪNϮj i99Iˇm54< IK%j%K Uw_\#>pY,]i˜Dip 凛eH3@Gu9Ø: r7l|,wP$ b7wV?<"Q=~yΊi&q+I19FKUIW-]/ソ~gjzO/S ~{_') :W<:x!ѳX̀_.~"I,YE{xX tgF@'%(E%hd!;rФ/S|)Ԗ溓2:>5ũV1Y[ae^(| R3!fƨ,Ď1#MZO '%ܶ<ݣ^_m5n}*4wþDG7׸/ '*l 1@L6 ѥMax]V9kR9C,O1XmY"#FtAjɑ+\10Hj$0!\Utq ـ ծ&]q>,H8`ȣE^qNZb1+Ib5"D$1ˁRȼy:?Z3LGfQӞPecZaT]5g NӡH^!kw2 |>#p~#S9RT #!ytFq$нByHHyOt'R"8Fī$_~TPjЫUz-$3_/t)}L:2fJ&'4 @b VJ{sRILà֦^TO߀0Abw>0կkeN"IбJC" ?HRhm̌.}7!&@MZv(z M?iѤ8ܶInt4%jJp gurJͤp}ZH&GA뱥'ݽJN PG*%/-q-Ʒ&^`7Vj9 1@LgDYp|6L8c1*͠dqVl2ĄJҏc1x…trV}_08f(tRKC:ID &v/Һ;o  y4eK4Xs,Y!iҨ W6@4_`:1X@#ǮͮX5(l6!UsJ̠<)l{d,(w^hS٠: 4z!4I ڝ>":%Q2eQ'd3jŅr DꙸLF1eOjSfے EtP8Yʢ>.H4\Gej]z2R̓tK+"9%h#thB)g=8C(1KV$ :TAv QUD}ngt[S1`Q^vn#XWA`P03br\;9&z*8IDA|Kƃmf9ޘ~s$Rjol}F-vOΗc=,05 L@KR*\&U6q&:"=]-5`Պ)P%D|\J]: ~2$9w'O+FS@yL_B񿗋ǞK"\љg.5R/% í2ХI;KBy YDHgMHuXE(~$ r`lwF(㖓jkH68z青.1}"yW I904!ksǀfiHF$s]E)I&"Do˂R$ `A#I,˴U#8Bq" H(rJxWv>1a*U# B|TTk7.py27gpl)]Yq[BVG҃ƨ\=n \ScVv|㹊j*u7>nU`:G Tnm9j'vUn'-Eއɧ{IP/m?P!ߣȒaCZ>VT(EG$ ݷR~t`sAr20Oqv)l/dKb3nMY9*lBIe_v7 B$ڏhXS!X\82eh%mr92&\$T0Llp,"J)jK!B*><>#J^!Aru:M?ܨʹ_-v.&odp9ۻZѣ7flà bI/ ] k8Jpge1b/ C֭Tdzov8MÎV=b Eyz ͔wCu2̝aRDZOu J wSA2D\gi G.ñ_DRnoPBb? ݟ‘T>54\_ ydFLL, ^jKe#rՖJ-%2R@=Z_ƣ~y/RJy_ cDb6ւk `O{ԺiX]h`^{EV}6G˃Ylfe.5xn!BH噰ja=].>3 I ѓ< d%Led ƙ#0DJ"4AZ,DyZPa8A)s,(E .OO Y59O_ÍmK:G_!k_%V >>4 i4 őma3x27ҁ_d\˻xo;˦2.@{mr>ss2cUcE:k_gFwfzYɸg7F{@#/@"` 2彽}.p݌t."9z^g3Λ~zDRJf`ꈢڷ2@ ̌(1eHA Nf/yYbΠu)4\'cags1(N)Z8\Đa#!>V:0 N Z.6V11QH Ka$l 0WQ8LDZJPh#MbG12ī_:/f0XLE^zL oAA ط؊=\4sͤ{VGskŇ81SZWknWQ[,2skRey+.E^ ?-[$KJCZp5cAT()&q)0GX6K7#槏H\9hD (/36ty e )5MXPks4JNv@m:R!Аj'XD2 ĄHGT &11KNDJjāԜKОK]nM d|GbD3QjA'euF&!uv[-I&`2]d05(`:U( 0| m$r?6[v xϽ>u<3?Ls^"VxӁgX./7F㡽Cl{b>MlB~Pf߫t$Bޒ,Fڹ3l%3Yjv Y-=Y۽; ߧˣipy4 U#CF'[@:ΊUIE8MbĖ8(ݝd8o&Jb(%5/۩E⊥% )N#0NDmL+DrojΓ_?{Zf}XnF{YM@_6)^ U0@,a* {K5q]PHiM21?tT(0X <睹 y>~/z:~-`?i;OxG]HS_ݭ^wu/zjgGͨovۣYϿ<Y*zkR;mK< }}/Nw*oVxzkܿvzw3u3 X&޵n[*/0^f7rMSb IR{IWy,Z9λV9zsrxy❿OfCt2MMjӳ<>}=>.ߢj ջ<]&*W:e&'iNx0# ;Ӌ 5O?UzNmE uļW/NO=jWqkQ"Wó3NN?_y_=O.둅Ϭ$ tޟ&ޯ4 36[IXzuF*j GV|S>OI[^55j_gn|?@Ժ܋^aO{Uy\Wҭ/c'_}Su)vV{@'9>$.{p8J??q8$7o/.f#\PٟlB2p7:~M^L_1Iw1ţQ׋O&/~wf<+#@Xn6E:n#{Ls$˴7ZYwgdY/CL~z7م;|*;)i>|0s)}>v'9Fs݃i G*<{iBqhOV?&6fkU:?8 ~*{x鬗6.AA0L]xm>>,~&)@5X`^tgb^pwF*:@־.~;j1`_l^{{}/Q'q[Tۑ ?'egW'ElEJDMViG6&:~ cK"_E"t' L_)XЌ0s4BT RL%JNe0_bo|a+^چ#gF̴4Y*X/hcU!Yr`QG$Ulb9 =XYf)NfwuQ5/8+טNzϢda%k@q pXΓg&Jc6ӫ!o8Yʒd,hØ29d+bujsVW9ez$]R ƤZɋt-djM*'DC MHրhLtu Lϗ`T(0G cuFk("hd YҒi$H"bԡaR`7d1)~FH4)сJZ#PǜtH.+ t\)]TۨL'/zf!4s)x:, %r.p([zRbuP@UkD:C4OzF C]Qȩ%Kp“CB;-!)BGBB9'3dwˤ萭1ͩ U6- 2q6?]F!(|u/Ÿȕppi82c At$vR #eɄB`#8iY*a:cCJdSIվeBJ;wQʄ-%.L^/nMYo>n1h[CĝpSE^rKba:?yZRr&ӝc5y8(";o7ٳt]<꫈u,LC(~=: k}::&rkPz:|l1d g^ $`Z|9Mu'oGFd=9NQID8m #P"ҢjՍ)]CB蘱79_s!F.//ߓ'79~M%7L[oDR {?o-m K]8qJvwF!Gw=w[z!9AeTP ώݝXy79~}};OC?ޯ ?9Ȥ_w;p%c[X .Q_C`:fTr O]?]T.&bnX#bÔMK"6i:ԇDSKԁ %ZD!=N& L{tutTx%N ˡ$_Ol&Cg!(Bznr,NcB2f:kV- __gZ2F[''\i;^W, !-vm,:Lɼ;(Rj9q֑b%A{I AI2Ó"htҸ A(DtVZg0ܑ"ei}*E}sbXRb>q.+fSV6#1a%Eg D3 nwH.B]E#B-B+n m{.xoE0 JMUd3}wVPh=3v.v[ӏzB2B* cͪKѓep\y0]UF Ï?ÁƔnufINU:j㶙C>\k2&p`1Ҕ?Ȁʮ}cmMDߐ\=j*%:)(+FISRl˔!!)h4R1iJTsgJ'u@L7&8' 7nLjGJ2&T S5f-];GFNᨡ?_I"BF0ixp%5.ކ~j[9֔YqW+">\lsr՗ý?s1~w՟}xtw˅lҧiMi3>9< §bE8v_ߟ3W'>:aK~~*Nöy-l9K[KfǍ #!q)$[A3nN393O9t&tBB~"Z%SWw$Jg(", Is!Q"Q3gP"#a ͼY5>UA=JOp?G^ktNsX vaL,.g ,DeW)6[Ns򪿓 ͇zH^Š 2NtC$EM5>C͉=ޗ /^x(K)E"=b2G= 3 91kϟwRuY Z*1C>ղF$MsYSGᘘTdގ&3ȑ˺eLyw3u"pay 8ixΐ0,%(B` #C4R@xY?HXd3 B< 60=n4崱Y[2c:%3 Aҡ5 H}҃n1\zҝ"  xEDRRJ4$F&UVmKZfA2h4xsSO_03 Csb4M fӎ7yшA)ND:%3 A抵 ,Y`l&B&hΫO_0 krRW^Njwdig/w5 q'1mģwTye;d D1:0$}yhec"Mp'Jfx|=tk@3l)1Pvd/aіVKvŠ$d;yں;۱l!$7.52Z|Ʊ_:VA>mSJt&tBB~"Z%S7Q"ѢL@ Iɤ wVõd 0"(ۚU I!ȊWY".2UU; ̃LXڸprDDUڷG }Y<+gn(O׷Ɇ,w;2 @X6]o! 2GlBÍ3u&9hbH @)pP'+  Rcb:M`Ü/)\[' jAMzs}Qi Hc N@8!ڠϭ*V=Wm9hXIV51c`\@H\W4m`PjxC)K>TyXD<מ ,fl5Yv\G_2B ŹD%ӵ"?3T";h%kSX0̃>vD͚9-?VO)5ڲګwW]\pQl܆ߴ,ua' buzh-?o`.FkGn͢nrBt E&NhTe&',Z]rNάEZ*lN=Hd~4;([C^>_Z\_\ܼ=0 I3ӂ_!'MzM>_[X_?yqb{Kε .h`48wp2%5MR;W;<箰dC;gx~Qt;Zڃ&ڋ6TtTyytC9ਢpXg?"nWO4׷|ϛ"g~w_^e.̇THQgbS% c6߮p<;-hpK |Ev8Jg3L41Rޛǜ!x]ænp=>|cHp 4F@ttiIuxM ._MWn~rHDyN5g " +wU 1g)zHRBe0QidijERIQ+@PGmm"Nmj*6,0.9N9,dHz6zI"hPp?7KAO@_bO'Ѝ 2\meIQ!kӭuܛ%yC@ k 佅 @Qe߷* tγ15);H)O\ 0Q K- H3Uz=ҖPtaJHAb 䅜q^Nr_vUd}N!"ՌyG iM!I4NZYXsH3(GŐsF<,tJ;գ塲ͽCB4v/!¼6? W/yCG4zd/?<+ NRw޸.,n޹WB.S\qI\$/|~ ǻvld ;/p|Na$5vGX_'-pJ[16z&xPKL牎wuaEp Έ pkTo0 BI@r@BfPm>ס!5C-_gO1 ؉@L3z0 ]Au =)jro44cUF)쿫Y?|S$^D-v0kO iws1|_W?]]PaNTݦBPˏ;C8Afb0(48.iԜw]nz/#Q)14 ͉6:^ *@ϊi:f4^#;@#pa[3!П{D5N2fq]gfH-.^ԮW@i=ɠ`-O6$.I ]M:H Ismg3 8ѯ3)8icҐS\՞WsŇvf_ѮPvN {IdSU {"uHӊɾ߳$!w2EX՗ɱ4|bNQIJ1*l/EA)I렔Gy(_>tp0x 斣9mLAt&q'FiD"D ݟU7] iTiEE$lɁ;!9wYLvX3zgP3N%MݮI>s1Hy8K^]B{E/W?fg7( Q RŽIG h [+ qah4QH!-jkS.Z% ;p5v7FT1ܮ0FsR6[(" )xvz,ǨI YE Fqii[Xzk&~%tUzhAJydGbqP#NX:dދMQ νw&LtaO %mH2>:|ŵR"tIѣ~6=bܓ@B=28u3a<BH!#b*W-7z]s 2J Y`*/-Dp2C \Jxx0lL:0+-3P ɣ[Hbs/<38đ!r{(d^Qh=N CM boN1`e$7>nX&'ݺhx)ig靜:1iEIz6"( O>=B{-ܫ){Ix+8jM؛ ^r^j4N%$x B .5<շ'AB&P^|x^}xϒP4vD'9uY"1FYs$׹::$) {YXIysx3#,^}.~s|L ;-VDžF`$ndj;X J8RضwyWcؚζNQDmYtiT7yʲxnʲ_"e㼔3"!a\T*A:ӄ;u]8O Cส@ڄUEr1NYUx8)~44*8"aEU̒cųZnziUXUS= 0' `ȜzW>?OX;!#|Fe1r)p&)xv/bg;ue>B4k%^0Bo?:){{௸8 8i@sz,wB*V`{",A[տO@:~{(oͦu×*}%u6{5Ae?{S/߯z!ijV$~z\t7R agc-Ow'h@CYaG0t %4X3Y6Wk]>Xgx}Mc$ЌL mMDHd) ._4R&9De А#Z :V  Yd %*u()A"L!i`8xi`NSH0ė%> &$,K7a1gCI`fqJ⫼4ͱ& 9Ǿ>jĜ1HХt$4sHQHdȂnqs%Àщl A;y cd!#8ei+ t Ϡ.ߤTDF(3 f_'c]WԴ hǪ8yUEqYdUӑ8NUq9U G! JI(+i%$w")4^o*_Hq4/zk/8&X|?ۻoχ1",a1U!mBhBI.Q)5wC#-da&W0Tb77 ~aPAp_|B_uN#7k=x":_sB=^{|s/$rQ^(bZ'\[gQ2f0W瘙EU4pԥHOM)'Lgi[rHN ARhd5F=ifQSH#tF5綷iZi6~fKe<:0a`: q` XBd>^^}@vqݜBX/: Aӊ mqxvШs*R4:ρ3n7\rPjCr6  e I]l|>9={O, ]ss P(,RL?K77_TLp_}/>񫛗_܀WpPЕeG)=;z^pApp Ba&<1Z`a8zC[(Xa0A(3a0z .4)3C8dgHs|_-$߻}Ljq;DbOuC\'p]pR$K_q>\K=\G񞟂^ol>-:xտH<]o_ 3{bAӿ˰yX󏜖ֳ4~p*gK+c!pͲ^Tٍ0-mmi)nU?ezrXGژ2 #pw^ /vUO఑h=;QHaٔ]q\A '0l(r.;ogRzT}Q2 6َٟ= 5V D!Զ=շ'*RvLYýdec0cH)OA+萔Y /PFl'<#-{2᩾@bJݓ +\}{4V$76(O"z.#j_TO3NUHxY{izS@# JI$SP.OA=,o%bʓ5?r/׳?|sTz~zxzUx4IF:#ҰNZ:N5B) s}N[}jz͐Rͪco*4ҢSOD gt #9L8@;<3@Lj(y)*y±7 ƕ턑BDF;RCs w"ʁŽh0N{ӆhOw>gJt#tӝvI)D,[!;t^"Ya""V$akƤRӝODp HGT]Y1i1;ΧPRc>kdj(y猪|'l輄5@9jLGS 3ODcRpnm Y,̂ % *VLSj#Y/'f_ ! fd/'E* Ej$ bH԰`8'4܂tn UmL6 [*J%9%J%[s%ڕZN[?o`Q>5Uwf{5(`b9LsP@`/,V}cr uq⎆\Oe&\w~6h[QG1hokAz}p!s[t~٠[.W4k?&V4kU'Q|l eяoKfd4?u -bwOlQl6ݽ !GѺs_ZR3a; I曧;G0'kU0n/ytY-$j%B5p'Gm$ 9gtHr_8p(s E5c@!-h! Z =qxz\9/7".OTNC2G%)X*_HpjAXM&hͨt{LL~;1<\Gv˻6J;I{X9*g?N=X'366P]#.cLig80}T3@%^z3N*ל&BLRyo!XJݮĶ#c˒舳: o mi3'_r҆S;6%kvW^Oowk1;rEbYS O 8esvZsGPrQ]wjv QFӐ#TڞykuA85#\\050BN+ _@ý(nIktHZ%D;nYbg)GL+ۄligV&lH惝l; do\kQ?bbh O9)#ð0r94w֚`.wHUزQ(t )uZ_ZŷJ^7%:8?sރNhPS,w2לY*aS[qN8Q>9cF # E1L@ %sQϵt[δ\J4 jMhp[bѧd=iXRG*0qRi7TXRT2l )՜s#Y%q L@a&'tFG*+/rB]'FQL; n&O 1 XJ.FR%uL(t:tF3Iie9暵!Q]gz8>%M.P:4Ӫٓx]MNT\4NZd?jdxs3 ]ZJ!Z//^2,J[ Yz9 h7)# ,Lc/nm|QPa\~b^ZSkhUDV]%[4~W&&~ +Mfכb߱bp7#~\ "L+LXwvQ'e0^0A/֋6n 9D¦ݝmDlg#JE,گ*P.7FBrJv6W{bLr[.~@,"˕1 ?Me%4Zl!@%AI|&nB sxhNGS(Q}nr:BR: fB'P@E9##1Xc#M&#Zj<{AFH{TW\iD6FL[#qwLDk[2 $+(:U'whi7+zn<#:hz4&vݢ'ZW\Dd ڍWbyPGtbDj,Xw->avK!!_Ȕ$zv#X vAщ}Gg`]=zu[ EdT2ҥ?IqzدՃt4ʚ4 4!;Bֻd#`k"O''_gBԍPk%1<&jN^]q%[v1}aT(W;[k\d(rōMqqYSܝ .l.ǻ_~^h^ݮ71ގ0 X^ 8wɳl5٤/0avRs׫| Af{t A[L5˨4cL33܁TD-:_(~F\ dx;k䴔-W=L(xxjrx!ArN"21jrgYv!'DʄR;9=~1oM'^V.S )6UrjbyP''M| QvBBr)uvøIQľvsŤ8mn)$+/QqRD[ʜMΟkX#"ӭU`c;9iSymVk'D<ָ!\r,%Jws%\E(FKS]Kl2E & =? oF&4!0 JE5*%A2v6xDh(͙ #)\n;% Ŏo >rMb>.(>[/Ϧsk@e<۽c[ z^l/H\,+*gx8x9P%znW}8ֺ.Fa®83̏ev#c:Vgr),QQ9E9wwD3۠9篳?fũ՛3s̬_OౌlJr5L|z>*q3B^L\#,VNZ(q3a"´ x \K 3WʾqVY<2T4sFex7l ģz#2P{>W^g#1'w\0)$6.#圕}8uϪX~{1zYJW I}5N5ڮgV4xnfe>iW̄hYL/˺2>vgR^+AQi}yi%^V̮CR>(*<"۱mBVS(8XsqFRn9.:VM;fP6J8=QI}>K*`|>{>PhYYMe3_,v s䗼0glXY;+)M9nxg3#^!JrL!n|uĕ7nTo*ݟ y伡T{!H hWᗕ[8|Rl;BîS021X1Yn_JglW?2vX@о oNF/^_r9j9)av YFUOK__6B!Ae#\U gr7)Îf`tI0K RNMI020'[X,j}ݏGixt<Ӈ,W )Wdug A7radg 9vvU[GS8IZ^G~2,4f(lt\A8f{KrgcsK`nGS\H[[pn/`DVe3p"}y]t#O'ɶcIk.Պ}ުLe)1mEB9!Rؾa=537b*.?I 6PaWeLaCn2gK/r*jn*vI,փonߕJL*Nä C KW0xWE {vU&t.%zv1!ѤpN# <>2P/#cX^q(zhN̦zc̓WV:GHpS0Apuxh ,b~gt00#A繅IಜqQT1rLtxTTKDO,!Ae©cS W^PpCYh~ifog(g̴x<#}4(LD dUi ,&{,E$ s<&2XCI9'(*f${6cMn}ҜpCBv3Y&Ӓ"|43oF\K7},Ύ#3w#pW^:}JJ7T֯T ߫7cU -aFIbwsզ?]jdo~]ۂQnav!s3)y g̨ǕT__Evtr *3O:Mi˘n|"jPʙN6I,+:p?%iaW?jɛ0M1^AqY64=XQb  %bD-^Sbe yuDN$!2f\y1wE@cNxm߂J)߯1qQ/~e<cH+?y\0_Xޜ/ⓤR);OrJ <(G9~zp㒸pvgw2i; ,v_~/DfnoeʇwdL"4xTү*ZO~h-T&E\TyK RA-+Wg{KaRU;hVC0? a uD-`L#9 ԞF7r@v{Vѭ' ͙U1\th|ʙ+չq$nNS 9۱uZ%shfd(̐~y!)egk㿖ח+EZޜޜ_??5›|1I>ϑdY]ί*'wy7*U O.gVު,1dȫm۷V;5@,:hX+װdM(N#~] fNbW!:\VBX5~ Ӽ;>/׫O:~ߧfvOI! 'Sr+:kk*Sgsts=N& .=*c@[jEʎfjL|L |lnymwbLZNm8ecݢҪD5rm1,\t[6I۷EXz7],u魃ytkm:e\rb^9(,mt: y6 zP@X4e1#%x6|V.հWڢs.]RC_GZ"ѵ[`F"H<&*튇w6هi{~2p`p۩n1tia3S*yi0%=Ǽ-<8| G ø;+J >Z%nQ{b9PE2C[/VJW}Djq}WF`3* a.$&KV[!"+qDR^sF &qHJ}:[^-UBfORO}oK^Dyr<݀cK{)MB(LC1Մ{S9CN=Uz}9c1y F f4ReXO)%F:F78po7 b% V@TTkT9lh7,11Ipl~D\B0 '$@$ xL/ <F8!\vQ!ÑCP4nE;m2`txURS 6cf'DnC&4ZtT! aC1W .0W4`, 1"UՄ˦)`:ۀVVra$ `T`ύB*K~7>48Aegic$wE˃L3NaVb83F;HE^QtWkQMQE{CLSDA6s\jaJi`y MS-aֈT+$ODa qbX\eीV +BƦhVjQL9" ؠtEi Xp:ʃ=y .!T tSW"K1a)UB@k|ju$ :*Z*: U#$LP1ˌ7ނ<z?7:nѪ&]Y (`A冸R ks٨lװ @ѣbҡ%7x2E8A)rCQ9,65JJ =C &; GjXhO5,|.X1#AbӦ*XCE#Jag b,a(" lAڰzAvZ zT~Wǧb >)OAǻc:3^~P:C7&ʨLlD$VҞ햲O>٦Vr]ń`qfcoST)`\EE!/\EKtcg;#;M·`J1Q1XvLiXw֭\ֆp-ҩyF=Z7Y99u+ GuJǨcN+dZdصu+f4׺!/\EKtJa:1CnNuNϒLU[1֭ y*ZSZ7ổb:cԱn;erznŌZ64䅫hNq2,] NJ1Q1Xe$*ћu+f4׺!/\E(X >@[)9S:Fv $[1֭ y*ZSq^vY; #<VA֖mNNɚdJ{g\Ј&a5aLLJ1v9/t{|{n=93żMx4Ru KfjメEZ";C:l1P/&UO_>hyw ߟ6E} >0Ɛ!Hu [y"cpuu;>2V[2QѰ,^1unnlM]VzU%u.4$#!v#=ev]n{h[RJA@}@"H2ZII,F"ӠᆥsKF'$(&~@q㆏YP=?[Z!?{J&Fpm;@BE)Ddu # k(*4` lhxeI#8JGoeכgRԗ}5Xl VڎY (Η9RT}4ŤNގlSL.݁ zTSQk1j|ۅ7Thvw UHsc8RȊ, GuJǨcNpi72[UN)$-ݳ ks8R׺48I>cT٬BvLw۫-Iey89< >;MzoV33,*qa>o9ohG7aSy=x(1H#b#3SZ4&wVO.dt7DSU!Z?uըw_Wz݁Vd`v8so)c9YH}ߝ)4Wt4sqйMG%P[Yg\{17'6 aI1^g P6HԈx"d@>U̱T:s R^RmL@R}*$3B8*(ZyJ QZGRA6'j%''FA,`:@M)0bp:x Ap3I]Д8ZB'B*c AJD>Y)Y ǨtT[=-ZА%:c(V^AQ˺{2mZАE:=]]2 ƨwtQk(d4.@p}SULZf-55ky޼feܦRmk^sf^s]E%5ݼfcV9JUWVZ juW?ke϶/,Aެ\nt$e'p-O|:˓{/ӭE VH"%pcC`^ʰڥ,p ߒ!7Sku}nlA5\]$IAO3EU0;d3ؒW-ҖeV39Ab>֍ErvX{gv2f.N?qO6U(ӞJ]qx꧳7oݬXx;ݞ/ӭ6|^v1=:w/ o'rg~d 代o99eר41">WL*`fD7*ܯF*@5fjb$!NHM?ReA,]RVhٗ}S!jFËYUD3PfbS P0Z65V^Ԅ>95j笏A @-~(ԩrJ_s9p#E]F2sVA¡71AWp!85I_>!ԧwQ#C>q,[ >.jMd큥{Ř:D?w ?3Ɂ{RƱԺǟb0KB}zX,MN&4rI_ϺhCh &0hHm{MN)TBLPIƖ8%o+LζC9@RJl4Fh:Ek\N=ir7 u5sչMmBÌeRnQu 2457ODѥbBa%(rO~WU3={#4'MEG89\q(!K@UFy@=fJ{RATv@Qxl^M^D"m  VIDӉA `.y Ay8$l,B&eMıڇSURlJLJ *uj+M& IA"8ttP >8ȷ?r- f0 N|lsRkUX2`%UZ>s&fi9ZfPo@jR@ (VIb(Yvjma Q`I1)S\<&`4*&q8Kjd-UΓMG49ɗmXojmNƠ.50,lf%r-.JbRcg0D] :`S*',5s:GTM@j]R-8@aj2B-;w8*JUڪ5zWͩO>K*,CqrQ؀}HZE(+Ej_Q5X65\T)1XsbejT%ugtEPܲ( 9; RpP@&j{Z-N!6Asf [i5IIyq 4Q%Z MuldڨƐD]aζR5,q=.UI5aC554GKq%[x%rܺh$9i歁SQ%EЧ#&\Aykh5ݣ}h6czK2jKURDs8 ^F,y;BIg_*M*'m=`!d+IvyGַݍʡj20CMfk2ƱC]>=3KinHQfX0P9Էb)خ,}]8<}d鸵BzV8l^ ߡV9V㲧܉ugN\@ӗxj,j@ܱRȥjښ9'&(@ML ˟+V5])jZzuׅ*T[li") gZ'"EW8 TIUJD )05S pL&/ 6 6LǠL~eKZm( RRmԗؘ[ѡgFacVMHAފ PC5!Wd>JmEC?yJi nOEw/]UFzW:ĬbzO453['ȴ<"I=a$nq%1|C4"C D[v'q5K<}ƱCd0K=cu_?tYQ?:>.j,kc)жNhKoQ3vB7KݟvwDZt3nq21.MR&yƆ惤/)R5 sj{ clNoP3S;K0Բ;6UgSs57K*d* TZ-Oتy 3zנqj-JMML`/һ&p'z[H}/t6QR0YoY˶%musi+yZԠ@ejZꍟI Ndۋ%f34B6!B.hSg9 u/m/Yjɇjcͣs R"5FaCSi%Y *pL޾oT%ɽN3grfRPjRI$y15M]>s9ěr|9!{0(su/^olJ?|_ُ篕 sT2;㹬?]]_י*LRxXR%X5Ix)ˆMluFVkcl<۫)bؚ0IO*Uk'4M~yQ{Woʟa})gwSysv~v=O2WO>"f}}͛?M:{Q]_!>n.Wg9 d^c(eWfR~ Zh`:.YJ ӇpJx HXׄ@И 5%l>)g+\vO={2Xf_{\fJefɣGpBBtHa],w_Z~KUɗ糴íOc\)Sy vk2+훳˹6>yܣ;ID"I(F"3WggjZYZ2{"+k W4_KD\L]xnW]DYe*ScA^e6k-Uܣ؎xwP7_ڍNH~AݬK3"Pɮ(902صu 1Bd Kzuhّ,ģكYٕǰ $4Pn̗% ֞;p YtHݍnȭm.VU 4Xez1/GW񼤋7x~~9/'G y޳ulFC Ҟ,Nz[q+w}/^8COqn;8?}[Row]^ҿ؝.^ݼzPtm|FGط^]m _-ϏΦg<^:j-k1 rniVCko?n9KRܢǼs+k?+2o)~s릜]S{)"eɯͭ⫋߷.Ͼ}t!|]]7g*ۏ6/3a{''_Č1~E]W9|g03?z0] W!7f( X?zb~c=]@'](!k9H~t|#@,OE5(=?ym+ڔmL-e&ooM nb:֘BhqR-~șߌ-R=wi:0`fVh ~V亿G: 1gO{t0~,}a54ƀm YyI!}/yt} ѿ}8m[UGݡ8lՙb~qn.IUlϤ}&3p&|yv1/1޵q+2l$Wz,%8dm8q%ds2@"}=#&;Ir $LUXE%sV)!O:$*ueI3(riF:Wl~XpD:,돗 <5I}3L8h-nQ_Y_߆?;'ϮҪ^&|C:h4өF&W|VT:|֏Ka?ݠC_t :XpΉ(9W%5#uQPxir{TILr-^ZwLR `L-?.~2WN$ճpK]XH[ ì-‚`NQ0 # LT.SbtY0V0ZΒ,JB2\6+4H&\SBZZTٍ \%)'c*Ibr,1Zbh}ٕDv7"SG,Pq!"l~Sv8,toWmdS*xv"ŭ eA\p8PBօ6B2Q&Lev10ԒX{%2gIYV493 xQc Kc*߫{ =;(M)@reyvc#;1%).̂Xh9Vq)xiUI(JVBb@A %tcuu{u6zKy{zWyC:-[==˃ΆT_~zw71׀;ASvzrGpyS U_:;ίP #op1\,V|ʎTCI;X1*g7V TRct[Ts-1n/.p,?]!#`)SN}k\]w%=1Naz1+qV|0X6y$ANTGDz`6`jԱ5"P:U074?dhTIJс}p.TB~B}1B(#ATZsC>oi·_~EĨ=x  $c 'zG$zmDFzF yKѫSC^+q׷!'R(RcTa·vy`씧WUбSwyE̋ͳhcl΅dRVZ[cHDxsέ*p%cFi%S/JJ}ysiV<` uP2LeyR0阁yF>3 Pnȁu^cJz쎯;=(a0 ;DE?7x@ vwP"%-?6W 8N d}$N`hw9$ ͝]JR-e-x 6h-nňy)#RSKY*Pahn(8;(Rm&d rJ 2 -˒pRȁfCT&R-U 60jL!Zغ@ 1FK2VZSrV:0XST 'rN 2_OKEo~`XfR΄P{-3y-22>(Φ"jTJ-(.;wѿǥ`@'[Vՙ Ӻ$͌N[FJdk@ О_ =[0]/B/@v˥ X򋻧m}@RBA< 1a-gtgq3QHCuc0 j9Q0osW?A.nr_Z?mhr9%(Hn{@55L /CRt# 5wt8o*^龝׋\^_}Ճ/G^y|Kvf>߬^CtNfJZ|CFhcnu\>ݤڰZSe*z^sR^_OV;+֩'WŲ7Or񃷆&N..WQߓ1h3N'KˇPWcIБ^hYBBmgY.wA07*٬,*-ʸKS&5V.SPZW.I2e-mjewnN;bp^ T2hnY[E4IjIc6'@GR1whsNx?Z nmH+3J%7Zpn^?vpjN 'lNf:c-cR1Q.\J~!!\DdJ$H5D u2;u`!F'-H+$b8c|- 0QgzIED.ZW.eJ E b`8>P?)_%632: B"%Dqr{,rh JKhBM}eKτGW$aXn 0Ӻ02%MZ_r>zr'5vUo\\k.,hBG UO_y%fŮ2pMȌ(@8!ěABA ReY߹78 #Kp:Ly%l'n$cSYߜjhbS͉I>X2Al.yٰa[c !#K<'XӰvͩ 9ݢZ [NA]m; RaL tOW.ubG$n_gU^%JL1],ܵ2<+>DuQ9th3T^v?5'#'J&?;8(t{&(T|^,޼LM G0כexgeɢ?Q̐kWxW]WqMa%RѰ!%) ɔ&CdKt6W 5bJSfE¶Ȯ|]+I`ѰRS0k3?%5gG| ,ф Jd&VY`p˫UAVYRR19 iw2 ou? $k3 <0xvNlo6Cz^lMۊ)w@uBi/[΂>?QȄ bOsozV~qחser'U,|Qz(JUran=DeUvV2|UIOIB昚9s$eGmNk;gVLxsy3JQ?(w2Jpe7#\Ѽ+|ף8!OuU| ):vtƁX-s۾Lh[ aM'@}uCŠv8 VL@ ͌\Ɣȍ02FsYyфHj 8UPƭ uk=<*Β+ Zγp+]A8@F̘TI_$FgfLYKKxN91m*kr+˲=QPh8j1tYIK& u\1m+t){dFA y8pB2sY KwWfQd9ae MrW-3y]f_raYj.FO*P~8l*4U/|);=`IdekW}ܧ _\!au;L>ubŎUx/Jˇ_OΏc*$Hg7VgSMOLـ<\p{qcrl'A!}Ig]sPM=zn@ٞHnqМѾu M)\3[ӛGCTJMa0 )CTn">6bz$o{DAc_e<|ڄv9|LMoZ`j|mҮ F5;o'n¶Us{jt-C5iTOd&B+I-1ܜcZbĂj7=6MF;Ahtd:R뼓VHqRS%R@Oݣ./=_E;" `*cHHVM{ 0ғim1RSmQ61j 1Ms։T ҰKr\FsCu 9ݢZQ?H r;b -x֮׺.noCB^Ȕf{vԌBb(QS[W%jAB^&8[鐭1 ]Y8+z9Z2Lî٘uG >%WQ,!b?2M&2?dė QuU,}N޺q{l!ZFb g5UNqg͇:ƼagXͨc̜.tCʆnz\XMFogϟC`^̫J[,DŽvDMr'I4aCxj`LlQ}Ah"cX߮K.l ``jc( Y/5P 9>\Yg~t/;$LrhқpfA#u6gmgDkJqhrs7 ^Ǘ/W$i=5,SȦ2#Ϟo|+l=BMkv&%5ĭ{]|̂뇯mK~cwhU"-~8l?sqŋwqNˡ5Ə1&b{fakf=`K6Pk@%k4BC𭠘7#O: T(smѹjJ(\rm\DB̀"'Pj;2tR<)¥f"%ʯlcML*j[jHl-+T4rVURj'Umw uZK,ӕIup ! ǕMYnsw:$KRI}"*35~oLfsRc&RE%Yv0A,b}7zX]@td#)z;q4V3m1߶tn߳󆓰Z;Mn|y{tQfhpP"nvmnͷK%x1, 2?=hn D0ǣ`] ~՞{(d_T%Ƹ]wSU6lkX{,Fǎ̚F-TZCy٬sI2ch\I/ΈN Tkm ;iL]a{%aæ4΀qCRR.aQz(k (E@!e}}y8rPZ'JI=/K DQ,J+>QzHyYjyR#J22+D ˔R*PJ7Tơ4ץRP󈯈J9C)_G9ÈsD)@Ju&@)@JsAї7JQšU7aDCRϟHQz(UAixn@aOz<:o2Rgq8 ơ4#Jq*/>f $F/=oC)<$a(\j(R(=?H\J aXyYjوF)eq( iOR\jPz,>o2RWADz^ǀ$Q#QLI( <@ظ?ob*h PR|Qzv(EGX?PJjH="d#JǡUBA8R#1(eRWp@)8R#IycJy^M0P1G)8o $ X(Ȉ0 mwHG"m˻_aea㔲E!dG׏}l{Pn66kҙ2cBN%]=Ru ~?Y.?]_!A㩠fk.a]͒z@w(ǠBx{ͪZ%yu. I%t&G NATKz9]хղcW6UH3"i˺Pu;Ӵ\cudqF4#0Mz3JMzyD#6gk czj~RȔYl¬Y&HfmFtB̴&Ri,sZRD2*٢m L dXH5vbAf$ $ Z1 D f VPd4KY"i6 `×w\t^=;"i(3Z& L4UP$S/ p^F#OeLvFvaۏ ׏wK:څ8>YMӓ~SXr^ l_`.?%㺍ۥ˫0ͮn\)av7~N_.Ww|7bo"X/ÿ\#%g2b0?mU)D(O lW!H&<_ ^*{oRIPK$qs~k%u5>ZSV(:oL\s5Mx£bKʻv>TgqKg^)ZUlve]Ʋt.Ww_/\id86o5b_C'}k]ED"zg뫁fNo6Q(ﵭR=h*07@bq'1mir3w7_w'ZR;\[g׆y4Fj׸?g.?3&7e7O6FNm6^~s+EҪED+eR.dPo| rWf/G?\|hZEUN'dJ@-|;PU 3>{ YU< Ni:9H8j R FiݎZcMnRKuy;Dȼ, K8Oy 4j-o&I8z9,6fo&ws0Bce"q.Bч:p|wgXK:d3(֘Tڃd۞<>{K׆40yjO`xKymHԻ -ݚ :MQŻ:Brҙwk^@ h#Lwpݣwk4F.BRڭy-swk!/1%I0RVGeOb)*$cO;G8 wRyb{.}-}ɫ# f,5e"™no'7_߇ٹyp&BҨ[7ejWAQ8|v \O׽Wx+Ij۽]U~+wUuskϪQYũ\jtSʵ%DZMq"IShf3Aqd~j&>MҥM򜍣ʡ`KZ+ tϡ@Z$!0I^_&q ' EB46˷ue#ݧQ(d, A<{I! 5eNS)(*|*$Φ)IvV5G:N!&@egm Ee?5Qf7?DpL%meˋ4p8ǗV5"yP_,5Si]"7䑷AθB <^\MCL0H);blXSfS(,e-WiB3p+-wp+z#{-Rݱ#[SOVOXNh$hjiFuAa̿(#Zi3"H3 yl.}F$}E+R~*ȤaЏ4~NSSԅJ@R JTPB Cq`.VRbcePaB$$b&L:a%HSSijRi~xMS%M(5n3ЂI3JKr/*]?#0$buR> đ:ʤit&x(H6=}Q2YtRF))!1r2X,r[Iv^,v3BwQ`#:c8k#4u(\ʰR~*f!z J>PXPRhz8;m~8D 嚝RunV70@*AYwWE).,.PmD=Yk c7ﴣOwnbnVŢ_ëZYDW*hUk[U#{luATк *J`IԑhxnF\@_鉪*sYq ryv3UUZ MxVUaWQSחZNH}VW wMSc=L4ۨ-5%L* ~WXo>5t*:+!+p6eB0N?nѻ5A t~womXwޭyYwk!/M0Un+' 0 :MQŻ3ּ=һh$ VkvNża'd Ӹu٥K t,|Hn۵TmԪ];pSM~<)UCxj,eC4=UM f&lA' B lΪyBuVo':V_I^FGͳ? N2l6* nSNO(Pӂ7BUc2;G^\32J37RbqQQE|fER=iF Qm(ȡQ[0쁲rꧠ7'&D5 =r.cIבr #M*c`UiZJ̸ fNq2NmN3GDSS2̹}UUHl\9JwG7oMtXtqN:|Bc}i4u$orf ޙHbtbm H N9OdYĩAFT0΢,EUHge?"9$v53iW˩5^VQ SkF,Rl2Lf1b 0yFR `8|S!!S# {W6_fwDw^R~ .Nd2U %( $&3SxZ V}Ql5aƊa!0+>` taYX.l3r$KYP 8q)2#P\?`?fхvl2Q<-*Q GpXV=0E+RƱc4c =2n( Ywb.:73 :聙 GUu]cl|'N; ;>cm sm:`FmW %fۦ]=>5]0X5=Oj%F M=B"vEN6SmDVZXsҰw& [nj1Z! gQbVt-; Ж]Mp kö!."ֺ>mz~xC{7Z2Pjd'cp'5'j:R DU7ep#c{s{xջ |TX#}& #a@w.6-.ڻ'}8z6C S >nQP,:݆\Ehzt[!S@1A hz\#BP@sNQ5z=L9J@1m{D:**;"v8K#e0=ca.O8F Mx-6K]Mn8eʦ#'3ccx?C~~ORug3W֒Ejŗ& 晃lA;?,a}}73ra4s̿&^湩vlg-y?,$}Xh4Rw<]7WD=?7=zuogbJ34KbH8 Fse20㈀B{Y{OSz]灗E)ڂ:rOvR_SɊJ' xhqNw|%K ' x0jt.[ȧKtx_Y,/S4zڏKL7={6Z 橖kbMu"N( DX<Za&<ޛ*EQM߯6gnjIt~}] {56èT#w8s#Tv@ZIl[iuU)j3v4yh jFR͔$2$ 8k:+OCD8;T(X]{*4,0P4g1FIb8iHC A2IC%fI-b빣iv]W )( |Yk;@)H7fRb= R n(`lJ2JI}Sa2 FcRߔ gRu|m0ȍ;yeׅ۟߾ȯ?KzSԲ:>sa0qTY'W&/;'-7׼P̲GɻOɧ4!xMO@]7 6Ilun(z4O'*c-'Q #^`7-BʓQNu,tz`2i%8T9J .%'87(dHkywkuNr NbAD&QGi>O=f'\QB,6zA&1YVl9_֟q43}DqކaN JCn*V;4>)dig Er<*h>0ˏե&;h2˷ 5,Z47Pz,cH@pt.,'DO Jg>)/$0 (@H1yD!*b"b m>L:F_[ ?u^YN;5?()pZBQpie, E"-Y,5.bT&11G)%8!So~Yďw~ĨUyMٟQ"}j,_dm7dbS0<SS d!BaE œ(tTJ{:icVjVtH~(J (:uO'(n0[ZDB)M& E"*TTŊ+ӐKC,%zfmhV;{^q+ر*N;sMYji|%Sʛ>J.r7pr˄'2>yO|`^\(fo~a&rXCv};YDŠ>}{ @ZcmZu㞎Z)ypt`\"3h%nYs"8:D ”H\U|Cqoašp H*ً"(qW XR(<̏<2uQTJZȫ\Uwg`vAsF!q7;HމV uػٓZXMГEnIY4l?sn}彶?B+J׳[N&[i{M KP:u?J9{<\rۡ+.Ye21j=jmZGbS%{ہ/)5'nx^L&nsC$P>} OZBwQ/L >9zϻQ(gʈz`/MM}_sTSw../:G ^Q5țD/ jq󏿧g@h +<n*'Z̩, 3" tgر8c=Nq C#J JNX`4)0u*Ti (A WRudHب-I$gT18J1io` |”FXVb`>ڮf42jVB'Z̭:Ei>Yw*/O|$哙E@~ӘҌCWz˂7zxY{~ZNa@NJ}'5qߙ'#J!B|箺tsYztg9]n=p!%U&tὐwx=ujt6od:y19o9Q9-[s;Ӳ{dl-*9][|Ќz7동-o٪f6_tL3DRu %[gX5bœ?nǵJE*JTt)tȶm-*mxΧsЙ-ktS=}`_hܭN':DP؄N0N҈ PZhJ摔hx4b-m9v 63VhǤ)KB1YtC)d0QJsn37J wC)Wʆ4 PF&2CSJhG BDG(JHB0% 1JSO4`jcʤa:fg|JcK W&JKIMHǥ)KM%F)n(ͪ Ō]j>I-9J,.%b(=&MYj7JpC)vTǤٓZ'JsC)e\%R?ͣ+wzS"'\ \Q53sZy)*5cS]SE9]`~}oQ]WPGV$:K lN"XX>\M 9./1NQJ/.Nf:ga,.zM ;hE n]M5gr^,#2r@mL6aڱwc:Q14'԰L1.ݻ@B wܸf"*XA>7]0Uy/tytDc4qÈh4Tk)62$28!nU۽ऺw} %p*ܣt!jpa)C(I f 4! c8eB ({Ztyk[۲eHb}ɬUY"yUm)82CS[uu V_رZmV_?}y/ZY8\wM-}zt+j Ky̿j'zl<7?#6 x4(UoX Tm4hHp{c', 1 1+IFZ\D_v/,o'X@Wa`vꈶ|5ϹBͻ?z]jvv-q Ӽ?áUJȎTѧKMs#YHWhADU|xۊr6U&w5чϙWl7[1[JOͦbuM ﮞKb{y~_n:n^֦FnG$=[9sFaJg wՁtbݦ\EP}Gghӻ3>9]W ~ΧZh;(o"9fqH+?^I K?,JŖo&GCi%50q]6Jgr"f=IP@]zI l,u0^J2-u. HXb n ՅZH)eiFVD_6s 13GcVyZbo~_E ]7kIc#.ަv/B\kwci}f'7K~Qt؈F>#&f63H[Ir0t9v8FNi'܈Sz My5CY/2dvL݀JMJZܔ jP)~Dhin8L_r?Yeos_'Z/,D%ژ\լ #1^\Xm{Jpjs8N*{s@ ?%NDD+(yVBHF9"=H^@%|S`9$F$WY3F̕e@,KMb,efjƤũJPA)ɘ\%baT*EA"-K.bjYkLyeXrFƊ j&b'zicVa*{'uFCauȍүd6A%XF#x󀜻'B+L= iE#ͿHR8iMIu|&W!1`| OUt]R*S)םmu0hAaY_AťRAIP(BUGhw#z2@^D[>"ty]`lgu@}Ԡa;V|bŽ5W6EGcNJ<ٱ@șC4 S8Hr݈ծLbuox k-$w!g(LZ5c?U'y.~pǼx+?u)~v9/Rό;(=9tZͥ//ng7|Å^1cMJlϦ:l&|DOEyí(<:_澼I?vWM+gn٣n:gI;6O}N/w4v/| NM^?!NǏ_>n,®ŠzNgO*_ԛwy&|z" 4dHSL"ҧ%Y\C§1JGͩ Hh'ǀ̴У; yI|$Y^5&L) 7&Q$l*pM]>~ʦseve36Ϲ\/C|504j,Njmmcq6 k!zIdMJrt1(ܰS^D_%X5ʦYeR$gI4ωs)JZk*2W ʂRF]P U|+)&ۿ⥻Dk_Q:DHZAD%,L)FQjcmdR*Y13ITjaW$Jh͟W?Ca5ݪ~So=ؠis=Ӽ(/ww~۪w 1SLXq4TN[ [+c17>M#3,IE[ST,m-- H( O8 Z-lN#ZMqiªHYQLNC"YP5} 4>m5lӝȄaJ1j4L}]5r aL_k/V~ {d>Tr%53R/VNI h!tB~>GhP/>ãRї) /x&vLI虴X7 `x@Kr9'O9sFaF:H}40 HxqE~#~q#(?KBQ0лaɄ-V>ЉFwrQѼ[|OB LyGRO/(dTOk6:O60%.ltۂջgk~7PĐgkG`Tm| gl?}j :`Op[/JhTYi %rZH)eiFvqlkwuMݞbwwJu$OK= --JS? V#Z+Ρ*PLY-ک&_;' 4zpm%Q,)eJ63!DLܽHC& KR(p{YRt%%R$%2.7rH&h*Hd]d%+4LS_%O_O[ء9v{}{LɴWrwl(CKE 缋 S(m޷~)%QʵO]=Z簋zgfu|Fj?ʚ>$k]Y?~dOs>ٷǁq?_Op2|kf* ifnj:onuwړ*_DfTm&#gSRR#KZl Rt!KOInJms!t(Ҭ\g*@)Mj7#/(]"JP|y.a>zz(}AZ]"QʍJT׻=?VRETxTjCH65c1xAuCRbv뢇½v<1Qh/pm%ͥ}yz% o}^ވGr/i t|& m;gs+MhQr@MO~H\ ͅql:H1NG_s=Wҭ&ct[Hn9 ? Z|\[q-xoI6Ls&Šӊqc4}tnwT]㪇K_NjBȧZèm U}q|~04X[~:J_<~frSUVJ>b0ŵjĮI_5izIOў=Zp=&;V<.)>k_ H΢ժYNإشbSJ]z鰬q˭]͋q6ֵCf?p]ӾO,_S% N g3#֘|nh`B1GH FOs^oݰЍф)^ k1-#mk/(`6ʏ.wZARG8IK’RG?iN~QTK7tiGG$iy Vȴ 2dX|K+|e'NjQETp?H2z(ԠkB.(]"Jꠛ@)M]|Q *"I3WR.H~(E*TH@)Mեd kJ*R^3ZuSj@y^6JCR,P*J+A_PDJR(E@)MelC)8s-`PZIM.yQ+i9bݔ}L~리jXztm4[s DS>#f!5c66J;Q‡phdrr5 r6ImGO<bTko>_a8>Eu㢂_C\j]~HήE&nJIn`Y&',Uj_1dex]j =!Q̍\HtyfUcrFT~fbP<q.,c.xAVI0! ,hl;ec0=/bt^ |pE[FEda5Ԡ219(pBtM4wJHy?2@{SR~|h,ڑAvIqsw'tH/A>!]WA**Kj(%L>(CcP:&˕BTr csfUTf]^*_P<2( ZF3,`ja#$VJ)Ѝq:'lB6z)]UZq̠ U(,PFKuMhHP7%oО#R1q@٠ N'SPU0L5sr= f"%CcjȌ1i{ xiS xb {ab s_nn-䡝+if0OM#Aq4Yadv YlA#fJzˣߞ>lԒs3n3-8ymJ۳T%`xV w/yBa h&ރ)R*NDCR|/[œtw\B"kK]Rؼ' Xёptl _ NZtgj:$Uz5zr) HU`ԉHIq$:I:ujpk \ʄNPpܓ:,1G+©5*kbp'Y$t`(! PjV ? .eʻVa+>j>{wYgc(kN^ws6dUKɣͥ\C}ѐt=fs0pΘ@ JԠ%i8hBxRFTN4cED .$Dk*Ck*yN&,l)͹P|UѤu&U\93ڝ9XdL 7 ˮ)S4@Aϗ4WDN"a%T,]tc2Z cpvDY)̫E"1$hɇ+?{RL”ݫ߳M E낦zV|߶ VM>Ίȹ A^`@! Tۜӏ+8\f46\;O:GU}i4M9r\ѓ*Ɗή;<m3uP?_"Io;q|/'Ojø̤ZwH' "W1 XHD.Z]$WvLb3áČ3h I^>,.+.iWizhkY&\,6VFׇEnvfm>L]ٻJ@CKުm-o.04ӿw>y=?r$s9hd eA˔=!'Md۞w;+o}tjӠ91N)"](f$!|Q)F(8K(u`gơ@Hgc]'P_򆗛Rt/"Ţ`~{yHD{ĕ1jQZkHPh² Tg{U)-Ae\ZZy*m# S0r.UD ƕ=C)̹Rձ0f {>_/nsc"?_Ua\ne~ P_/mncO"~l/025y77qfj1Ǭj~+# >!'9$a Wr)(RLay{ca{=O"`9\ gô8(Y[77}Y/Vwemf Wç&=xf#0!R&4Fj Z!FY6bݫG+IJ2 lO+H ſ{>e{Y~r4m4@n6Mr/3eXT 5S1J 啄<` $- /6%2B&|c 㪑od5mȒFIu@mDtѴYD0I'k >ă OXxH!2=IAdv&#@,t}@3<_9 hnDYKXډ=Wv"H 6{rJ&Xn*ȥbikAqacz7ytU,`5zr) `u܀>}zs3? %fO{b{nڐӪ:|rkU_qg|]q鳿} *Co䴌߮6i_ m\sby<>DU=ג|˭ 7iCUqT{]iJrǸrHC޹)y"ǿX[_NwX eSv0bƶn UNYsS놝#Z<(Q僧n(bKBEܑ [o m]cuL6u}A}l`evP gTj&NzYVxQ(saa{~V#*$p~7=KUy{PaMNu؇uwIoވn1[bi ot)OII.)-Oh:5B}wGX8]36X/.{VXp r͡EDZb.x(c qΑolQ5<]⢣ϪO./|~V1,3I_VW MKt׸sJ#׶9Fj2!-~ mtH1$$]V/J6bSeYՎQ~4*KSҫZ9|~1C 0ZZe̗.F*lTm*3P Vy{ |)K0 8>eMj]N7}r(Ҝi!h KRr7#ҀFty}Y-!|Sv{1%>{D%`MvvC'uS ? t@0ʀ ]^@f4f]Odg)vW둮N:"[#8t`+IьB.Q!!BZ4L=HG:[1t`ʋ݋CzT7jR#?֬-B'Eg.@ʳ^n6y~x6Qwc4z{' NqZ`"í֎^H>.n5q0τ[mkz Ub QMF>2q,6!gNN1$ًn;JKFD DIFTt[nv`y>[#6CΜIrj3fQ!b N7bےFү  9s8BNYWvR;[=znc]E/pshfgr@e%Ek,uP7^XBm}8e}ʚ˧w޷?7 D^x͛ewo?o?ioݧ OMKvk(~꛾tI_r~VRlU7}KymT[%0]%Y*z*n$P+; d)Lt 6KY8s;3*X_Xe;s Xz꛾d \Xi$XA-*X*`yҍȗ!^KXZMp7R܎HY8V3:aYJ̯KO[}ӷDme,ݎ֬e,m~V-TK*XՆ"mpo\YVRN˼t,ճkkWRFKA]nl,\=q[Ty2ESVeZ^5.k9F }7-a~M8ϋfK|[`r o~4%|NЩ0l(s|fU.>`,So54-p璄=1Ruz 9êzU m02.X(m*4Qɉe#cd3 }Fz"|BÃXٽx،Q>W:+,J_iɒ`lQB͞wzwW* m_-:g߶=g& H'8 Ϟum/[u N?m&Lӌ) FbŬs_u :^^7X_yC[F'Z)8X5$vNK,z Er-4op'xye ^VurTQ!ՙ-zu t/2Yz}KD2M/Ƚ՚ r#4w֫xV9XcBUg.[j*"+L:gUzY,Y$dƦ:=(R`7cJuG({!M[9W *|!>C+K Ur}e3Zht/*IxM2aDUafWt͙P vdoBWeq3A`/&"8oRVZnT]u+@(X>OA+`67Tժ( SrF'2:\sw*/+Pbx•BzV5z"qAJ[Wk%LJw3FASL`QXu V:=ΐF:?~&P*>Ԧ-D=o{:<x`¸`ć) Ch>Hh4 2񡇰ݷЅh{a#Xw CѼi03htIm[)\m Tԥ)u] IB:4{}8E6d\}8YaiԚύcD]y(;"^ޅݎ U2{$ ppd1oVxԱ 1ׅAxi ڋoiNDŷ6xh| o5\|t @OT)Yzsc#!tZ>fT90ٺ4nhF ÓH&/KFAofX4[К&*/R@LZ%. h[gg` r FP&zx& 8o!*i0Xٽx،Q4<ەD 8cXjtYhpZB2~9):GDxvޏ1%a.8KJ_;ӎCխZ^/yDb_ݿ0D*V(|rZ{m>Q{S%#@ í&o;ozVsSpj%\Te;RR%j#0))6 nH싕Dym 9E8z3ν>gR1P'1mUDO7"03`NYb3dld %#C?v]{́$\K8yO~2 ъ#-7sa;DA`\\TعK}ʏS=j}qPONn57N0jKO=- !L?1W !}.F`șS4SGƺ `tK@BoDg8ABzat!gN$N1~ƹݸLY2%b N7ۂfgnɥ֤m9Es Kލԛ-MBW s ԝ^!&,K18DٹcraicFxԄ͗b)ynoYcWĄ<7UB%r^'W)Q 84M ȋd:a k5pBp4ijOK!/,d/ ИFlk[VKr2ɢld]HJeY#9$?`9-L'Kd)-졓ay b ׄIRqlFcqQS7[d7aޅ~4 ~[e w&|CIƓn}v12!s4uǧi_kJ-ci҉H;,w`ӻ^OHL5W_)J.HӯuT(;PY.Yۊhܛa~,MnHMdtؙ3}<ڔ3 ?.W%Qs.뻛O d kSҚ3cLELY]V( 0"/LLUcaMF=kx|^O?{qL%C)%Нvӑt+?̗%gn.yj7$&zX藣 YcvwvKaKwLOc$ k_\;7P1tƄ#:CC^g_Z3~S+\ZDb%=5\8#WO{a|s6_哅1 q3q|KHf/b`!)&GK'hv2ٚ@ .VDqRsarpr1ĥ-9R|X xQGM65V/>=.m4n N^Ife18|SOPd$i'HiaWЩ$b!'2dӅfQ 2d\Cu03L2-&y.96&a.mH ȚePss1 dǣCflA7cB*{m׽!Dw[O>F:Xȳ`/?{5?Mi>yW7^n>@ć Kqsx~?nÐRLO,lx")"iNz~_7QXwI8yѽدG=w x4 EJa*6k5v啰pq@b̻.'bms>OBooD,P$IGa>*_1/Z5 a[FLk2\y>|;2jtLYӬ{Nۣ DТ7n]am6m$.͡č$$YJVlqF)&ahσwD 搪ї Afr=3PB6-EXCB516y,#wD 2{q@q /+m)q9^.xIFi@L[urE[D[FQgD#z ))Fz 9_:h۠.pʡ [CDJG(gD(tts,T&aK1O*iyDžOB;\ra3eCv(^.} mb_$@0/51CiGqba[殾ީ eQ +%d1Vev]i[éFO3ޅJ`pQ```# C@!X^.f@ev6e-U L6ޒePoJ߈w=m)" :W~XKI;bQJ[E! (x&1g;1& (.2_~a`T9 R bTNB_Q% `;l(j9ㆹ؈ Pp =h2{q@ x/0 ŀ9mIGEv.8ol>Hq.-ސjGR-3uޔe~_mMDC`< "K XB)D2#ns 0֘^A̕282&)SScԾB*L[[STv,cc%Hu\^pH{(砖Ckr4,4i4ь࠳35H/hjo۸m*/ŖiV-豒c3=Z @% y-{!,=PF$AZ?c~Yy;0>.H5h/,uHXΟ1[-xܿ\_F6~8f<ٷ2: ’z*}K*Am_U=7b!ݕu8}d&uX>*:$=Za)Fy",N<8R{V-R' 5-t"V '|J<^KJ] h*y{2,t\OvZ ^ri?坡}t)@s|3wӱz-XO#D^H7؋k\Xmku,Qc-=,i&yx8*,  g* XAPt^y3c$$DN"ssA,5S׎2f֖ȽCG>#,!bI5}Q#Yӗ>%B7޵0m$R]6YQu:-;s"e'u3#rM v*-,g_ϣgLzf_. C~-茔5b /xW}߈<Ǥvg/}?7k@D߄rGT-ХƇxqz{ׂ´,-Ms^x ăc ~nBEtޒN~9o[y`s-`O_$BkK&w6}&V@h\z_]yy- zAjwNwXIS󫋩_y 4~]y*I`rznӱrF,hDa]%k!y H( el5)V"cd +BS-mu UFU|l;^[:U6c󞹊*iI͓*KjP [TBۈ0ACЄH0]81H"!1mr,Cr5ttK)υ&Y>Og?t֟)"2`r0t)a pHC8"(- 24LZÄ yPP^Kj@YP&{AzJ.VtK}/oV\iN&,>^LD%09tBFk_.~VMS**aGRtj6%.{ 5yUBmwobq 6PZjCKAo8}Ow_ߋ }~ˣΧرS ":Œ+l? hY8&`}0 X5MϤ)ӫO;^>1~x>gU<5@ǦՏI\㿽]ywWiJc_W4Af2v0SӶwO6)mN隖kYpw:CۆRgiwhv}̀ppTP12 l(4Bn  0! C8P+‰,$ϻiq:a0h~<~ֿ4Gĵ:VCw ?7ӹw|'ihp=(a.Bn3Haf[c`?./^\n0΀LDžOY.tI{ zc!eu۟}ݛq!`nSso_BMǩJp-ߏ~~ׯ~voA{dP4#5k &lfcN4MMrMΛQ隑lrl D-8#:J3WRKV8u*]Xԓ?$%)N_lۼEL ׿پ7q|m ԥ'~%T&s0/͔ZB5[ƥ["7pv 4{#`vкz\/z+3681|Ij E}G8#O"ws|5 '0EYv&cߴ:/z^Uގ:4~ofPrnm;lI?^B;byY2|W&?m#;`w_iॣm6e B}mbH2tR?FM~Y_Bm$ş/:Ɵ`Zx3UNtqhɪlVHkZ)v/2v*޽tzj#cO@lf' pj p&#D[A$)wȯ3ǵx/q1*> jG(\VfgCgŢGa&#O $?mhsԃ5Ͼ':9 xh޲0/1e9ݐ;V-3cRFn=?rANP4㾸E:, OK2kWN 0b('l#+uJ ƈg<Oj;E9։qR7̞Yq-)UQR>]F7ε6xZ8bzJ_ki\7Ptt,e~l/ʦ) U,`|9M`CRK0=᳸?bb֩d~#փm;&K~)/_RBS62!ZQIF(0ȑJE#A+i(8 t ZݿM hd|5jWqTc8{ ^.k9 6UiLr\>236M+dw;_$vs[8n CtJ{}.-`ݠ~(©]2Qo^fM٥`Pjx_f- ώZ0%܄ްohM/_a- Z7(b$SV??~o 3!JjH=4m&,_/Q'L1jIyV"3a w9z"d7Qq)JsCv-druoCL= 0Ra-$th iYV:CVH:ELHR_.~[/_W.rUHIIłRaOdP֘ϤJJ|~qК[T{U6Oa:v01CY= (\w&"(8*`JKa$ڋ '2Kz߷l:K)N)cAݻ"rM0%et\v3F?8+lAlЀ%v܅Ѳq 6ϱXR}k&cQ='?QWhFۃwo*LWxӿRq~I䟡4RD #"rK-7i0 1L)qJ"TX)" m䟅v:B=ɨYP`' BVY"|Is6GɈ01X99jZ 32^ϟ=;Oj vX搙HR`c jB{EywR3 rFؐ:D* pFX"鐁NI% |m uZK"F /K^,?VqBaa7J!pk{Hm- PeY>KȜǹ.H2JN"QlA(ܯj-Pv 佝d еU݉8H#al$"kflKu\LJ8T 9}N_Lnb" #R*sQ޿{~Ts`86Gꟷ\GVƫѮhl1&KC C(AD:L(lR(5zZ$CI58b\eL(me0SHK܀se"̂%B\aAԊi )7$ 2,Gq,0\"!x=NʋXYS/q v(w& Sl@1"D XF(lY@5 7YF~B1o$7MgC0nkh T(Gd8}#hg*.%R8ԩT0&E٘քF>)Rnjт܏U&9G"򌢣!咜c;LԱ09&bI70Ml#H5[H( aW,>bYHOƻu]=ԱqB3ݛ.RAP j) Lu 3m]S>0سT*O}s7_2u5u쭂yLɲv/:atO(2i&}Q{Mz6h1$jl]=F6 FIzR1yyqn˓uyyxyU;PBz/s0NHrZ9ʝ'(ˤ8D1 ӔVxG Ńj5<(pYEް`qjrZ=hCVIjA }Kއ'NUeMA1Ubyz0G#Ub@%1< _lE,p* 2r*@p(u(`NHauΰp35H $okF&8|dEqT!Z8#ۘ-wm'@ZC7r 8'Ŗ3,ْԒn5c$_f*$ zwWew]ZZ1a7h_rVlNJCW/i%<Ҋ @tIZeC)g!)n}Cp d]=THN[WmwqwG9*.Gy{MŃ-;%1)sF9䊊G |XA?d%mY7?fzlaxfPMF^'̥;_q~bdBX 'pWMhɩ 2cq~ƱbO(F5)QI<|!$Rx-Δ B@ܡ#ͤN  )2l9͜2lNP T+x | PG{ e!D,rR-? ROTiEC5R [KRdNz`SHDW*X932Bw:OE,# C_(AjL¦#XO ,ὴ-K¬D# _>JH0:to\C@ R YqMYE0f-5kLtUdSPG5.9DZaPM("#z]V SV`JORYo 7_fno3 39y1|{}N62響gPyd>7!>?G%%Koel]LzhLbO2?0V.=~~p6=|Z5tIxgxC\߸ Kt~^]G QAy6yp[N_fN_Nz.ճ-OwﮮL.tpष@A:A)'(R!@!L]-FMwDu˦Hj\@r- "ќX! B-#NiTA (C!KnFAJ +W*de|WzAX }Kcop?'L$wMLjwwo^]ѣ]f7{7v7zWuݫ{ot];攧 ݍ~hiiKr~>QvYT}L{`Fe;{K`*Fߏ_u70(%%j{{z\D;TY0nS]Gn]1hݎhDjr0yڭ EtoB J\Ꚗ3–i2'Lj>! e@o.gȉMܟ`EPP}R څrr3fQO)9}OӃ :f ߝ4$N a@BњqFفDy7љ^.GEOhnnf\MѤxuK~}~ձηtR\2HG$f%ڬcŬ j_&_}ӏf3ﯯto_o|,z]x5? "_:QplƒՄ1V?SYԁ|+{TW¸f+tYZEkJњq]x~L)hmaK S ,/S9Plyc8shBr^XxeXz^Y٤D4FŁw,=i+h:~޼PLr=pyՒ+Z$ddO7 jMxjg)Jy'7Jdz<tKr㜑I%+Յz?X~Ca":#}S4*O$B4Ɠ0WH臘[\'Ϥ3?#TޚL/:,e@Ɠ~f~Fjk6F*R[^BN.~y\yvONGhJT-8]ќWBcuX}z}|DOGP%k\G pk}b}6EXEcR 7~󓛇ͫo,.81%?zVo)|]NV&V~&np.)*7(oo byw }5LjӴ߼z. Śe,q%0J rW\ LKJЮb4 p9Gμ:kQ|ʽX~LX寈3Rބvua5Àq6~%DsrB0ΑPypF^Զf5Hi\Q@_N>^\^&흴& P`ƥ%urw|^] jVǛ_Q )@ua~96j=ݫIʢEۻf|73ϸBu„_GK,^fAN ^AM3EU7/B <핣Nuo8t4RJE*4HE\ q $~Sr U" ^RPT v>o7Ƶ/ 6dJ)LY+5ގT[nSc@R sjhޕМSCTUOH &LUXK\!'eţa?Ohd.HNh,MHڝf577_J:l&<]2#0Z] (it8.-FQ&kυMGA׵ *"D>MH?6!@4/h(]>bi>o^8Ek Po|jN=iLRu4҇"z#{9̒ࠔF1p @wʘT r%8H =E#|3 8}U3d, S36%%%3 D lڤyLGҊ2jdi93P4|ף SN8Rh 9 %ˀuW͌Xd wf-5gSWݳk N<5-"YLg#X2Fʎ "*LV 3^ 1FPRIP*BA)sAbL o|@Fs)_*J j%Ȕ귢w pzNd:#Ed0.^DF[| Ppc1+ njMd≄Zc9ASv כ$}JV]B$2d_Vo,/8\̑8-&!I*N~#O] 甍0#և~#,EF]ÇMlBkMO_dGΟ>83%=Ga#l #GeB¦L1x"m"MѪzXx2ohkeP2R$ʒ1l6p(Pҷ؁\$RҚn9 v7=_ /V\γ09Qf"/h<#,)&9"Zږy"F+/J/e$> 뮼A=۶&Ȋ"fG\=>^G 9N4 D`ȦALwDV)9LEB.!fM7ekyaP6LfVi(2vQvj`%3ϓR#Aq:XV~ғR)b*d(:h24]PѼZȌh2*!{J[i?jS/=*5E!eIV`/ndҢq qcٝ˨JٔžF?Q~|OXH_xde*Bki (RT>:_waUs#>DVcEw3/F=}fhqJwef;ȨL/!)a(Vp|VjjM}C\Iўs^n |AEoҡLnP#::t49 d*lD %i]Ճom fz?=ffͩwyiJQ+M kaݶ!\p)u*A(ڍtd-}~'K9b( 4+%LB8dCֆ޲SR$>U)?{ܸJ/gڸ_\쌓:UI&d@kG\zqP)_(hpBs2j_ G}/*zvpyxoMzaN.=q6=>3%j;TqXjb}Z%28 ߨ,s9 @d5e~{d=ܱ<G'22lxR X<!95.I<C ٗ$N@||f ͝`^< sO]dxoLJ͜IYM2oj哻-4f2R?͢!.MYJ?=7B Hba'n[ŐZ)l2up|tNF KSƢoǻ]w7O7l޼[0,4 ї`<|uPڮӚ cB@@f|L,F !h .<+igŗ3 ]=HGe` Օ@OIT9 xu ̵ 9.Iֈ ՛6VMiǢVDT\ءXEHaN9h8aj֠y[M=*^BԧI Oc@PGI ;7Po x[Yo滛_}nzZ=t8ξ%S3][EpzЁ!8k;T SQYbॠυzdv qlޖv%hRڱ<萫̶A|6^%lmG\.?hC;F%mA8o-i$u(]@Yy@n_)]] vRAR ?K :+؅@F:_NW; 0F+;l-$u'߭e?]숞~Q*%E?hpU)X(̰)]J`4Ǻ(%&igW^%-8%eJa=s)UYQmvm[э3r;mpC*9ڛzTQMšWTHyFK0JE|K04lZ- (;S)S Co@3aXn AA1Y- T z8D[%Dbi!vj1G9d56/#YKGE ->{ &d% A<ΠE4۬NA1Q༵*#3o/ZnU*NuTzJ zU]U)Ba-UK;a;# HR%4 g () K 7ƦvO{RPH ՅxPΦ[3 vg%81iS:I EZN3&S SY*040URk/b>\j8EVMĐ KTf@FISB˸q `5O ɚԙIy9\[r YCxU!]SYx! N 5EY-L ives"Q2߹1ít3&pva (noߛ'\}j5~;6(Ѱ"m?fiCDnQ+e@1Ƙf 9B*Qg0%,0{m8{7 b"$ ho^*7aqpՇ@ѷ359%6,wGgV^UfAzBw._i kS~z8g:~s;[o2):P cBSoaBJC T .p9dIXj져";$ +8h]Eպ: 15UpH[cx?Udm,t;Ǔ/E1%[?iGfAD}h͂Bn| ѻ:Oavd"R[L#Om7N}l7y&_'2fD魬D\&Wy@JUiZ\ŨR*bcZLƮB*]Z4 \* eyi3=gIL?*DJFHT!͌UY8d֦0yo`2)]]sWsRؚ\۶Wh!KH^-\ '%s[|xL&nć@3ӻ8}7^fǏ,OұC@~Qsó[W?3jSRȒw?XR6+<6??,yqrP9ғa2=xUOK\)gVNWp$StW4͇Iq[U BD;h#o=1Vyڭ y"ZEDD)V9åLQlbD8\P+- 1i .?YJ( X -qЉ;]Օ1B҉,1G5(/$ǸXNq\oÜܚc^UBK9rn6Hɪ "d2d xDlkz:l5Yp$S$ l89 rdmyMT6cR$p}^S̶bƶ ꐐ.d5js)❵p斄nW3/~N#oY*JL*ABR$T)d(% o w<5?b3QJQk$IA,!Dp vo'!/ϓI3q%OA+c{X4ɀcnT ')`dsWM0 tS+/rއ_xljnL ש1Qr xbLa7ǔ.z J6=x kv{Mt4:{SI0ъa\j/~o^;M!GX #}͋b;NhuȕeCY3wܩĕ=NჀ͚` k 3&"̒q@Q0`"`a/;F %-赒7c37c gBMeNBDO ~ʉ8nw5ndr\6=!3ʌI°HD;R Fb9Qd %9+dh"}N ,y`pN؎<9c?h,;󜶱5xC˨<'dԨ0.ξhNr^FÝАΫ؊?%Qxn% Șs :GT⍌ι+S;I|c%r_@Ju+b8,T,M( }=cs0ʌQ>ÐB݇(vWl[}Qئ1;@n>G0EYC !sz_` 592v嗒.0C^b):aH]_Bʟ=ԼBl~ 6bW嘜M~m_f_|V*/S38|x3>}qqQeNreEa'BB*r ]@zA؍Tc k쬵&d2NݫsLviέzL'/++_E?&Q?xӢE1bŠ(Г88a9w# uì%Pbn3&rHHf`+U7y's/-ZC5b?z[nU~ђKl`~0(?LJxpb`/Nr9~=D|tb$ Է4Ӭ{f8f\ uLNݓe&8QHfH,N0!OaCo 74́2x+sP2ja)%ƷN L;`[@q_R3 a L/hBHov)ᄎZn c6/nepy^xQv@r ^0*kZ mZJ-SDf0T!*L@G!eӓ5wE9,uŹf0NRbRFVfY R0 CF'u{4{>N|Ns{:qErp Pf Ag>::~87TS$W{O>܌:Աo`n׽"}?grP)}~|Cz5Q[/OtZHݷS-k/^Dzs$rY6jO8*Q:$B,Q FsmZ {j '_9~c^w{Ru ,'-n? R/!˝C{0qnl:w85y}g3tkP.얒IHQ-d NUqymMdj"/wDP𤵹#fjvۉR*XIVI HCBqBʜ򷩖HAE%Sq G1 %h85_{Ĩf:qz}\l9rQcX9US(noL,:( $aR Σs ~^ga0R3s f΅xMwXeU,{_cS(r9ݼ~/y ex݃S %?WnN8J`p+~foXg ~~w=-${w $Sڟ$ބ? -NY"(#mT0kcJA`)3 ($؋ KQVfPJt`&3zNصY:lמ]Q )jڼ1][sDZ+(+U~HYLuXeEñD"yPO$ g1 ֍`gou,NvEITƲkІzܤO^=W,]WvrFr@7`#_C~ZfR;0vQJxt ¼@|S_p ƪ≾/_8dVhU?ۻ6& UtUõ彞v!L, Q fEU(nxJ9$ x(]r6kkM0RʁϜor^!;[8No6|;ĆNJk:ر؁0IiN -s~.F]Wf>aM׋ԉd0zL~;KQؙ JNplgb Cڲc*ӈjE$O3Cb45rS LD8.\)DEZPswC%"`=6Xpَ"rktjnO#7,MK9+ErC9rm>.zkt"LV> iXGAG2^՛QBތ>G𚫄sIGӻeÖkJ/^+b|d5@ܞZCXHkL=ZwpxӼN,s~.?8\ÛW7\uDw37[o[/i}{}ݦcu4f W^q4_fr#w3,֊(-8˯o\ >zgssWhq^Vݜs;!}6<bFs1:ngUvس'ّY\Std1-<XJH}٦4Mn;oL6HlZGŹp@nmܥzٟ8DZO\Z߯}/.ڞȷFjw)w7~Z ΀x$u瑲{I4Oٺ <qXvt۩lS9WНr]`E-m%RKQ}٦:k)j)iZJD}t՗mPgRM9;)o;ZΑG F1TqԐCN)Ķf*jV=qMbm1j'gS^[%ݏjT2u >v/T3FvikiX`uZ32ʗk0*q;MNLlrA8ʻN]餜@ntF?f4tߛ49׬a#FrNXqF߼~Ff:zw@@ @}o>5eVq)S{B:/T>reC0ړlΧ.z~T25]cw>'$3ctutI5!TI׉a7]v3.bͦ&HfRej:C?X>7?EvcӋ"BQQQM5nvhEVx'S)ZP`zYPpkOHƻ2n#zӈbqI@,N PgJq'iяwz$wz>zp)'n#[x,bH i>nU i H}P3h$9, *vDE߹)QёJ|#x؄P>a3ClgQ-lgB1EFfwAwdͮ3ȢZC;ːސ銚=  ~/lSM(\!)\pCB5Wȷ+!_Oݍͣn4 Wѝu* @갻(ȧE+)-_܌,-AY6#LMV4DXL-њj*Z(iZkhDQh)iZ:Z3L)kLl Mm@ M\IMR-4k)NygI+\-/'B.C]hq%Dn뀲vS/'dq ,ΰOD͝*_%Y4р|kҸsh^N~oc7U-H|OfmxWmkh aIh LPJH /r:Ӄ-I+iӫ:@n_o(j<)9+mLk+QA N8Q0͍o1movo̥_Sb 'rɕ W| Xhw9XQԋZ7_.o>o`ܚ*; ћEjtYY08%& j5}}V/5XUלQ׮Ҩ28 ) lhC7:^TfRӷ}1H$3yZwBr|0DI١İ4D3= i!XF~؀N<Zs 4ut3機|6$>(8*8*8*ظ`KQ"=&IkbF9Yk$W0P iYuyS3~1gb24N.F:2X(6]#s_T{IlI0a#v`㬆c0'z If+e rJ¤xO,WVX&Y&K18/, 0\"LBR+c< E*<9K,ET fV?f)%-xJIuxJ&/;J^O@q+x&qUD)R4*cn5$֛sGX(ĤrGl=YZrZy+Jhccp:,1z&g_hu EF֘ ;:vw\wg9Պ ضqw&&4a;J;W(]"!8btW+2-4fy'N,+$H` C3 BrWj%js$XWOH?߾[>t{&Vh$vt˞M#v$7{HLHYN&⛋E;, Ǹ8RƊ&c.CWd .VUa4պ&ٵd4?vgw5$BLܸ9I8k7"p'7NԆw|wuS yWKR*8oaCUGH.9] pr8igM&bIObJcvh'&KKDB6I~^P .X&i5$7*C;̱mItd0ՑqyD-qb9\{0^6HWL*8'/llꈓrjVJDo@^QN7 (+ Ħ;WGI˵:FC^"ΙjB@ Sq`{xgpRZX(!LUDz_#I/ޙRy=n`ضD%{ $fRŢj4 Q$+#82"#2Fy@$G/'Z7hHbb~hcDCLzN+w@H! <**o-  iM}fI~ 9߬W,ךH7K.cn|j /gLno*V:yNLInx|Yn\rR+A>} Pwⱄ&850 ^(Շ_kZ]ڊ<_6KkĖ4=Rl5 yu!MםKJg e- M37ߙ`>q榓y3 5QlلXʸjSH0Oc#@)plDWu َ5eM (8k`QBlFFI1[if䚨FLN"+x@|Ves63[ػ72=4X] ZE0zLzV(sO] GI6b LBLn_Q$qsOJF( k)xrZ,AjGMV7ZFzs_hcoTD)N@~WBtƠ1J:t 7wˬX+ +t HH%"lZ`{cPZV,>ua˖;/ૄMyئNa|rGV.EzꞀHD@/28@ISh-N`w5MMծAGA/v6  oCy]dV7*u UUBu~FIwˆXLR. >7^+@gD@[5M'z˹X1neZp+C9C7OOJF( mJ+R@7#,H׌yϣ R->fx~P{Vn`r2wty°I}X ~ow2ր(44_qZs"ŀ6Pu537t%RĨ+Tq<ҖWQ :rZ ?g/"|?xYC+$$pO mlvtCc%BA궼 D1h;c _M/,ҏ0扊-cxy$쳯5#iz8fw?~㓟w#g]^G0?{?}j>88h4Z!\ G"G/蔺h%z#cD MAUVJH\ %kheIF=ysz~ej*W 5UfP۪hE )fcH<2iTZEL 6|!e5%[wsj9"f[fu=zp*iz<*ܬk( x:mHN<@]S@!ل#P-pd(x.G!BMnU ;u, k҂kG{~5]9#~ _ bpܩgZ;4X>ٕjx## cx;>,*%m>i0In[379EJik"9Q $2PP 5B4  B Fd|Bb;[K8&15-Ry2Ji<=2f!hzS)CĆIwq`Rb78%юgʈ nX+,7 H>@bCĤ;ĨZ/mEYyrVJFy(J"iOmBQJ?]/O<\'ibAI $_B>Gfiw~%%4G?fo>0[<~\48L&$oӏߌGq2o|ryz4NjD;f6r2;+->usy> |<#)`=8xZQI H#\`b&^IV&}a.Rn&ː&y9^"{k8?"\ċ$w/B2|;Uz}$G'9$dwgFM:[#H14oRGa1Z "2_9ڠMgo$_eQ9mxD^6%6ހZM-&jLCKѥGIwbXmy Ihô] p0ɉRqTɔH#*0+#V.9fI{%?5s\Gm*Tă/urs\!e%䃐/wf۱FhE\bk'A׌ S&Ъb]^ }d:ŸT@[+)l5X^SU/+ ']rPz|Iۻd x"1zs92ƒHBDt J'8`e(잽I"׭ȝ5]"S_$3 :uloS Dw"ƟuIK7ffڌF(RWĒM:{%&Z=U({ޔƦ֪{^r]+UmQĒc C5gu7. sv=s5.W~xW/V7 ֑"=<%+'z23=+xT$!L p@j|)ڬQy^\ Ɏ MBqZG~WNQ0N9JmA&d!P$ȣ$eAG(!ul(u|$-v <HZ0=D]- agr' Z`ŃDC}ё]g|w겷U k+v%ܹ5ekgw2`8+WQfJv41*~38t-" NPћ]:wu1%:CZPp+γ9#8K%s]wJrRwEw {Oֵw?$K^njRguNԶ](W-Z=-2G%4.F[f#hCVnVLh 9wLiVK L֌LcLH^.DJsPͅVIG9,{c P+-TD632M|ENxmg ѯ"q)W[#ai \'jNQNW,RsՕS}Ord[/KrB+.Cs-깇J˨K`pe6GR[Ȑ }.GT ~ LeXwBnt)MwLnԂ0; `:%DKy45ofG˳`guj Zt6$JKtAۯyff>M'Wa5`xX(< /5%^E5CpQx?GiR8yua5pn!r# 0JMZ(,FKb,m2?M|/sEvi($W4=Ҵ HCY6it~O=~ez֓찞d֍,Pɐ~8uD\m=3-zk$I7iF52MR!(]_/h>ꓻ b~yۈ=]N Fe%Vz}4">!>FiV"& [A8lʝmBa/撐:1M~8O.'ۑ$.9Y4}hVtj\a&Qv7|;Lg:[v1jTbȖZ0ީZHɑ[\"2P#qj<>oz꠹`~Jd@ZI[I0Vl0YyɩJ(' ˌ6\ AXPkE!VZQb„=CTXuHf{c`jS%!<%E D-˃;`,zXE Fbk +k( [بQ̿"ʘdc< "i@N#RzdӁ#LESH-SfK)k y%sD 3-6"p;R%$H4Ong`&.b= 18(DU56xs"fIzVZ'q yy *+4 T J#0n8eXgXq2NF &)m(m@Y.ƩZk=4)-iKC +7&iYN f ,)p! Ce@6R1AI Lkݗ| P,(yo H|,bBZ ^\C޼UYWk_< Lt>tv·_]A:࿻7 Ew:+ghfokt7Qm0YpJeD/QWm.qR8!iNJ%*jk-w,P0X{a  &.s~;&*O6T["?\jcQ͜2(Ye6fN0+=?0r{aVL#=%5✗I9e =;I pCԌRmx  ƶ[yVNʩ$C"E<$mKTbR~d/~Gٖ5毿.~f]jC5Kf\ #1o; ~R'H'ҫH|y5 |8W:>wCWYp1tbkDTAjC/8ŸlQC]Jq-*YaQZWBN9B>H_*'[ro¼ʂ}~$ y&bSZoz%'nT16xN@kӥ&[rGzcX37Hjû [*UL':d1UyVnǰgn 6ޱWCqy&ӛ~b=߳b>]񠻆{L Sm}o-ٌˈdISF+*bJ' 0FbuDY%/X_~ڶv QHt;E+յYSu_hK,߉;E5%pуת>+gT5y/o'DgQ-}yr&s5+|br4*=i5wn*sY[y]9%Okd8:f8%( XYB(TBuУK7I\)Tb5Pq~=E*)(ܯηp4N `b%k3c ǭ]k_hq !b,x4HfW#؂Q&lU?F~VhU& Z8{fLdALfAk}xf`ZΩW&ئ6DMm|>HE+&5Bǡ VC֨RJ H-#˕# KBU5b0HaVS#TzpL_ȫab'9vGn#aJwi୧# `<䮻ӹ%tABRMoQ!AW^~q}}{w%>|LD7}ap}C Itny2Ot}\Poօtʸ'%,P:rX TE T1 I:BrJQeuW %9x`=u]RVn{M.!̓K%8Yk)]-92t)vAt^|SU[[G2X%p|v]1.<2rNw"nafo64䴂σ~:k]H?f fƞ; ;t۬H٫"^#NVQ)y'u'MB'\Qzi~vе: KDZ'$5Ob]9t8z*6()m32iEY@9Vc3<5&jYxBYr|7SK%H oovv F)ӷ):cvQJn0Y#$tT]qK)LIMT iqg|ت/E62_5:ۥd}ͰXӃbyt5>j}s(D<HO ɖn¼"BTJލһ%bPtۨNJNIe-sUv y&`SZЕckں0}zuiKu.e9Պl? '}SlO[ /:gr BS_ϏoÛ3cKgt;L&1b:c_VjQyo3}ۙfz=Eu]1H+%)J@g„xKH+e`6@~46}jn(TX.&RF6 %uXEe=x܀U]cTp.aLk +>ie& }J#8,.{B2Ќp0ǔ0g^qִ=5ⓞ`Z1]|SY]|wF39%h@ac}q\ 0uJ"fT%2RU7'zh?Rn})Xs=VؙHl.nŁ2,:>' M8Ü" "I*P>ѫ_}' dTls)kZp-fև)q1g*M$99g zY7?.e7ݿn~ϲ$˵(v\m R)&;=8€kO ;~P6h|f10yۅP.R9\=:|̧,!vHM59dPi5/>wȵ]Rχ|}uVq2?o0Hj.[ nӵ`pL l{y߆t0k-6q[I+Rg uJh- ^};;8oEp9p8椱.ulH`u&OfĕʖYB]P+.YL> GT 킫C%q|U!!T4nd/LvM(/Um}^ܣ1Wℤ02܃tW sv_(P$U;4s-^Si/de*C]b|ꀻӴAzQ!cN2pmJF8YN Q"΍WItN4/{#ǭ 3.CYB.f0-![I;`0a S(TO5*z%t\ȵ$Ud F{x4R4Rñ%(T2.Q҈zR؂VX+\*'=\^OIT:,dTغ .^e̓Vf)cdK]BN e{5_:ժH#WHD;:(NʷGt,vi p^f6YV+u$5zz> 8h^X%D+B{"c,]XjzHd$޵Wbof ۾h q e0pf&FjT(2 Svt$shoɭt5f 0VS-&x`tXU=>`\C] {k" 6$#`8"Rg×#ߖ%i{ith $isV#Ԓf Fi9)b3\AF!R r)[D\ b\')یK)j[vCcG]t˧", ٲg1!Ϙ74u&./?n>t+VV]ם5D}laZqIx#Şy ߾JYQo{J'z4FaNC׿ͯ]]fv!YX1хB0`B} h7r%-aP]rjH'K}PY([N6$!dd _y(Wlq k9(4- {?w@KQƥBLg9!,J˲};ZW??͆e=@ĿQJrpj2z`pXH{)eX2A|#sw36%LQELXMV`y '\5N}~]bJPi=vggܾŒoWlBwZ:D绗%3T/#ہk b!D4q3=fNZ4̍mo*-#wwlbN$H& ӡ/w3j gWCs9:*2 ͧQ0 F(~< khzS)޵rAsvKDԪc/hw]C3CY54s?FFrzVdm'h6J8ҭZjP zzc)c|e߭!Z㎥%Aq+/?·śxNdsbMmx$m/5h@d/w M%&Le<'@@4眜";Jو渒/ǝ5JhedAQkV˜"6v`fP{XGiَNbnsfm)Jp>@IwZ~bIIfZgwPiG5SÇtv,(BQx֌La¨L!S4 ϟ35 QJ0Q1z+чwfbF2j¢DΫKNKUp I!1zFHN BUCR'9_-~(#Ӊ,r /o> O/g"[8I>4oPg]e^FOmbn]NVK:ӇN 8tGIp~oߑ:fݏqIؿ6_?604A]*+3gj.Ez[vc}ٳZX`$m֣m[X F0Nmm9N| V$4:6 c>zoYZHc>I0m%]WyOʴhi@xV@GntyeG$NH|y-(߸Hry)SF<d-scT JPt y*p}A p.uԉ~$W B4uu`?€ZZiLфmYP? ~8 :M㸳B`H- 7mEۅqƛ4eA~|?"A,4׮#޼tsbҍЃba,2a]z={` jHt5ڣ]S%dnঅ>^nޖ ;)!$ HFtu`u)Z*B]+ HfJc4nϳ&mLx=Fuq`m /d )>LGʫaKH P$,60n%V:v%" Kΰ^i|%@N7yx1$nL7z86B&)"5yQt=V'r X|/J!f, _|s |2[RqR)UgDpPT h,]Xh4R˙"#u9ה7)-F y);T?^oFfܛHKf$=fo#9L vT1;V~vkc|ˇV"Kwwë{eev-mf1FΕ&d'M&eI2>~_.:`dRf$ 9EGA RP}sF =źUۙvelVܺ'rm㼀Ձo<+!}F1_5 hbm3E"N?|]ݸŒlAvMaB^zD(a!S !8e)hldoUؒZ2+]-uJ44 ZȈyb֜D|_eciS>T"RpݾK{+J ,o'oã:Ɓo\n7A='^*,@~=/A &r|1}un_] gF,LnTǣo&D};~_xԨ'ykʫ@k|_ $c=-˾Iٰ;4irU7ʣ/vJO@v;K+ ^ǚ=0!VӅS)GI_͢}P*IIJ\W'XR=iij%!&XFLśҎN[I2"k5˜ԭo M/HzVd7>DQ@, q̗6!;% 9ȍ㭗܎,XP7ػ"цxHHh YY+b$ۃ&r1f r;ޱ 5BM Hm껾c6бl:P¾U쥠FctP :OB <ɣ]XpʾADr GaS`_8XV g?H7ߓ-k;RvM,Yey㾗JI$NNdck\HMZ>QOhZFȎ4VW,C^jM3uGQx_7go='xaqD`xXʘ0'u<;qKnt<;;sYq@ xa~?BO ~[$ѰLWڱwJl" Sˠ_f6d/}!})K͍OhCG¦V VE6I0Doj+d5nO 0%@l8_P}EBv:wZgo-C"&WVL%sF[=ǭ>e@|*sX etߌ"e8EvI}q4$qLuЎj<-'eCL=:KO͐բem*pSIPϲ(ߖӻmHRӋQ׬$ZоaC޼dV߫yN"LZ [068;g5Ά[ t ^ZٓY'ۇe?Hq6(X"TïpPi,'R3 SYE.,[ !M]OpooȱQhj kL,٭;qrp|cs5*ϹUvc|`ݘ_6=+ϓGt)DXT(S_aT/T 7S7 BA1şFa}˅ ؊B[-4I܎Q$X3J m3jNyA~^鹭eyxڜ*ҵXuֶ?y(5/8ng&vx ?U@IULvhL|O<)VVl:ePa9)R#L#5J%G{e3O_o:gI@Й6bNx(DlncCO}`%*F*'FbҮVX8T.nz,T. Bc(#kIXHK2|MveDlXuF΄Z (&̪DLa Y2xq-)n[ $n xVpJ i Ypw+܇ߌBWIJ>YS-KbN)Kb*W+]*-m,E5*ij(%QȌfYUC&iek(%3P:*& MZ9+23T9TgR櫧q2)&T$zR2aY]dࠖx99geE`49gIa,n-f`>%voK͗]d['ɓ#[jI[X5d?3;ID?)cl>/仩׮?5SKԂHq61%1GwOnQч;/;b8(m#A[F@.XC0OYtJjXiǵª>h'/_(M!MUGNQ=ӉiiU竖l@8AZ2rԱN&"#4hU8QDWN4pn䵂6Ζillq.&.υzFQJh %{qa" hCzjuQl0ִ-# T [YcGYW_$:^Vi(:oe`X [ZcJ0C^`m~q];`B C[[JQ*J%OաD11iMCYB-֡=,p6j.6g)VBUD/Ȗt1_uXgR7xi u-JENjy,wOZw6jmԺ۲֭Ň,g9Vi67 A;4G@m.0Jc6y3jG?|,pF\n󇢞EףE <tۂiYs쪭'+AU|]WpEZIiJH`1cr%Brr#(1"d$ҭk-rM5!r ԝh3_C^!1~O^!c_ާ/hmx# \!QZ+X*Acg1^ /.Pxx'j6MLئh BR,v[r)1D(6'<)hpw36vktbUY'~U iףLUi:oH a"Yۑh߿{<_5j^2!@TQXL6D a]O^ [(1e1R:YN%-d~ytd!4YGx%s*-S]]nCcfo? ӲejU$ Lr7%黆I4yIS䅙zv>2s4 ,$vtbŭfCaYe$!O&md_& ۷ݜ^o&F%~7XS9ZTa4&$1re^*6 !ě$ ;2y{B랩xeS'ydt!\Qs̬i2J)ya5gP^4. f!Vߺӳf Br~Ğw1j]!y@j]}> @ z_!D_e^^h6Ew!M##}[ ΆRI;X %FsbWp$:;z,1zMoY꼒dfqfja'@L@ݣ j D"rӹ/w&?wz¶skB#,- )M\Um3܀Vڃ*Qj5~=H8ixFzR7p\lI]n9}Txɸ/#b"kb<31/n|Yպ=,Ⱥ ǩ?KԽckT⿿14,>hq@-^,3Eac,٭>lWUCg+7)`V>^i.-KTvZҐ7Mt 4xOr%*yZNwԱn'bJzn.%Z64䍫h#0s˺a\jTpJ$iZtYƗS[ Ux@(fjg>w-V[~)&0sPY,4.QަlTOP!ˁ.wDF>*b×DLcl=50p1H"Y&h0+3rl$s*\D."Bh% &jqؙ^`VL>B62<5ATJ u;)+ h J4zrsy%1%o2Fg\sG-4\Va.p4P'`s"95`o-FT*o=b!K.,2T$D3`f0zF9h4o鮃; lܯyWxȵ޹{aG|cVb1tt/CB)y9jDJ H9f;hT 6"GRUva[`+/h*Ed: cyg9!&v;m#X!9C\gF<&&U]h!ӫ> ݝJ(q%f7+ 9v5^IN=A`[~{ˍ3]\:d.c8Dzѻ{JNNkRZ#)MRSӲ~mR<%)m ErM%k=gur!؃*" b]ً/5+2:47$ssCOP,mzq# a#)$f>-dY7ۏ?SZmZ}2\6蘖6tmB 8&Ӱ}<o/dǘ R?.\zȣA^Dw $IمNC$s4s4gjO5qس=zP[k^4’w|abByĺ~ *,)Yɥ_/`e6mC{,j@6qx( u,}rC æG1 _;:GLFb Qb:1,wO6N8nt-O}45F0e XC9Hlrc2ǴX;%* $q=-ʋ zj2^l5OϮzqbhC. 6SD̛)! RACr L3!e6.=DL ʦvJl?{WFEc<"Obgэ>iF&%{EJrdYEnmŪ̌Ȍ# F(ɓ(dkpA "#0XXJBIIS¥)-Z!ic=|/ 8__=:%3 n u  kYke]'o[]?Oc_j>xf5gKu-7G5AQs;t4iGm $u}ZJ~sqS.! "+ 7f>BvL-FO1mDҸ^{n_::;dyP*1̇/CX僂ûs}0LrG4mҵD7fv&f2ӀA΄(tq_Ea>͝|)p;kXMkÿ`Ç~4HFb v8Ts99qfMQ ^>,oh˛C-$.M*n|'6g2Ɔ3EeC j I|y喿g U??C'B>v?)E'J#Ȅɒ>𩋐);Zg̝?_vr8 1[ V,T0PifJO4]`!nT4ꑜv`z:ȦV6ݲ*-|AqF8Ǣ@0(0&6E3i|ʪ󨝼 s:ZFrɯM ƣ2L͏$V#cy>`Fp?O[1˘Vz>w=qcrIO%ٱ|r+?H0W RsN)4x߃流pOɚ z_MOVo)VyӹسH1RӞ֬hk)r=c~F(W{򭥪"=X5S:ET(z{ JEq?^$ץu(  (B/epP4(ZOqѐ!KIdIі6 `(Z(UP#V$0'{<[t&F^ A]n!feA,CV%0]MNUߔyU+9~\Kngv ^ _ R;k,W_/J3]Ä.(a_??K/?^^^?=Oy}_W߹t˛Tt?t}%.وIsB}ۿkgB+X{~݄?f7inJJ3.5. Da@jH ZZpqSȴ3!h+[{* MpcQiлɌ)y;n*5ڣdB tYoCH!J}ak 1ʶ hM [ylN/\e7i5鄂Nz^m5&tpNJo _ AçS~X&?oC(ؔepe2(֘EP3^UVAQyT q6fk;crYT+O~ؑb Hy"ȈyG"Aʮ;C Hb'/L-9ծDZATN"fLӲ :aSK%~]('Shmn+otZ Ѧ{zZA}+mGc'|DR%cSIeJNw| k5:Sv]|2ɏ:;RX_1 HIpy"&)V6J,JXb,lE2A CdBf!ց]r%暩 ^H#53a0l]XE}׶3f243`Pٔ[r>rV_G NZ + ":pmPy#Ǡ& hܚH1Gɠݷ9A7TԕҖ[=AH<I%y9Q*i"U!)2A鄉>3"\"Ӏ%Sd |՛ F hS Bcs-U5ehvk *2n]55wRͥ؜lX2ȧ)L*3MB?X-7ӌ{[9Rco~Wo04=rA[~)-=\γW}7Q..9_m]Y'P?j0nxƤ!G6uS7)l%|HS` 䗅V'JeMwo߳G(~-g̲ʏ7b_>aMyLsws>ߔIO.߮فwyKWv`VIfYI#ws"E{'Ixm$kIxC0y<&o~siqb >:]>O% jά:5rTOz~6-*E(r7G4_>1:(&,}19(v  3N;mkp^ktAQHp鸬"\u F?݊hW/XAO,I-B~@.yAr9lt# $fUKkb 4̀16A`:'2I'eVwznu|)CH>R2 zfަAA i> ޯWJu5~Y9g׋W݅gbh3o|q7ݣTeEoTC[f\23i %^/fd$ۭ;[}Q+qvYX,2{zl-%CF9^pΛٝ]]U{, FZr{05hp3߼v\?V@HIs^*!pR6E&3F,,yiBi^{a_hb"`H+MfMΫ!٬ ; d'!|T%SutլCVm"Տ.u嚷N}D`qNpA CO/C6eh~é 2 ͯQt횻cdK+0tX8ߩ1{t,t;Q)j܎ܕ<]{aELcݩrgxl܉Z RRp+=ugWgrȡ)$kq[E_&e*/wwlwGuɐnElu0nSOl#g#8mq/f(:J{l\ FwxyU&`Af2Si0/ [}z n4JAFdGR9Ά/pb@Zk~ZU  љ .]Uf,Z&'͕NۈT +`$vq Aue_\LUR_O_-1YIKύyy!+픚8R@CYsG=^ˊ'eޙ42A>{ޠ6 9 "o8 @ӆ~p[qU>^Aع,vVx֗͆\|짇 jڣHpW'd2&)*)ObğHJK r[9 KuQn/[L&b#_̬|i/Ye]u}M\7u|3M҆z>wā`Ģ/r5ij8P2Gz?|~M~|LA݁1y:4{Zo~>ZXW|i_F(=|B !*rnj2yvOeJBǾo9$VYѸ"B&aQύ^B{&{`sd=r `+ȅeb6vtz5B}~:VVwW*pjkMae n4 @N 0ԑ HnN8:U!YBnf:"S61Ph!]L<Ȓaeb IPVByEB! ^SfU ;T69A?r!5 BrϕE ug=cRRGƪ Z{Hʻ(ǶՍf"T7KT([0:2AHI@3t0ՠuJ]ҸKNX73_llnP9KQM1^҄&siaBpvcUBOl,%Bdt}ӂ/SuUXlj+>[HFIu>8 D^?(u$:^(KĎ]-$ U<;_y/Ǎn/gd&Z_i 4odO9hD.ya\GhGӻ:kIEޏvU"x_2ޛMND2g=X*MZN{n6ᗒ$$l* @'ȇFo"[IщRf+00bLr-eBE uLyeׂV |<[pj_"hi-N{t"c8v4~{f *W9XsxGPLoP NQoIP̈́ ^:] z ip|bM8>&gN!j\40#aWe@ؖ ysڂL;ESGfy"G`}5Y^2WavNr⊓KT\,(d;[C !c @^wi5" i٩Wrj>8:vT yP=8:8;i R#R%Txt5x$1.jXS/B]㫋Ƀ. /jos?g}xk<ǯ_1?ߕNp["ބͽ_FYoS Bʋd&8"J˱pD ?*Rwנ8[pHn^J eDCr*Sۧ}R1QgԙE Vm=%hքS" VGL)ņҚxFđ1|<"!iȒB1Xw~cf͌ 837=WZj0xeK-KYߎl,#-+|qvu?Dˤvv݌`-Q0z򗑗rb4%ŀ]ߛW) ʹ+PU|fgGL Fk$pUߌAa3ɔ5o$ڭ@y.vx;+ME,q_`BWk_[iA [0<~yC/nL\7WOjry;/現 C=j|9GFb1ҊU2D8s?_`ʕTbEDPDSArh0ǨeZH3Q$5JɆb#lX 774-BCͭb[ @#>~9A,c2gQ.7|TS8sD1D+|%Mc5!h#CoŒHX nbxᩤË~9W`,%?pA]=DZC[.U1-I_umuXtu;{9~a FzbqI@FȞ#l!b\ zwBr`""Pw0<g *?9׶#ʛАW$+FvATU3j%Z= hȉNH!p{b[{LkcqL]H3\fpi64 HЋ,l&ݮs`cGF6UcmO/T/%K~YBOwj>*C xrY?,7X=m+K޽*H?}ꢦѧ"*R*S#R^ BzCE?DF%d}5t`G4<(7y'_cJ XT4ӫI{&UN @ۍK3y1xU vw!&{"ގ}1tfWO2c:0Fcdރ3Z(˴C4I (ΑѐR"PD.+772_Qwo.kX@*yٍ]72[%f檺o1˝kmTkȚo<] p v y` )Zi@8*0Jފ(g ӜyB5F&bGc'BSG ;Asf0v`SZ!EkɰY;;5PR,_-LJPb^&*) (z7ߗRBVAX;91T1E2N̗˥ M ')4ELLlI wvw|{c-Q?x2y\cz[D:og~ý#VWt%iPNXC-eH^tR«O&I GH"ّR<[YS*=!02/24<mxq% 5rdh“BmΝW3&ؙ\[[N64W~!Ė(m~GɺHNgtb*{w[q52ЩZ}Z<[G4g[{7rJg:^p6D] mfi@uP'?u"Ph-d"N;Cc5PG r`:l@L1~sFEdܧS^>Y^ x]jܧіbbc)RXn3%i,5*7DS'rWظ%}>|o5Tx5^tbE1xحkǁc,'aܠ mQ`rQs %TTbaI" <7\YYۍ.&V[<"%!AF,h{ [5`]%)b>ԲjU'RHLrX$C% AQ/RVp\))HlJf96w#H4&9rߩCX&6BSO/ [,ꠇ%G4W85P8l߼ RN{Îp$`W!hkc`דAfdF ou . DAwD*8\\NZ Ĝ5D<i Lrsolv&ly;YKF/%pþ"{`=N/}x5o' hYxf'cu(XQ֎Ań1{9vY,1uCUꜷ.ɪ IN7$YU'Y!$+FP߲5Yx;qZ3 Uq\שmVp_;(̈_Ah RhЗ lHd>4n˘Y%[^;=w\sN3S{5 KUŠ EP}ϸ]<p, m`D֐^I>Ͽxgg%A}z^BKQkКms>h-13;)꒢Uptƻ9񟇃2a8pi4.?߀eOo"LF^[t룷ME$j&Vݍ̭o9 * 3ʩdm3ǥfm̕o,m~g|;X0ix 7iԭFHDkM6ݏGwғ#=B2`|o $8!5%oeXۍA3z7([7@VynnA5!!@st!Bh.4u4gc' \1Rhᰖ [ɑUZ9^ pk@|6=/{g`4_5fsvI/W)ݏ#)> wH *^6V~  oxo ɹ9"\tsVbk)+%oFwuFE r|(m1:T\ -BIĒL3 =`XeP9Ҁ&Jv27 79„ IP9α"DTf)%1FH@>vHs@mí|wGTc%{t,Kӱ櫧c@- >Q}ǰ")#&8o 2DH_?pQXoxF!Gw6zkG,4ѧe빓9Қ;Ǟb4uzϨ@G1\UtY{~pSpxo<-[Zj9kK?yaB{L8g\yj!$ɼ{L\B"8$MƁ9O%8 ȇ t;)F1^%q2sgt۟<>sm݇.=9; ~o6 3jO^/c;L_7}NjP ^ .սa0xzK9e&cd#r$Kx!!E|ip7 ʶ}Y:@Y!6޷e5FaxF7υN^yk դ£XC^/O^|EAwIr)*x} e02rݾ8ՀugbH!Ϡ.BkL)r&W9ޝ*C|RῊ;M*v75V_|6Ń7q%'J~>h]PM؄+ . ppu4Ŏg.axƭ?]GL~>?.2/ qpp[?>߾94n@d1҃^mf`wČxp?PG%?oN^_! 60Q*˕Tc*~9}g/-*S;*pϝ`Gӹ |b4E8ҳ]Ϝ;dEŭE}#caf0N":CqP1AY~WD5 iWƣC_H`Inlw'Ǩwxm&jSֵl m1Z*V{콲Vjw*P:mqss HgHS ISfO>QiJJX%'K/ iPP*,A S^C *!N׍,g:Pڲ%8CvVƍE.OX65QP4( W#e"51@R&xg{9);@ ˇ'w Iri制)%Qi04>a_JjsŴ_UR+C4uBِњG^OxԐVaq|~TiRdCVFcbJ?n7._뒍"mTbΝzRErFQ|&zM0 ~ݐBA餶QŻuxښwKqnuXWnk6YvȄSZ3E>67`ܩ]FA;Vo޿?\koUvilLzdf#f.HH_D H)A%uل/$q)+Yn 槦r{z9?ޅ!_C1ےސ-X<)AFh-A㇊^?x辄ݰ7Yޒ+DI|mi,gޱrf B:m~qf(߬xqi8:xߴN 'mr )&B&^9V`|0% #LF8GTyk9JڊiuD;wM(sء1Ve۳ۯ ,HK8QC0nHWgV읆@sK{;."w_ON|~y<2jhqLkaU@`-%ОƆD&T {TDrσo_%~ Y7z~ k0( H ۮfNy4? q]pwqܕ ~Ze"rcDQ uXrQk``NhT#d 05\f!n9<O]BQcII{^xǺ lT) c)tRZ* ! J`r赭HhS%X6dmX2!H'HE rAk3{%42c(ǭP98EҎ#s[1YdA '\έw4ViM;t4?RPDH(楍4$ HHM#Ek#+$MRJ+rj<Z ˥s8Á3d]u䲦Izv;C0$h-5L0"J&.I`$ȭR#,r97@-IN :gqtp['z})%#+uKVP*IJMѫh16Fl٦g܎Z !6wG`=ģ]ꛂI@3ԚmwCE^!DwpaT"yx_a$BSQ^ 6Q*z*$<7N{+Q=ݺ2Q1oRjF+Z[h_Kv( `%Sв$p[]s4 ѫF!)U^Z㫁C""p Ag#6uۭ@+oSdl˵ ,wb'kɽKĺO mD{9XwI$9"D/jK3Ei\[kKWRNvJ`ٯ=zM[r@glx>*عHœ#JMW6UI*(گh@QV7y"ݍZg#?:Ĭ~ώ|fH0ÒZ%$ @ ?ʖQQN5fROF*`y iB -}xŻ?1dZ>>gb@Yy3 ~ǻ'ӹyx>}¬3S|2v;9F?<=烀d 7l. .f,tſt㶗3˒0zJ\)GQCh58$8g'mV/#B[';\I@$nݡz2qt# 6~;Rw^^{l~;nzx׭]~u#Y1l)%ƍ>M/!AaFdbCTU` b۳ۄdKa0Ny{@h ӭ+s5QUҨ_>Cjkj,K1!En;EU $_*< ds{M=劂.wFbBN;2HNA}I}u51ե̫10:E"xtf,Je1lG]}Hvcښڌ=yE:5<1q6QlQZF/׬KO^+&G[%I]䥟|5  F†О z4{s?b7r<o#{Afs9ϻ0;yԆCTo)Hؒ@R 3Y!A,$DuN_E`ZRA{+I*`1Ik'H"$NdS"AS8/129 2cJA d0Sa-ZI{}M}L7N9φ>QZF!a_iͣpl?B}1kNC׺P"  =^0tЧ q^{%'P(ҵ\m"D^ !> ^ ^pMQ)\ S a߮fڽزS7P:eP*+nvʃ20/A tZjm }&(KU71YZ!wU9)q9 Sk!txQƋ) @Uqi{iM,nym561.~ k0(FUh4{9yjmdj7O2ÿ W6|d3@rB[`i%T@U+ q4";+Пx<* A2@q=آ5?~4-Lm_$7ję-Hy'3] x?L,;d/0N#=3 ߗ,rI ~\uy%<qunCKJuRhY-('u.)E35]Z<י(,Dqh]݌-lo.& Ge[X\T<8ϋ닐:8F1L!]؊,ᒕV JnZ0fج+ 1P( XT <o6fv?muާi~idLWڬŸKdslq2sR;IN`{S#bs!{;7ZbZì5x^.8MllGf`Bpl`sS;.RB;TJL[jm#% )llQn/͟?/-J+I_#slцܿY޼PyU6-TvVSb~?}uVk՚k@1bN`KRO` !V_Dto}ojGoQ=| >S ZgCJirǿM[t I4jZ kմ sJO.zKd{H0A@;Yk>1UkƓ6XlI;t+8J ,f$L  j%bwvuwF ~_L! Mo4y:Pu8 *NڄAUca^իCkbEݫli4*mptB{~@RstG:j)eڢ~ߡ̢_ XȦ@.Eت1&lj@ۂ]`H!QΠddgDYk'Lոdj݃%Ŭ3_W<=׉}[zyi柳h[i( a^ /d!A[JꖄYJE$JR&?ejXppSNdϑ:^ ᔉ;1u_[\8S"pRL'y\DEa.㚌uC4.b*黡E3V]GuW_e2`{KQ+&če_Sy)me:sj#~k= \(B)_nRBv'tb[}r>b^ݺ}i gr ^.8<9KU3 t_Vfjh|Hď={ :`~a|XLN^e7{ASǟ N@3?g LMlqrg39r,|.JMwLlѹW.(9-W:9𪷘}&^Ùߧ[@#IQ >@} Dhm,HFl($bTk`_5:J^aҳm1O2OSٌ"| Oۙջ_"Eԭ[Qwg0@ R%9`͈:$($08UHc(؇ƒ[(- U*#`HTnJeDc8(=0Fʒe:3n@\Ӏvcem|2Ee; Jrv;h{ͷŧ'EsJoR]Rpx),"㾛g;קd{e2fn>66[r xUrT$n qw_.VC-d X,fF(%Az~+x<4"%v5|LzqtW}֚/.dˇrKټ/n6{6πKQc9geD+IP( LQ QfmƤhdVD 0/Mu!h)6'VlAKx%f $CpRh)Qb2f2ZfKih0Zd\m$Eʢ4LqhqV%"׻_ҕuclّ֕3.az㾛ᤞU3Xgș@(+r(asC/ne)E,v&_AVRԳ4!7YEf3nQSJK3mfh^(Iwhjwnvw9Y{ҌhYn?KQK%ΘT"+3%rqrXyft p T՟Dʺ9W[bo1Fc{ r"WPm ^urRANti mg*TU@%BfHMDucjkj8jm P,ӹt.oR ~ɞ.}}˺_֫. tm՗& *OB*s1͛ $0ȋ>A^$"d0է? g {͆;Dok2Zb_]{T TɣG o( s9iЎdpVg5.i<vyy+$sz굗w˃Gs.Ps%AK>ysYWezBGb|F:tA݋U8un7e}k6:^OY0Q;+bCo?m_Qܳ|F pF؟yꛦՔՖdg̠(uwPl`JAx7M;tw"/J ee֭϶l'9lM2jI:Py 5j/}~k'|!xϟR^|Iv!IC[=6 B+y?EsuoF'aSRlzrcݑQsudž](0c*Ag4u[5VtpahJv00L`2JO`I/{m:a]v_BLCd /.{Ghp}kk'm/{ÄZY`^o|~ #̸xmKB kGߴ QPIz:Pe0@ R%9`͈:$($08UHcU](;QVS*Rsl~f"C̲V&ߕnV+T;gըJĝ5f#Ņ?*7s7z(6\<>?VE6Nmo4 u~g_G~e2h}kIUA W:f};ʑme OGGrUjv!_yFŔRɏ`٭.8bv'tbFJⵣ}g×٭CRt \O6sP\';&邠"bJj{[&rH3(:#j`N̘ iHd QDđ~;]JY C!ι:eP_|rqIhrH5]^oҤ pP=zǎ⇎'q'+z4 B/E{^'Rcy *OC_QǠzB}ι 4dԪmU,53MZ t~hű`DK}R6@jg88IU=O|ZRB}.#TTc-d1**PU#-dG4?iA@851=Kp\w5~@BY d2co漆'+Qnw^ M0tuߚ͏,X2?Dԭ[Q.n-96y.AB%L:ynԨ @&uXH`rwמUX_߯&7Hu,d[1AW+h ?,LȽ~a%h\+DK((ۛhݮU܁݃ɜ&{Nuvvq}q%_>G_Er~эN*TL)]hz5.QrUHP R``ʝfgNrZGCKɓBKDLbe`sCi[!Xm{Yù*E(8("P_^zFxŴ&*ڟHyHkВMrsAmxF)I'>!mj,-\I(tS1Ox;& M*Rd[NEnn" -}P@> dYcR> .ӂlad=%LNArC` cAq糖5w-f=0kM'`ltjH fS DM}b`m_Ֆ.ߑ& Ôdp60Elm7be^d@?yze^=EnْZ#U/J*Uż3hڝ `D>=*Ӷ?Ξ"ҺGw槟> C%cOP$ALL r8SnH3=a;<;e=<,Eo;Dӎs L@:܄p ))%pոF: %Nd6%_(qhTS'f6;`UlªB윜k^onAl^1t`²^G|*4T5,Xy|4&#x_Nsfs^*d'!rzud.8>OG M\@Πw*!$LwMu iD_.$GL+@=ߐJ\Vq}e+ >G7\L|V+dSڧ&ZN[z&nL!pa5+%k =S7 E4 L˙C9U*HƝJm&%D%= #JS`V|ɢ|8qϼ{x@ON 'x^`e10be 'D#q2xR MX0aj9ČڥF)A "M$h磡iY>@"a fZ+薹ۧk_$:P̧{f6QNg}$^ .X:菱5g'3{Qs\IH.JMKsa3S+ t,('! Tp5Z44`5'.)\Im.GF cD P,Ш@-H@Ўt݅F@(m[T~EȼhQ8fW̏]K-s׳  (25nW&@- uGym򭼭\ч+!tR~= BV*[sTgGo:LqyZUfwzo79KW"trJELN\: \V_ُ\#nqLkb,*'9Juk#FVSG@|Y]J$ͣ8}ȤP}|2sKIѦZ0J0 @ 1rpanr>o)dpWC|㕪S[;Nqrg&(B{5 ffPWdvy} f{O_)ґVHeuRA+mFj |-FY͟[."3cUӒSich~?X^>fz-i7)_߼#DsSC$DߔV ų/\G_<%/0hB#0Lq4#⽯kVk%Id9[kLKHPɎc1\6btJ<=/֨g X1+p=j%#IxUB+8hUZ%ؿ5Oֿ9ju _<`p@Kc5n}Lѕ re-QRu^^TݿovSH^lJ4N PpQݘ+ƴhXFsubjر=5qxER"p*CA*uwnZOȺT4% 7"s*$0m];)VEvR>n ?~(@|f#B%o=If&h4=WwU$GݺXwNiI4%k K^Rˁob4a:spc7'v(:1YW5Q:رފud_ZRE&f渤Ggays[مf'JܤK.R(*'"(ʁPyN%L[3mYϴe=Ӗ͙,fѸ(h@g +KX Pr>pxWתP~~Y/Í黝ckrJ1{B3)8KԜ4F 4#I?\Hn!(Y^Zpُ~Z&C7_gI[dojt>Evyj\P$ihE~ {^,ʹ5sP ;IІ$gRk"gڐ i1KH-Bߞ&`[vCm%n7vLfkd |̓/xZ>XK G8.睇K g? Mk_!cu؎L}Zeu},>TTOJiӵ"UHyfWHn\BFEoZA]ytpEaE@NMMꨦtD*ֱΫn=mcqa<^Q{D+B:,k*ѕ iF OB-RuhK:T߸:ݭd;*!R!۩ / e-cbm(5;V :N<NGuS00I R/sjp.|6S4I'keDX Gb$ hw77nOR琛J{C+КT62^IəTySx֗_M,$I2"=?֢W|kR!k75xR΋_4j;onÿR]pU+ ;t|E D׬W9AwR |^UUZzmI@h#M@dDZ">؊Bb?5.ZX[]@wjRt[Pmj1j%)9-_z$Q5mtэf\M>F]K|^_ *e +BZ0RCQ9oEqYը7Jg]VY\Ee‹I^ ]{. Ԫ~o [۱Jv_~0j>Vw|{L+n{Vo!|MT4bs g+2S5ǻXc^(oEw:W2OUw?~WUb:_SlHױW]бy幛/I~߯n;֖D4[ r#| Z\kKuoM p+۬蝃83,{qRCO6zPr: ^)){LJաtׯ=jC sj9\J/eͱӵB/5 }٩9vuY_9s +hvX;)q`vC/dUY?® t9S]W@̎K7ˤGf~|Mѹ `_b粞9^ DMiژP7qNmJT:!@HƒF x4iڴ5q2휗ϙM],%㥴oU +PUr Gg!Nvݳtu?' ):"LJ'M;K" zQ?O m5\N-/~:'$־Ve MG=P:hʑ!F;oK6[ ܡI:"@٨oUJowߠsPp1솷{HDޙ]7TS{@3e .l$|1AQ8IA(@X^_;r fAG{wpYOڄQ/~5+Q?jol|S~p}—?V6-xu/J?E4#s2"_Vo^>GZ}?ykM⇁iV籖6rVƟOH:iY#-|e39`Iy'I#=W@ˬB %@ry\y\ӗyxˋO>#%h[l\gAv?SxN !7H-D$d,#ӯ"s)@Z/6,(/COyۈmrWs+5%IZR}l!(5*T" &D)Bi *YdJ- Bl|&_ai0>l 6DM!94 :t:+h."lu i~稶 F|0edyVGN8IPD9,h$ ^'SLj㐼ZhQsr\R`A%ScyBZd^!lR0/g=E~-j?./?]h.t:ӿ,*1j#fmU:9u*b0K,$%k 92ui^9D랛X>-ŷܯ7-4^ ysr pn-GtvߞߥWՏW7!}eMPR5ؼ+I]a|}}u!Ƅj!߮iεyXqOX|->qj/nU(tԗX/}qKN$m'OWX+(>I:RCޛۛѩQg,շ;^IB~Cb#NbE"$&[.~ ޵57n#뿢6W!g줶vd/%$;M忟)KH+5K4Fwh4\ 4>X8Z2< n 4"P DVT)P*)jP >/i|=wvf^6_I ;06hWRx]wBsѴae S޷k - 7:*-X{quX3io%#aX0QEa$(! )"YP*Q` FUuWUuBQަPQʾ)"(xquXKM-ٖݗ<P> ΥNs4YZʨgD:7$k`X"{yW f{D핓\mX r`E7)XõGIԞ kXj_H0z u&B3qK5͇kD̈Mz;c+ӵ2PZ-Xh>NhC Uk ,ԃUJ@äշ%nWT"^ )q;]ķL4pmfEhm~Q+%P4FA TcͱuiJْhjJm^hbLj%4.O+!ת}ػ U VjSH6MЂu @tYv۲M !9;/"Je#[ʾlcCIavzNmln=u`#긶Wy8Bt,n%%s/;3?ݸ,1eُ~d}㔽6\q2 AH;PsfX 1nZ_wqdrGP*֪AV P(d]ΗXΞ9iƛffṽjJ-%FUl~EwL0ِfu@aFQQor_߷CY OëyGMטVgוl륒캦xAGXXTejGILu客dOiޥ. C_icfGXtPkNŧUfvl`8']a' cW-+ܷ84{G4|1sK0a,bſ2H> Xg20o=|RmWa cWc0v_zT;A;gzIK r`ǺjZܮJ +&JS3ֱMz1xڮJ;m'sp]V)1~$/d[]8|^jMm 6E˓dAP0'ԌVf/o \wOڶش,>sI?i4䍫N5Qj#Z7_ \TǷ2x:uBC޸z=$9g?^H89RO8JG"ླྀ!VRBXpr.>%aYr{'9Xfl- < I2ؤ|lٿ;m&=ۻIY2l{{ٍxH=z8^3Jx&kQ  .LpM '}fB(C{M>/\?w&T-T z*㥾 FTKv)&a|cӟ͸~}[6vU, g@UPH@bdl.+BS yDi$'Ytf{QeH#cyZ'Z/u8a"hpBimR@YWQ9vVN1ubiK_U48 wbfBh2q"dC9cMT=;͔(AУ n߮J%ꋁKoeS›KE?昊Tq).DW֬G|1pQ6X#8k̺4unUhWQTZ%nP~ Y1DQR}'T88vh+n݁3ZȂh$. Ds+6%M%eJ=a oY˚?T{wZO݆9 V!?T(A8`)iLTkIJrPRٷ43yp1%R,N!]vo߉Mo{QSԱcsЄF9Q<LD9+o.+tN+K1SꞂ䱷q T4 emMv1;I U'")Ųk[.DYqhF/C=%:N5g~d-Zh')# 8/ TkT9/hY8Q |PEZI&lR m3۞@R(>^w/_e]Κ5NR^<λy7TyeŝOqˀ),"Ќzc iλ}Dړ28fmtKk_fmfCM2[v,ꌵ"mm Gxѩ, ̋Rꕹsy n 5뮑DGz=A t䭀` Γ+w>$+^sR…bPhdKT,wp fw9)R{L_+r bPcn~*ƽ_i1/f0~kUP'3bfcԤS@{;}mA#in0I )B xI8SaMbqD4Dq +U֖/ݯm3+[)`؟ffH@/n|E? [[ HwL3fƣY~ZXʮ77`UBaJfXjX(D &8&"B!8 `)-IZSW|#ߊ_%IDHSBhB5D%L)EF(5 1XXb?b'J!eq0DI%ÐF‘iC7I$9YO9§.?G)tHCFHZh*ۢH$8C02HTcl8 YʽZ־3DJcb!T:NSqbU0Ca+| $8e1ޖQ_߬f/Ybӝ9*Gx:?>ZRRnIc}S it݅n8w+:~z߅0h<Θ{! ·8tNɣ g`<VMH,~tW.` pͻy`s7T#,kJhDw|.wHSmuhtoܴ 0m.JR1#Uٽ }Y3r`N߁"e,N?IpE.[/J`D/x6TNL^!2p^zQP ]{/3a@3 ANzdBX1KyX`,U YxW^]^%1"QMbJM̹&Dw!Khe(ӥy]c4vD`&jٚ~/eªmeXUV[db RV7[حXK'' 7L?FhR̲UՃ;zO&ƜO}YSVKoc+dOE![_\~^e8D6̙ }}iAU'T{;a6C.j8'uI'\ic AO& lASS382q3C1v{Eoĸz3P~}/9?Ó2r[\yqWEjRR2fqV^o׆&$qTjx0tV ƶc/κ gO? G¡7G 2?զ;nܤ=EcD ararA6?5Ok30 U8A.D]ZD41*&hEMQB1J\ (cT8xi푹2OQ%kn,>}4N8f^#tt>9J|R_lP6 #pKz8ZvSt0DV@ZȨBJ^^qRFJ(OL0Sw[x͉ c!ҥŖjP#{6j|I'v_@uP,RcUu5+B5|:<@J(h7tQHBKP A±LeS[DpBC[X B(d\z •߉x8 &u#;ӄ_ Vx0W˜ƿG*?Qk@TSJ`۲\.7ֈs1pٶ;p OU]RvH!uo}qH`Z*%&6 p( @FF"[6I%fD* 6Uuݷu؞"+uD\0(1si0feޕFr$B^y0/%kE _?GO ZE_srqy|-ޣ.M9K^!ixD]QjXp J"xP# #CZD})4>}x&%e I4OwSH6|.HlUbjFUǵvU`ky'Œ-4Хrb,?QVls޽D'"BEݒYHW;8g(B!;L_}zT0q͑ Uڲ`^@ 4-G&L6˘*rNSn^3w5Zc:7V^{#Z(S|D>{W1oķ6~I2y|Ԑ{pOyP,^yD橭:`?caZ@Ayt.&I;Dm5T8!حmԢmq"}JCvdj1fa8q nҥf)}\tȂ@Q ÚEWLr†5&N,жѵhXk3KB#o;fIfL7-AQv T\c1ŀSB) Ȕ4Wg |.Ps,ܤ%spnK8UrC 0&J0Kâjy 02c-Su`bl Z@6]޴FM?ڊA   H)d/&9Λ8o㼩:ΪSfP}*=1p(#ޕRJ"cD"Xcdl nP# nށb93yj)NU6aY_0bqe'G%IZ8&P&vhA a7x=#ξj?Y;DX$=#m!wCy< z },NiR$:DL(R.NClBǷ{O_Kbғ&CE-',2q&+-ĚRȤ&@]3)nub&e./_?]&؛rο]O _o&teL~z'{qݧ?r)$/I/}㖗*Ga݈ )Qvs총=x۱a+"z#lg-/2iuۡCCIZqiq=pz&bIf8:jnbf1CoTKs$o]2x2zDq͆!K֙;?H0D ;:70DMKI;tS:/pĈY/a\ ,=%kWq)uSpE(]=)mq2 I;Ơ|ə#tV {%AѶ͏(@]Hh-Xة(LG$D !Fƺ8ֻ!!pB"6XKG)jiV'e(ėFo4rCq7Wwa: vF5,x0 ގaI8wUV) ^"S1ؽ2S Ʊfy.JYt$_t.;%2,LG3r11t(?]Wϟ/砼4:d?_ޕK[R5)*{*_ak4BH~o.Ʃ|&jWmq'0+֞Z&;IhY4ъwPRWضiގ4zyr3%6(JJ[,wUmJAb= X FxԝhwE'HR"8BV tV >0:қ`JY|#S%tUKvJĂ0X?||V-$('.vs5բh1SW 1/JJJ*LQWB@Y+`( ֒,1SP0(hmǔY0bLvBoL2FR{M4J/WΖ+(=G (ԃZeT`U-2YՀܜ RYd^_^Ui YgL4tM̧8)]x|쯳&s9dDcc^0t U:"7T^fx-KtPkhpNǙL *f%3F% w@-)&}q޵O@ִWFFv2__pJR>Łj'\y2[L_|ϳ w[N.f+y%}~T<]\c&a7+)jwyyrH:#:q@Ǖu7)e宗4R0 W۲V_:d2{q.tH^(KEw1`+'Jf[<<ˬGY2-hq@F.h'kZ{U@jaxX`-=HTay##$#00N^>OD 1MYzd5ycr<# CTW1SЄ; +KN1H9G\)wY,8Ƥb>MFzSxycď*F-MPePk -G-Sp VU@J636k);j.џ9P:x˽ސZW,^|_R& YzCj-{V}ϊ<+:8F tR3$Pzt2/:[lWTl<b `R_Kͫw>Fp>i"MV@aސZK1ZY[)%yVJɕP"Tk9?LaS_L~5_uB<Κ(Y0 Kt!WVO}F(} _V7.vʞH]3DNjضD!J=_}~t wӻ_x5%!z`GUnWHzJE@.8:uO&U1џ $T>v.iZ׮pb^%~7Z\h RLU$&)ڔ-vrSoȯ2 f4IR?^NE&!Yμ"#,2gt Xy N9Mǩr>Q>m|R*aeŌw.Q*/_deSȎ1d/Mt J_~x+kt*D Ķ!Ȁ$Ҙc;eEb' iQFL;Vhʂ>dJ:DvG'v $E{3NUlT:DOcf0ɊAڹNpp}MAF-Yu{z1з8 D4)>h2Dr."X% /t(z/91X,Ue4z(\Z\@OR"1=Ѯu A(h6cup4;^*72y>y_klYagGxOc`0cV9G!V-qꅂ4N`v ZGdAwp`WI0A9])b)aphJ`\1.ӾVҾ'/zpqn+RrxH G+NSm w$ []ocq;%cùRCpFý@R'*),I-J`6Bf"JECK Ճ rm-E(>ݖ/O8͓6e&wZ"1k %OqdtHÙ<m ĈĎS﵃֌nybY @^.3ZErf,:}IU >1ӎtrk독әQ#gj۸"KOo$- %\rk=l%%MVWdq*v$E7 3?\Q\QUsOpQ1:7:%i9I&WDZJMO璛/a>S,!T tl[keم_oZ0v/[ҹĢr]xWD>T_vBjq/\Q\QUsΒ!B='c&9U2Xc*ao.EI.{˴ %Xº}⻬Q Q UBBens$Xa-8qZ&B1Mc8BTX*OŊ k@!Xt"ZS_$ĩ/[i\;AP~;ZC)G$RaЌV2F/cQ˨* N,0L6kT^fj dnxa1*%)ԗ?V#RdsmHjsC2и81o"3׵*bBcx83%*HZU:$"+ qFœ*B4 ,wXBH0]yaӂQj%93C`E /@hOHa)ISJ5̻(wlm G^a%f_m78ԅ5 4+"r5Zr fRs2՝Jbc8_hhOS{F>슰QfhPS`2YjGJ{pL8ȭ/1P[q-*C?kP -Hz(gUxXxތ$ a$]4µZefLPs^{'-c g D!a3Sp  Ib<fGVys&w&QAߙkw)sqk``I'qa v)#]j'w$][\oFUu4,x6w6+_"^yyY8 fJ8C}s2`xLxzMn` "YR|zta - :]~Lg?.\/9 7f$Do&g?pE ,jXK?􎏛#0n GNAEDF[%n#Ir9Ě/YmI㦇rp $%AoJɷ݀oGt8 nFY1܉E瞏4D6;$WOR(+rB:Z:KH)({cY`jiay2WR#W}dZ6"LQ0%[XxlUi 8.hFvN'VeG m "V 8E`sJ< GIGxMsa؈p:UaJi0ILQ3Zԑê0%4"LsfUv?ܪ8tC(``UvS|dUQj*\&٣G\WW$ldrx ie.02BNeE=a g+Gwf2B+"8dW [̏sܣjp<f /}?s.W*n` v!gFiF Z9,@F.Nu  gM:_/}7}p\NL&eqϩRuK|kl!UMĶݟ_SM9;|[ ҐܹU^ BA4-[!|D @FcáJqgMa435⪇/6rPm{:or詧>l+dQ}W?cy qT Ĩ{ +xSb^bmwe[=qOH%0ZQ}VYBeAeН>k?S-a^C½"M!.rRtxwT%ݹ!Q;DtM45f2hIMш;!6=Jjta C`4*s8:wX]Uy@}T*K$o\tzX9<9 ^ =ƫ#+ɳK\UDh7S񊠵8ɸjWSg29y߳jtEZ6'qR3ᯙ?oVl-Gal8Nظ-*&]V4P?IY84~Ay1Nǿ^M@MdBᒕYpuv0c/MeDL5 Q€m(E"P$mň[\QMu/-J…X3Ap9*\(dyai̝3PR*Ja {^(6]_G[ʤpNgXdaXyUʂ)|w3hw%o}> .:GP}frs6 an%痗Ά).Tsi=cs+DdL#:aJPXM˽]C!x<~cF'O_?$bNPU!r / HHP OaG- nFxuoשW}9.h(ӗ{V԰gonabΆ>i=ͧ8n/׷fzta+`>Z(|-X `{m|bn$sbm^Y6_hn ha3sP-e^|iX]O=T"^*r KO>5`g7}5&_!?/ɖv_&T a:&'6!iPg6Xd>6f-g?_JQ߂M~=?HlvsF']zѩB ZuӰ ԙߗ^_^_a;  `|1`^7a|#q0rB~QJ6b&37ךYdwZFO甊/5DVtb\~$ gxy/yhgu!AŒPrZaza=jk{㬡eȨG-'\d8Ѽ^+LFC@ ՚{bah@nE)ey۱{s^1ZDM/ڍ\cM.Rtjh"ҽe*Opi!9ȚPĐj[Ӆg49*{1 5Cm\BI&OQv:B*IP3v^G<@ ߯8B*7{55ΕZRbZ֣bsbЕRkP\.3+B"fTdYOػc~^1Rc*%&랔WrE! 8AogGk(D@ B3Qa%(Hp֛˺B59}j4%|E?lS;CD"3~wHC+rcd* f4w s#d.$Vjg QB0rY-~ܹ&ZU;8'{SlBy Fe!rEθ` FT{76G9#+|[dh?>dOΆ/_Jhuu mrOT 7˹Brj]ZZqC ]Ӱ^Ćpڀן:\jr騯?HIbM~Gch)Dnoj!ůnR"R= O;}Vp>qU`!v&%GeaU:}Zq]5RV%uF#y?,>"4bP/iƼUȔ\;D8;ٖwl茘gopIc]XxVPE2&,62w6/Ͷc#GӷDl?)#jqlyѬnQKb߅MhyUkYKuҾOHZb[4cbhQ_kSRJYjxO En;xb>D4ꮦ3Ofs_m#Y_vsJ[J+%[``3!)'N*R8 HvٖazX2 ?v#'X vPEw'#VYúk a髗=|f6Z M*)—H/_Qk%׻mjm&y;%m)}{=3 $ոL})(+!J_0()i WJԷQ,: <謫Nj !akmUp_њ1ԛѱkfy=}ܖVX:6PoX\QeݧY<&o4!R#e;Ned͗?2\10F)7PkP]G^fK sٛ1EZXߜ#zWmXz =U]Q{š1Q u;,j?Mtv)cX"tC*$G8pԤx}#ch;jµ{4mOGZ Zyy*=4\Qĕ5]pok)^T2]&#h2RW͵]j`Z/}V5hHK^@TxqvN-Bgg9ev9ǁn???|whoo^fV_7 &;d_?h-cߣe{=Zoߍ(>q80%+c51T32i4ӂ9p"U~>ez:\Mlrs3f;f t`m9tA?&i;36Vnl^Fp+CTx,U{\fec. VS5dmmWۻ}T}!sE5G_bbZurFBTߡN!F^\[Z.xzݑCk֨ {׭dӟQBW6U7Ok™{Ui[P88UW sZ۫8_ƛW3.~7+_}P~}0K إ0r$$uqM ͤHe(Hfnl"Z"E'aZfRF%Vh`6MT($]jmHi1m\RQiT‹o#_.|p"gpR4 f 9ki4"0 9s g)tPb@ӂƹp?I[y<"bwR4W˧b ~ ےWaYS7C!2*V?t6_r5t1O(3@noyO`mL09y?bL" p.8Hv-asp" hS75IATSD^ȿ="^&DӺBkңz,lTlb}]iC5!kYkb$U=yX>b7Cn )3^k‰ej ׺33G wL:jӌså8qXJ&LN2PE4g~`98d !ÁٌiIbX/,#6T!X&(JgI̤0JDq `e$lca0`Ec!qd",K*2.CeƂk`R0a XY>7!&X & 1QZ3f| 5[.rо:)7|zd3`% vb! N/qF$5K!$Q0&-].P SB󢷺]NXh~HЕ8ƕ4K}aH8 s2FXG%x0* @f"D@G2G"10LaRSoW,Z#S2Mh:[qdȂԚ%"aR( b?m틦ZY`"DE{D%g1uu&} X%eǒ2S2]^(}yY*(!BaY)05gOKx9VHqR[`V8xLBM{~$ź2iQTB% _,/Za&,)FfD< g9+$r(l$Rؖ G)`Y,e:Br4iLn3>$bS1Xh5(7_i:U64W|V+)Ӹ !#=B Dz F"rw5.@uMX{$Vg/t.B\{±T /bSTT3.RJ$i ci3,֗YWI,`iSAkf }zB˳x-=SE(2 ? ?}0TDzrόm(<f<Y[zӻ}1^ _),o[/QMiVR RQ0X|yU|W-F|ip;WlO']@LrLBXnLBY><z\l\Tx}a#n7љŤs(bz?uc8k-jx4(ЧM)@**s:P* و>lH ;L-s~gAz4>I]y2zY.kCPh1; h C]y-FRzhiTNVuV| zPh5[dU-T%$}r+"(M} e%҉zb*6xӷ Lqq 7TfF[n\eL![ы]hB^Ph5OE\ R\fo)?&1 y@LӐ׻47Ks(mH!;oݏZJ@廇.?dx`Pbw_ ݤۧgJP~9XI~'#8mstUP{VSl=Ÿ[ DQ-[WqVX|<ڹ-hZb5*ռ{zfL{knIsd[9*/sgrLbK+uwnwi>Էgt&ֽ3_Q( ]%cPxpZt[ڛ:lG%skT)& }_mRӜ+yo&F$#ZlY81EiA7#E +љu ³%_5\.s߈cAйyzfټ7YSy ! y"#S C}4c-щv_xLW\j&$䕋hLL֐!EOiop%h g_$ TBiO8 +AQ[*oƕ ]*x;w{<&ﬨ$(*4EG@T[PMwW=*X:]hԇIE3KO %FڇpG֤LB7ǩˮ]Tio;|׹O}n/ 1oώ(Z7 ܂}kGO6Lmuj1 bq an!{w:a7[ΘK#?$3tӻkR!gcQHLMHe28N:%]yyc!̣55C-)c^8ӑ=oK4 :+Y_9āCե1{3=.Ԕ΅pIx,c;0Xx;*eǿezOi;IT@1k يd_L-y@t@)gWKgj"z2:wàrqO] qq^Ĝr‘o$ZƉϢgyT#Щf?^ |27דr7>Mr'/`[B-9""R<,q2aF1!&K(VjbfTa1gL ÉTYϻJCm׮ѝ]wZ*H7ԣPqJQiR5LT(yBGFJH#Z$G`/ҜO~GMnbaWk`-_A,da&up'0Hp'ulJF r E`NTmN7j8Hr l+[3-f*oا =xS1Ɓ.׎F+{r ʮ@V ;>($E5mԀ@Q qB0Bgժۥh9xzf`,՗.iW2jcV؇Ű v:9Ti 'ԘT(ƵC<_,)'&E$"\6|e4}9|pp5֯|l_`E72Nߖ-=fI=+mUx EIԿ-JWV!:U`iGZ'k /9cb^^UZve OVhRRuU]-֜y`uG5 gkMvkP-8&&C~Q ҄5 Į%=lgv:]p]טJY_Gdcr3uycxAL]TzQ< |s=HҒK'5';' ¬_,ܩ1qLk1q8jc#@(xڵ[ wԮn Ƥp ҮDJ_xdOye(U RRck\\B!2HL%qdi:iq  9ZU'`"RaMR*!;]YsF+ z=̶卍v;uvkW׈ ) !Xաh/*/57Θt:Ff4V㜺Qq#L(HT .YD85ELSf-L6IXa\Z `4w]m1̲mwe(z' 5jS+-Y52ګ1އ-7`eRbR nT`ډMf֧^?̞g+ţ+il_#8+^eCS ZQğ"GSXy҈y@"A*Mʊ(*9Ѵ8Oy5c#!i`kj((ETJ `Iz1)*̖9O#eegakH*MZoSZܵ|>P({Z*/wOb2U_?|tA g `K_\x-܃eBo wr}Xz:&47y%<__MJ@=}zruWhD%"ÒRӿ}JDhM vݧ.?VZem/l FvW~TC#0O hΞf)'J_ȟoAP^THpI矛)KƀQb-BJQpT[ޱ!ha]1[HY91vdg2#+tB=krd Ù[Ap \o$&#Oq+˰B+*lкo pZsڳvM{ :JʫherБ@ ]*˱юHI,Url-n Hhζqnh#vTxTL\M9E0΍C8nx(KGغ;_nBͲ_87zSWF2GN0jWKAr<֜O6 x`˘˦̄xNlH FZ1}^ߌ6{cSD(u>-SVn$?Ov_̳s2wdtق\݇Cι=ņc]b:wg泐!wvjl428\ omW(}Ђ R!nezDvıGc3&87:y:N;oeCf`][i 񁫠Yc& ƜAƫ"zK/3`ٷItF3)ՕO{:[P+D\g|`+ʙ`x,!- ]-ւk8nxfR" F)ͩM`hݳye$-狚#{k.iuC}$U¶J 0JB93!=ʪvKu͛`-ŽtLJ8ipQjcArU\ z$FmFvR\X 5b,`&:JD`F'1LxQ nd]߃^44f o`Ec$xhy/3$+HdO = lbdlyXyH+2!X)Q(l?Qx4Ci)x(Kl )A .٨fSΔbȽà gBD~{ERQ%^<0B4y3x}аXCQ^%tP=8}3MQT+Usij+"Y>ϩmOl:7dlGcSFE"n.#/N]ur"T;P*d@b$P^Ayfm%5C+ɳ,&lyǑ44>s6l 9q Qc#YÁ]ĝΚncY}Ϙߗ"trW#6_} '_mK֛!'du_Ⱦ.Wq%RKN'QZ_8k,}ԻAUw ׅp3i9ԯFuW"Wce̙`x2\krt`EJQ\jnj =B "*TD!cDBR]0 (~y;P馣ND:8 dkGVpj|bDߊkrY̘0"b,cDqesʬ #`bґ#M0Ņ5ʎnthWODRt, RdQ`zym*U}Շ?X9nyȌ*bog-*.eac\ >\뱃܌~v1ê:GǨ>,œ=[o:gZ#. [StJwR\] / d@Y:8+./e$n"uG|FlXWANwwS,T3ԅ|"Z$S~jjR rDm ݊ݺ/\D)M,:cC]ȽuGB{s:|; 4R7p4L-p.EFѫJQ+0ee"W:Cf4"FE΃B 1}{C;-lc^v|C:L F% f ])kG\D˱S8 h0U~L @EX~fR;wumV*IZwB0Ĉ^G4U/)Ew=,B '=x5 \O(=\( e{?Myy0%b2q_g[EZߤ&zOf.]TZpb& Cƚ{U: &,h0Ki<jߢNΒv=o Z!L;o).|\&q3e) 4DcwNnn??9q 5O7?7odY{Z0xB)xy/^|:nThY^|Z>>"Ξu(tƆB8?6J Q|1E{pw>:REX= nz) !03riYwwssu{:  ߿’v7/)\U5Am̥ͿY֘D&{@KW\q}Xi$.l|qCH1&VM4).iY{ O$hLY $;tV]#5R0ĝl O QYg0iVX1ď+t0{BR@Fkkn6ŗ؉I~Q^^\.o?d], X1.vmOcHÛ!fC\.[׍F7З,#ހw`Z[⌆ 2+z5}J4b,=Bf`Am+8@!)ET% JsKa'*@FWL^.$LRu &ڙ#mPMPFofʼn亭)i$BYhi7fP$>4DdhAA⏧ctc5;/(Q`!xf,XXEL0ӍK}f>sF =u'z"fkؘ7eO8g_ɚg$ΔC(:-C^hq1}&b΁w;(ğ~y; 0@qݵ\/v MEOo)u oD ̺un)OODRMf_7x~(8:UG~?mөP3WB8 B8 B8 B8* a9<G0kD3 oQшq'A{BHP0 %gx9f^x3 FO6R|l#R3cb^(r $yY# ͔LgTK=&>. 8 聰6 K v7B#pBR$B;!hB 1"5"4]oPM|8< Mi{U1# P62Aڡv_`%B)  (~Z|4Ü"s(*yhK~t|Qa 77z>]^\`4$C/"w"<Z+4;0Lư {ץof\n:: vM>2!f=دc+ أR^8ԅ3 vkbU:nhP.C]Itj VR~nJ(~~~}5gl6pёaHlPGv's:ʴJB9DgkZa18~ۚxaCɌpT,gxP Clxj-) Q>4xqva 8B:>o?[8*y35 Xy}8=~f=^CA~bz=SW\T 9X<J?lDW?5ԬqKMPP 7s6l > mPO4۸ ) '`%n1մ9y=U!I*TWCZ6u[}lg/E=|O5u[LnU8H&+[g~2jZ_ ri]DҝG[gZ^`՗{w䮧[.9/o[B3khӾPcJ5L$c55jjQh/jtRU1p4VD3>J;v#*Tng['ɳK \di  u}Fgmu-; @׸َ3Sc]^3~Ng@oEr1T~(frEQ s&Wt XR5/%+QƑ-0*b}xYAiًsrκpDR`ʥ~EWR3Ţ87Ϻ˖VA"TE"2BdzD"\Rc-i L tH\ZLq Y2n:{c7v6vQ+xt2F*f3w!N孤xq?xF7fmxJo~{~l6x{U ~7 nQL3Ѥ_{_q(}jIkf!oFcW@N8MxL3tc~qWw*]lnjچ[+c4 cXt/]RY ME浴3K`'jHHXW*G+ɫ^Һ|cp=}OCNzdU †HqtW'Ы1/4Xa &BJ4GA4GA4GA4Ge,WB!-UZ9o0+I2*'0R I,`4Gr2@1EIzR_/o7 JJtƳi(2@">p,Bo~327 _>Iy_]z3N=޽ʐ0:aFAB*nBLoN2ﳟpfKF&EŸK+Ȭ}G˅=r@ Mh7c.'o|d=U [\ۦ 'Afu|\!,[K9wz@?NC`ru_|m؊r&xq; 9ntG<2(rc@@e;fs+ Q"=9cJa͑8 %0dDJ56 ⸞o1i}[: #92AJ䡢V(13YÌ-3.'.S!}L+koUrn`K\IBpMsҭvcTz-I}G"OiUlZWAnuH.I2U:ݒOMRlƻTH'٨y6)cBKWRK\'P εjK$Hv~l b'4 R1ѷbPm[3 #M:$ Dv~n NbeM&s_{kOr#X E4AFu PzTYa- S0V%-mҿɌ:;ETf& ҄wAKF&uhƺĂNG'Ubk\JJnV֠|̌'#ĤnL4uH%ZZ:|ZLY!سjV1eE 1`dJXrr sg2Dn٪F\.b?Y`A+]!4̀0APsrve &M1D{l``ck i1K;:&jqM>ѝD[jU%p(ls}l&d.IA׸EӪ=D :4gXZ,$˛m$6ێN5z}U=ς/hATV K_ ੨lSQϺJuaz@zjYQ/D% uLx|B%.]d'EeXia*kuo,Wm &(kKrV== +Ze)$Wx/Rq0ү?3L*t7FQNWB*gEIo&΄χ9P|OV7'3/BbTS&/95O>.nj>yMQ?mS^SV$X{= z ކK+fnw8{uyXß 7 oߨ#n@TZI L?1i q&T豶DYrUNmr!q{L^.{L}VwIZf4ϯv8O~񒫤ejD@9W{+M󌣜 ;t G:sMVJjƴE1D `RK"(ԥ륕< F(`FpY/\ /^XǑ6Y1i1UXL"XNX& ;K/JK/PIh+2(9D>UydvvowȐ3=_}|tzִ7}0#?ňÖqUoS!\悟odAznnn>!f|1WI۾NYVL ,_{~OPRQbMhmBXmX(k. 2@?r `"5[ gN .I[^% K8]*`7]i)\dWJH]y;4 Idw#؏[1%mIP]dTBE^g#7VB8o fNyOQvIQ8:YH0z>n"Uc&|܀"J5}B1sGΈ:6"OJAّE~O|ϋ.Ѱ~+ pt/^`["ڱ#s}$Pfof7*}<ߗA!)"#g},] W7[O38xZ<>/~77#\S(MBـ.uGLtv`7 v`7Ub:WCIehQ.Ŋbʹs-&aF@sOJqR;J"K\(ŴQ9=rU9:t vofK~_% G @?v?7J#VISÛW߼ybg}nt}(y E(El vtHZ/FW" 4*%jdokg2znp-n 04=|蝾a(y3@Kd=W< eP7pJpIxGL 58|y9TqDo)A߾JQ@ S PDi.΁Z$-(#2f ћ(Be$K ,$z6*VrF% a Gf~8J- O XF4F2b0"NTWV b 2o&@?Ej%psmI8qmڤ|\KT ')(҈ks=*8a & cSKGI0b7vp%$";JEE ð,qq\X+ro6ۣBEJ ѱw*?d!ǴPc!{yS+k9Dò#V/fJ'"+p3Y|5Ew9u  uJE0Kϧ}')(ы. pG~^?dsSGݶ9vH]f q)l6H\Uv dsQ%w_eP'I)C)]A24mqWx T>#4jS0jqtkͲk2~PU8Li揊"e^蜂)m7tJ SL{SzA"Q%8sh sq2jc[=yepfXύ*=/IVk4ЃDD z9qǠXygWBW&Wu⿆셖 _qTP~#-url<ևH 4iǹ𣏕ЈCc*h[V#q@12%p@4W)X;}yHA1N$@ P*+)xLcH<ρ4/p8;@95Fzc# B*FZ$E;oB~S Nfk5:$}w~c3>obڨ=C3 rE{$m[V& CeeF4CHtӪTJ5Jl_6LFZ69b@9Ql8hj)O?xNiu<^;"7m$S ˧W~1:nQ*Ѵ͵nxGCZ,VHFƥM-zm96-@ x3E3&Tϭɳ`!^QP<-^-J WJqu#5L]Õ~5!*f'kI^L\MY#*~?s?;ͪ~#&?8Y,H+v3Mkw.-߹$Yzwxyx]qNoRs*˰FW>DƔ_W78cXR1:}4YݎzVk^ү#W.FW>D[)NbhFBID'L )#v]oKUO`z[P؎$pH(R.IT 5Uh^%q3HY]p. dqʸ\:!ǧP_z#&C.v-R8Ol=7,AZR!6e(„F+M.Q&Qm`D2гPWbd ~!1f_Lxܱ)Q E_~E<3 CYcD؉p4],Y/ @JJ!~J 2S g_~,Qk)"ݐ5*tjAE?ֻŖk,s :)a$fވVuU"qş.9S3Icpw T@j{Ƚe= n?)ǻXc̐sA޸cȻuG+aڏhT z8+1ԷF[mw U~юh%oJz*k'(ͰVSWsГ$M'La(N&hݮkHJ``WCpJ:-)zo88 tЃvCp"+21LT#JNq-49q͋B)E%SRG W)VRY@$s1|L#0As\7{ڒ&3 Ԋ(IL5%|;$3jlRFyS@y_ȤHTM/MāAwDW}(R8PdbL u5옚Eb!6*=y>" EE9J2LdU&Ϳb J:4RP:UJRRنyM5Nߗ1 jT*˞Owr>ͼAx|_~,K)*q%@\@ |&L懻[oANC̖oq>\.nkO'}^ ֯,&*<#{Yw`c >vR_oK yG; Gnoy'Z9~h$&;+IBN$= ԃ#@^ȝ )H%^2q$'& ,G z[j`wޅTQJ(m vfn(-dŝ{]oڢKb3]n/Ѝ@UT ?|H岏2疇l;Ud;`oEq{HX:>׻rKrt),I:TP]_eNfK?sgڂP T^؊͓ L9+CD5Żx.I|^PFrs3}\;nԽs[<~o^%Y +Fb o޴^QM %Nd\)D,Y[Z{;`A0PhL%!=1H#}H:>>Y55E<YŋOO VCu^LdUdB"KxnQQ@L1 ]q}]3:k[WGGiF@ލnw>*{B|t]g hDiꌮ:\ڭHD_`1 b3©RBCFknc`f HH+ftHҜC?cz 1qcpi+7DzN- O9߄=1tqsPZ:YЄ¸}&8^ rJ]gŖ ˉ{;!w^&ױ@ӠTS54Mx_PTreWb>eЗw~lؙmJ??*ꪮސOP-#hH@m7LQ H)(\4bJ3)If6]n+ǫE8˷mŻyᵋ|vZt粒vf6/fߺ%o"fPuZ&̦wW}r#[ɨ;(eҍ8#yႸ"K ڮfYt|fFd p=giVj++2 610X_Z¬nj doZΖ6Ξ("fV>[*9)8t>(>TyZspy(M?S ie}%Zc_ ƙr0-IqZ8ß5)iގMvܑqoq$W0.{pV[J6U뤥; N&]suRAd (dCynJNnڞ|}TIFTQRs!ytփ`1(,'.:zí (j:IF GOG5Dm5'B )S|U&Q:ɜZ% G$M󉦧QDM[,H>k$+Gi0,k,tHZpy8#V'zፖY=4PƏN x/[q[ 6 rRKQ>zlnoWɗ-nʷ_}oKZA_Xz&=4߿,I&+T?}}wW|&i(yZ!LTU+g ?q2=SZͿ.y5͵er;P706= ۜ̊dW7꟝pЪ:9Vs>^&n.8d_<?n.cp_/o̾kr8-<(țIr!B`[vkV{ %Fx9 X!$(SxhR\m/ 6ԵV[Hť$9 W{< rP%fpITRF\QhøxGB %EZv'$мdQ>8k%dGn8Ňiv}"\B =F1M#~~dI! l-1.z⸓ƪ߲՚#SNG.w]68ڲV'gG8GnpeikRg*(6QӒUh<1'JZdJs/gJO`ͩ1QxОsm AᎨiռ-nr}A 5p2JdB4Di %Lx)5QP zFK%$/D A*ZE"M U.;Cw-oc8ۛ3M.k(d+aOh In 3}nK|? E>i  ˋޮAˋOno_,^׉OOP(}S؅'s}BSF#0xPr>,2Nhat{ql\dRo—`;j 29$r0x3j.gee_N}rB|fQ>X5A8{sjе7prױUy jFGy޵#O"%lI8HXjnr-B˻uvvUTEsjc%YRA?"u7f:1W }c]x^ ݵdMY74bw!ۛ/ޕuh&5s^ӊ(ZK<瑗i[1@gnjZTtfJsߥ mۤtJKVuN6J Pij] h4djKBvZHH? ΢]cZ3YJ(df%P=;9Tq\JρqčBj<=Y@]J~.V Jj /1궸g$`LjE72gšᲹP%J=M9XQ+Wh7,osLFg#TH@7$HZPh`tFPE%E(iT \?Jޤ5Vm`}0.<BLQpGm}uu!pŖa8@|o+_ow+ioد5@fA`[V2?uL[ҳE`5FT Af }E=ŝmӵm{=3A"@ w[J ԖfX|?(A1B+{ԸT yzuFݵ02PqTTJV9uLq 99Π,1J'n*FX`bcDvf&|iܯ/*$V'Å'R}ee}s|3/fzb_>42[/K6WTю;4RHQ9 @hrIV%ohybxq;]X'/TA?]etUOWuT70bXpĐA(=UQZ1 *5QJZp5~6ԪOҚh=+)VI\.:j:8QXAY5 褏ň-r IMAl$0](hW+j(ۮyz5\.Ò$>Vj zI13nVNpRA՛zq$I.lMUjmR5<ɢ:Nl! gp$>h6>H\p7P9`AAӃY7`4\#$>xaivV$X]zG3Y7AErvPkokvPCj1[Nڋn#m6QUAd+;նw:ȩzt:ELApCAg3yMG؇+Tkr pumK A5sUdd=F*yxp )5`N{xzp27ѕ{Z(Ȉ:fC\C$\tTQuf]5Wb~趦js5_dBȎ")+=fNfI 5)Z{T-&+(Rifѽ5iW!. \Xylmv*KQ.-X^c 19Aѷ `>.ZZ} ]rlh/"0=K-AݶѪZk1 k52{EՃށGO|ЊW+fB2;U Jy[eR읙€DF:;&8ᘨi.Xz3xF0}kn>~dژΞ"'c/s'e^&WWT69ס[A3&yO ӡ1@D_aDа4n.yh1/DE1T6Fꚇ6-A~^ kG3(:W@KА5ڠ;ւ莶Fe豬%#v_ګLU*S{U^EԼT2B7;0HKIր0HNk`^$TZmCm||mu|SHbj)P6|)%N[ײvE!z HE}Fo2CkTP^r@\?۲^*UƊ;+ŵ$0] kWh.DAEi9J 3ZD9MЌ44rnP @: l;E (.mMYn ӊp ACo:H%OoDnvZHJ]8IPVQ>C%1@x5F,i~cZiv\Qwm~ ^,ryHLr;JOQxZ֍0V]u!YJW:2=!?b =rB/aß/ Nf.9T:,׷}n+' E۟%Ēn҄#ѷg+%F;M΅7kv{1!S5 Q6I1#j=qdZp5NҠYXmP1hŢ,mq6e,ȫsZWz$*e~!y"|Yۮ3YoϠ2s&'٦ó8yJ!dKv>ǣUV~l@09k;uDyv -g'FH+-ˆ!4|p?:{ܓBvB*skBQ=Tn̬աɻ6vti (hD!d"T!3 [!n\r -zbve5eZEnQu0_dћJcȯ=zl輿۬YgY}ؑ,=cLZ)na U<$;(4"iS>?VOd}eg[*\vu4 |.Cppсu8<@7%%VB7jnڹvbJ HnyJ-61)7W4NYJ/⮵pD3DzX M}-8ByJ%)Rqu \cyٵۡH?F &Pb<0Ljs1pۜT3ܘ\@!&BSmJ{@κoYi A]1lvO8U= ^!Mo-یruxWU4~+]-/yއ{ip<[i<  }*^'͖%%_XswHAj}v#rZ XQ|}{; L#7!nKn(ǔ3+K=>8i彵E:[),>_XwNסnדGg{V%mq\Jץouyhb8>R9^>o5!pru#GQ%u SLrʡOS9l#a\밥\V8+I_xSZ/xp%3S`]\$Ɔ +X;[Һl@NE>VZnնVy -'#['͆ H$*# 1Gl ,5dY[fT2Ƣv5nV6G)#LFl'|ؓ0\Wv;S32.'ggr2]ݴ9qDx?o"f!ܯЦ3>V8pVG])L9|LYV2;|0ǝFn#\JѾ9p&&isth_)*LCu?V|,$CnY'(ЧHE *s(li3\ZYS4Lk -ʺq&vZ\vMk)ԡ Tt-`W^M`5tG6#(^7FS=W#pzw{ ;/oytz9-jz9]R'회y `YW}>=ϻu)۩n:u>Ԕ9]#5jE4_D|} p?d3Zq8e\ &S؃ɨ6ܭ,O@Y_^n zM~O$DwSk @Wec:hTF/q]lltn;3*?7^ڃv\L%i^ryXs=zR<wNN1fƍ i@a[Fo<_/BVVu-;U%ҪWn$?͌9I)-M yPe`֩ 9J$,0h)FJDYJgr5MR~sN: EкK $ ǖ_9_T;4-_)EA("/IW\/MR(R0-m$ e`Ln Z6k&6-`+&I񮫚:v`$ԺiBLJV#bF}bF` ~aykVOOXSK~^IRKq/*̗C|7 ˏqC:gP23&j@_H'?|ܟ>|A&[~?$;+n8 bom۷'w5S Iixc8nTseE|>Av@IeF0'͗5\ew]\[ $VpXI)8'خT @'I=) τ8K:γQKZEFv0h <b+A5>u}Xs% Z]$=e n[8|R@*蚑GԸ4Ѷ9Cզ 6d - Ki f!*un/ȭDz?r;E6a=lcqHw*fNUuPxv&gx/*9MN*α\8t9/4c( iA;`6p{֞!w{NkANt7F"4BiMY@R Jn &<) +tlDE"o;Կ:gb-1qA.TX[hH6Фs{xM%%|$ =Lr>QGR&k*f9 |c<=5 RZ7O m .PI/N\nz*ջb81b}M^ٗ_B#v\<~z]ρmoA,e JЬ/Fϙ)ܓ Цo(puD9~'I~7HA&#TE5W㫩fmȍ%6KH.A!ИnSٴ߫K[rPK&2X>=U9fPhy?0[d[E->7Yq)) `kdC?p[( XdIkEvgP#3_w b ﭓڪ4)Ńr[j/nYyL9ÁF2:1.9Gk'! IոO;h@~v'fcS狀wšm:wE$dg Sph˱y"Q11>`oisvX6粫ޓac_7ܰ]qN_,3&`B&,VA˷тb:T^zr<}}ϋANeۨ|a9|,ȗ6 B: @U2 OdTj,X^IQ^bY m7)+]lȎfr! Yx.y!D`(Rdi3"rـj ̑Xwe=H$m!'@> l(twbeW|S,i ue*E02!4jX^ԡ 9ոx E(,P` '`4'l)-Hú9P:@BZ*!$Nz/?BS)[o :K(TrhYJ.Pt,Lj-7bɼ\ZPLP][#/pYR*v%އB%:8-%J Ӝ:HDیdiNd(7}Bn7WDICZU{ ~7v|f, Q2\iEMd[p\3( Kf&&ۊȍ4! ؾJ-ц^-!S d /OS\WoUf–߆_FWw(fvZp)2#,0,ЙG\zc э77#tx0fk\%x#>Gs@=, J_Bz6n9I@=lp1,8wU;Ìi: RnF:+}rimPJ!VjV#w(PPk d6m$5}xrGAk')%_\DZzTbx~z51Ѝ}^AyltTlP5ɇUW g&,]΃gRJyNUlJe~K^j=%Ŗ*M4f*h=Rj:.xqm+мDwzxSG'h0#'rȴŧ:hԌsW`xysrl̷k8}a(%iIeFB& MQ^GqqZy]^LOFr+JIp %5^!q0 ꨍaAyP> _~@U'&/WjwjWjCq4NOڳWjwjWjCnIlQf醌D6կeB.~f,L)Etye&oBH2,ׇBwi|sy ۲D-\c[-=}0D@O2i `rBozڄ 'ꄙpZ9=`sRJ5-gi' ` (KB|&bhʀ>VnpznzPkR+sr _ tX*z,2R䚢))UnR2R2c5[CҠ3pz^~.gۜ:y]!Q DIx)}R$eV"jnuސp _|EݳM% a=݇uѡR!u~*Đ䪬JY]K㑛^ ×۫y,FUw/6Lo6_ LCkR1hT bD'M[<]DXg-<&R!!g.I2v3hz1w4nG" nn-Hș2U,pR9ÃAz^FW}pN=*ܽhXkd70&$Dj;e>'m1#lh P-lihǿ:3Ž]Q?+>*ҚZ˗Z!M{L=𓅒o)Ԗ /qVY*_Fi'cUYÞE I-!B=`r Zfq8l0vX> YXҽ|G' ~ L7o!Hk.` oW3)F7>)]\XL$lfLG ^&tpuڪKa5l_+nm+X:A!SrJ+U(}Q(RPopfqUcڄVk ;!dYE*d] Уv1oXq w,iwr5]?E:Z ґ   2ʝ>k @(Dq \4Ucn(vVUe Wspa?.@\؉)Q>pDuTmǻ~X HwL"afÙsC&H% Aq.LHQԚZd-QA iE1j[;2 %Vcejq"(RrQ %RT#@r;kܠ"F @ O_%"Eh><\;.5-)XtIiʒ\R@wZA^RbdS`/»їp`~Yj.%a~ÝW,XVm[*' =&Go_@?R"!>}/G^teFXV:#_>\NOs|~yx>ݭ@$)J!d\OmOK.0!`8^p05Np+pֈ }}Oу~wݠ TPHc^/wDrkmm`_^Z!@s\ W-*QUp:2Ҭ(JT @k0JH~ |!KXVFX]rt@$8;z #Q O:p%1U+A8æD(F-aAܘKX[X£sL^2帔&6.U>_1o<{08L+S@2ȸ%䧪|Ca ɽYǪ3B. ̼R[Y'akQث;4( _3 ޶k.pp2*-R0T&(-[=B36 A-MGaSCZjtԐm?T3m?U &1rGFZks&K\"lh)uE"#fZۆY= )kL%lļoh zxSзW 4yvmFέs.'cG_QiFڵ~%wssqW$($Ĺjֽ!&/yn!.GuX#™cֶBj'P LDSGޜl@aY?9!j Mt2̞ 'gM5+ٙͼXfOMD"HnVÓv;5,gO8ֶ R8!SR&--߂${uϔ6kNkhYua\scJt;)IqNx8N E L͈VDO'~tFn@|aC׃&7 } u7_&!etf p%- |u_xx ' ﲮ+2go1F?uRAΝh%C:j{s󘍞#v>Q/ކ=qfT]W]]UQoh%?G?E"# 9SfG "] P xT Y!)9+.fwZ+]|kz8sqv_8n".4 mkS;:iaLN\l%!JAXMX m%?I9 qD$bJh~uT.Lr+$+"7bz\P'$5J0 Y ^R*)ȅ^IinIxQ )M%Ւy\9Վ,xs80Q(1%DŽd-/]抑xHXLh=Y|%Q OqS!% 4whzd4YMJ{ p}[RJ)a |)\(5.- Z {2-Id:DaCyJ\\YnON=IdkE&"=)ڂ;%(Df*씹`aggGB\.uRhFӫԬs2GdЫ34PQ^֙uU_cc!7 lD\> =K "'4 fW}K8K-zs>bQbr՟ZgWZaigO5S.VX +I1Hsb4}:hÏ]GO@y0Bi1 $YVB)ȡ$ shޗJ1 h-Y,PVP+>I}̬7Os+g3\W=v^&)lIZ-|J^F_‘%4})Zz̟*^f*o܊(duI5/|KG8I+}Gկ|*t.:_Ow)ͪW|_N?Snb<꧗'|>6WD2e޾|x$ O;u|q͠P/:UǓCsSr*M&J3 .i_.hcc-]A,Em[Th 48'a @} ɽ9+,f%*x(J;i+g."p1g΢ޕ5q$鿂llud]վ؞q;u)AET)B);ˣJ)IH Qx D b2$YV3 E\ζupNv8-!e4]dHfe hXdP1X2>w(!r*ôc#qTr:83+5"54;gEsơ L+D[3f-MI$hi[;䭻[P̨"+^eN1 p 2/754mZbo3rHR)ּ! yyjC><=4tK},wZ3AdCRY+z*5L` % )tcx\.^ 1,Ð;! r21q4w]0-P v8.T>T9E?w)Dd3,>=iõq../i>L;a0:~4 D }}~s5!N_//-Wy} @Sނ-ּ\&"E,cqfiah'O!nS\5sC;تFsS>n@?pm/Ulf6JW4ϡ-$#` YYQ+9\+z1\2is +yW ,Sxs u9*jK櫨[*dy }}8n"-<8hyۥU9ϛ4lq2 S]ynR ƆFHѸй3iOl֓cT>)G90Ͽƨ嶺Qr>ȯqpN"(jn@G-^󟳟 2aeRedݸ#ew /b]Vp_z[ml['_,z!ւXZ" B~!]Gb(nĊvǧ|xQL&}F؛\3dq-PB*Jk܈ ~#Mu3%0wD @qc0h/iL 2 |ngwJg ^"߳9 ]19; zk9 /Xf m dgYɩ%i Z|Gٴ+ y|%`l,8<^{ & wk=DOLC)ײi8 snz8n+ uaoms˲s?s..'-+Gķg$Ԛ8P[G/6Ο/~nszn[۹{-9paٚ@+Mu* Nt:WtA ;3/h-]eD ı8c=mQ;I<1\QL{e`WaCͺG[G4~?SoUF5!قXHCF[AL+F9PuЂrzAKҧ<$, /cp3Ĝ6B#9}~A7 Hϝ15_$P;&qT }`ii#^(OzK6*loFή;_]4; ym"+'e$EL5¶C |֭+%m\)3y֭ 9q$S \nql֭+%m>[煖Z>$EtNP`W۵ͯ9n|>;鷴 Qj$Zrq*}z4wYӤhkDA( PpׇIgT2onͶapp];U˶;~q>^޸/Fw_tX\r~ś +*+="ehHR9x{%OcfMyVWR%H%+c-('Z|2n*_K7sA*&' >^9Su1R,r6AҊj MwI=AHt}W -uʶ˚O稆¦b`CGyD=mdO';%#އJR*JWXؤ@;U v PJ|~g͚Ǜ?*e +~ eqaOQC>L/Iꔼ&FqD:gy4W9>ƏG&P|O?baGu'Ξ% k_,SN,il;dKXd ~0,yhSYʎ?~ir X™W>ӥx#<Ε[dkJLt4ĥeMn!!OIs޿}9|˭+SZfh5jA7Ͽه=D%kL*9(Ul>߇Uuu 6'YQ8ob.<HkoZSN2rhfd%R:&|> 霒.ܿB_B#s' >q`a1[2b@c;f&QNh/bȲ/_3PPwVH+SW*{W9!Ƌt8h5 h >T4 ]YnUE<*kl:V\T;,sJi&E SL%9SarN^I9Gt&:(%)a gQ CT`9ȢOB< y am4'sUwV&>w:Mޜl">*5[ՇE:AN/Z&vr )TA%h)v7IMMR21JLy猈" Ȱ`稚yT2қ-@%?o]kfs12?9hj]DCl@[FE3|DXK89p!A @ 5SV J,\7] ooz=nE-49lLeroFb8|GS~znP5]G @Ֆ\cZma}.(@\nkzys"˾RƸ+ǁO)bɊw+(c}6<+`|m"ܖ؞w7ך*3N1 'k`ڼ΂S*3kʟ][s `)p*bsF ⱏh0n#6@ JU(A Qͻgނ P8ӝdSPB4- P{8WN4C_m9qո\˸]M5I_MP@堁љE1uNaR ga+IR.m:9m` W廀;ZEY2SH*01I,euBN!ɽ^˃KI4B5՛?)3#CEU8v #Toa{*Q=mPwi: #R9c f]nir>G'm8M:0<1 >Y1Wvqv6Q9ьpB Lbf<+%Hb AW.9ZiE.jL_fŷ7ɕZkyO*eҋ7BP 0(!ݝ)My%+qY{zn(¶㞳Xm]B8-OH-o4(Be28ٜJ7[5{M~6cC4|I9DIJI<HBNnv@.Ѕd!ZhuʉF$J%;/AA7흽LT_IEΪ9lzz?ޑAm.K3퍏/_V.1WO_?k&Ț'`xs_)9CBi)d&AY#|-(][oǒ+_fO("@1\^}EA8[=!9{8#%cQÞꯪUYm,⨑ѯi> OЗћiq_NFffvw;b`,݁P ?f;Gz5oXsnd^>*GksưBƷAND.06J0 rb$p)edy@ YCNg"$ET[P&%u@sɶb+r?_}h1ps׫sYr7 ˧R;x}!;DyۿQd 0ב1HvweA"mٵ]0.;dh@[oS$ M*u/} y=tU(k׽nΎG <3WGﮮV_xG8z4syOz`yPyDz|1bknld{gb|msHK57XMS|@]rҚ[U7YACߋRhB6W`[7urwO,g_7d}ۆvւl+iSa0o-қ^ȉ8x…fv6n6u|;EA(x5 )87ۄqz6m?%&jB1% Z(IޮWđ*bņb \/Z+)yar"X9R %3Eoy?`W:Ç@u ՚CL}]k1SQ`&`O( ;X-(f(i#غ@\ZUhQy0S|ec4®{҆IXv(b܃P\i1uFx+9ijpEl̀#/][Oi734dG]v! v JY+ S{+f}H;[e3KYxoxǐ! D, ABـ 3]ER . J&w[]yYiýwth.GmKaNWܵg]5wVRqW2F0fTS'-O,!ڻ/tYS09Z$ PXnWMfafqvpv GhX͵wXNgߤȇ[?ݥ]M׏)'߽{;攧 4?/ng|,eg>\?YG*?$-5S5 Ktvߞ[Ӊfk$䙋h#*nw3MҘ_lQcJ 0&rKa* R^j`%:bE~뻳[ 621II,=UIhNNjiaDهPKH>6?zw`Rj+Qn5Zf쾩sݜ:-H}v[jq "'p)mkK1 hj X=+Qjݳk5єz7˼U|DDJS٫ygJaLx)mtW*ko(,kts4tWE̚iM;~hV'!( Uǻ.}G ϓS'Ǜ(.{;ú ;IdEG׀bD YIԽּv Q蝭_)ۺ*:9a:NQc8jU|v:uAUߠ}SL$.͞`F6CxeY/"~NW7Kʿfs`tTH!oXuh(Pz$ljODG_/q_IW*yW]49UǪj>;{\>N? RȈz>k} Mu)bHE!d md#ʲزGiw3 Kvs(+Q&H LqNt 0ƈJ0gDFD) >@ib0X 'TSei^m~o@$_"&|[ە//E/fr8< ֣%v%J_(/!AMV&;L|!ވ6/#ߎ;t4^Ȑ*mc#/ @"d,x sw0-Kᓧ-ף=%d$C5? 8ѲjЇeۨh.we^¤lrg**'ao#-FOnDDYtFȴ YIsϑ` zKNp/8C-œbWPϐœ8z$;7dq#CBfWj6o*)ɐ 34+EU)6sA/ ʝPjh5v`Uߋ ;86`j-DHy:Q#J(3T/s,̂tfEu/f}6pEDG/~f>?8RB FŒ3BI;>bT-Kp[tЫ`"wףy4Ńt`ͣ@t|_^ūSqNjw7kz5\1ux-X !"X%ل)&67$؈ְZWoǯS[ E`H)$  :-X >j*i@qa9(T!XFX0X1U`.I#4Һ}ܤG睜fL$H&L#>ъI4r!, A>Z *DHF!oРD@|C0cN9!ՆP( ШT~6H%)N7[[h$?K$1k.jyura+RR7u>ŴN)j2?h7 xͷ[Hpbr @  >N`R,Ds;cz2be(juDyD@ㄛsǹrMDzh@NȃEhxk4#,7T2璧kuH:qޮoMZA."aQ'V"3[n+a:b$06iŝ3*1HXP(01~+mC8\t2Qrp l H :ǔdJKxĘ(oZE#SK%%i,Y@2OxD%5Xفr,KobTd4h:jpkQA@E&`+\>KW \T1Jw!`SuxzA7MxwucUXhpd}T'uN!U΋ҝ 蛇aXv"ϛQQ1Z%n//\xqbAjCd7,tY:K,%vV>Kl1iIIE4N"U1 qI%*A[ @*7-@85] r2z6-PNFrfb{vV +!3X! fDuzrs\Nг|J`qؑ; I] B1rpE h!o500eO!|6|^7l>R$zo=$pꑇ+a(s@L)*EL0wnId%]w} ooI_;͖d00#BIdWCp.8qK9,=x()vqb]s9;h@$ulWeҫ%w%3لZPOBJ.ۉQ`E!f28惁I!^rJ6_ƀFr1d ׁObM>Xb4(l4iF$l?I0R xل+]֙i4 RG+ h8M0̈ALO$]K×%G`)g8V Œ" I1XDYR qIxusXCR]вB/l4rK5 z{*' nv:6Gjc^=OUs ʳhn*_@ӉwXiة|dKx1J=$ow\M t|@yE>SyFvLܔq-wms&0'!*uϼ2ܳ@ N xqA58Gb5Ҋ?!UEz25\'m8w*[%*_ R@hg^pD98?E.zn Eu;߯Ǿ(v$@owyy℉3$.!}JLk+8rʩ( WA;!Pڌ[;Xâj} ^^M}8EM .GK[nFjqA{阡L+IDm:VX`FёN]`,p-[Uh{"E#vS{&ZA$a!ݶMymDL4GF?{m/AJ&EZdM/ ΐc;eCM/)6%$rnuZ󐇇乀慖$@cV F"A%#ƼhC=T¢S`zG3#]J~xHaFl0t鱲'"RrTO_nE+aCFnt#)%F'=5z1?=L5bԠp[̴4)fC-’ L)ׇS<45hZ)NnĂG8H /!d A4> {#;++ZQ3wލ})q 2yg]4/QՉ;GJF4 o rC `DDuuׁls62Sg[uFF8r39JLS\NCzCNSP?]S P%_),a#wG{x>xᮬCo=#5=QT}%zW՝r11JÌ0tV77]],JZ6mrYriCDC,.kBxKa5I =+ZMt>—"r& Bzk/_/wˋO`>e&-׵tPZ@U8aMpV[@iJbr}qtfHkV9i kNZf["9b@疽fMńʃY*G a#rcҊK*YAZ @Cd$ ݝzu%3Jm@a$`krTrc@8i܁e Y%`̖suA `v6D$\6y!"e|O:Z(6(@)YQk W.5HjR׎,$L-S%w݁i)T:I@㊷L+Rd7CK4@q{χ {O"+Y+UDEW19f~mMVM9Hg-3&CHfTW\k 5 I@."M3e1)|f7k~"Ӊ̩A9fZhdF5VE g.#ΞNF9pnH>QRTN e5g 8!gIT!` 1HCN_rZXzj3!QnɺmSHC*F:iM$Q!ГV'/;k$u OjC:yqL[9C9PZ̢4g*`[KUlэZx_^I:&Oaoj6fSQUnDxw1eCJ$n-oŭV]X(z|<vzk}2m3ү(K{[΂wGsa_3NQw6"Q-*2fa挂mR=Ϻ'*xQ%-5촗d%!"t/xNeJi:E ٫s*u7D!Ɯw ]ݘ^]9b4 (JRA)bݱBX gHwm\zvdZJٰK뎯cXd\FFIxfy݈;j]y2MFYDJEuh!iol o>\V^U#1  ᬱ&P >&/~)"Jo!,RD6B hycPKnP?bR=?߼\S4p|ߊ9,އ-8BQp}}r1~韩dDaCD]x8{h\{]㏟S%N'`$7@EbIv*QѡT`DBE-p`= (OQR= x )h sCP!{CE==! D'N! HX6'v *[%Z݊&Jq?ZRAڭ8JUI[ReㅠK9qեRž 1yKBLQQV > |;.%~rB0dgCi;̰6C \Kp;^żezrfS+AƉq|EƎQ@pT:4̨vGE+lqdg>?|4?yv+.ScH ?`XqwgaRVJ1A8FU/&\H ( hK_/RP.{PB\6&fh_z>qD5Ř j2TJ81zJV@Nbezz CY%q@3ip}!m4H H(@W)X=:ps;?jFlj-q^&se$YD%%,jVEJ~ZSHrL 2(63VRqAE ƅe8"\T}ä`6/ Yg . q_%Vpk‚"QP]8E+!a@Y"4F!K0P\UnQk h%ANrFiȀzK[(/ el/~Yu$`EM}۠mP1vʶ#oI ?|>=2o ľzQ$5$(dn-o@Խ%wU.u8Y [Ow D|^޿|\}ãu/||~wȇuG÷Wa:x0{wOxXY眏i ],t <#w u1XbJy~]C_q,y(SY:[=i:Lތ<Ӝp1Cʸ*na:S2w>uASb-fs#~̞,~ɞw1 #Q{"e##m:zM sЮj-47d " @ $ӮǺsOj݂twDǏ(:[x_ْ=w/4spEpGeV!y]̭PUo ++V ZY.::?J}Ƌ HpC6pG; S`σ,Kzݨc>S?A`\EBfܾ9ln G|7Cv-[ozK6OpiH5P0koԟI |lUQDU1ќڜf9;oϧ%77ӲĬv4em] gQS.zW;{~y+0gSreQn1gDji(rÁ rTPL,ùeIF G{e&j1Kx$j|$D6R:mV! Q(d 3cB (G  XQg964FN7"FQԠFic@Zp#"z(I` )1H/ gBy)@i%9^S^##!#Bqhhɜ)%G+SǭUv<Eܒ6}\hb ؓ!mSjZQYlo"A~r`qc;+-!֕Ǣ.WylGFj%RHB))3F]k EX>zgz;b4R6Q02Z}rftuuó\0(xNd88Zh mB")!Zm @0~~~:$A1\2x[œjl07}5{mm n]8í@(aC4 mڭ8 ŵ`ɹj83JmgC}ؒ  :PeWWA6_win6h.x6@Wf|6Lyj՛#v9̹v Nn֮.0h#٬ AZ2bN[qϻw_?OjKTyjSp0e5"@C@iP:Sqh)!E\<;q+㓚y XL(FHa^[Jeb{>%2 5IE/@9qkuK.'@{-Ø/٧8tտ>pЋ.A1$)ъ_)=]*X2u%"<S;_n@zz@7v2[-b<\m##*iRLW R(X*H1g -5e%K/3+Q~p(l`WwW봗W-b"?SrforW7 ˑRRQ8Î\1 \Q[l(*rp]2cLju<1uF'!S*35B (VӲXU!\u%,rEgdhF֒{*&>\]n'5a8ޓuyrs6FEL6$aNo?MXJ}&HxS5Mی{,s"Kd8NE:k<آY;͍>s9=LCk̿"oל`-oc7Ѓl^bf:i[Ixշ2Puu]wNjMaߌ>eV^>l'Z:Fjv]?u-v]ySbp!|<^F?a7Rrm>{SsS{6n$ K_Q7  4*߇Kp[KX (-E[߯1Hΐ0V9H F?^ͪj7Y`>](4&ZWc۶/f Fkzmh\1C;!ұP=ğ(vuddE)k  k[jGgFDcAkS(P+?^ڑ9{(P_Q(9`a1Җ|=h_ h &ɜKPщvb[^ˌ²H㧏tdgG`܌VpNi4=} L6U̘J&1ZD-H.`jf.s`>O[ioF(nMOEykN9+<"4Oܒ5Qe \T'oӚZأ4mSN{-sR^r{p8p9bɥplq쫿Z UZp' 9r3ۢ<2c+'k#eh1F4f2 #EgԖ0G ۦ}?DMUϨ?:]9bqd:HQv3o]k+Pͧ=fLfpjϕA*ǔ,@s[ˆ0#\@wM N!ۉӯ!CLO0cי)&;bqTVsّ| K_ %.[V0|ߟmrWhPre  HB\ݧLj pg_)`pEKO܊ɥ? (`u xJɳDp mB.ԇ;?tͱ"'Цgo-cƻ1  Jʄ[͹FNxBB8fף~I%Mt?{5oWlV7҇o !_)RE*5^RE*5^4KWO2AikZJ:nb$b(, AAAWSW_fonoճ!5TiYO#e}lKhr|l,嚟h,0 6~I/_45z#p :KIyKM%IA Ʉ5j]]ߛSԔLY. !ז\J+j6h[cq3Ĵfm(l~-a 8jQFɦW:B/*k mX)A)#cIR<@EDZWmE-gvjW =SxrI*`UiVaUF(*nK;%wNʧHn`%`ialѽZI$3k;ܤ+,lN8sa$!|+(YE1Q (֟gj [統( j5>L'ҍ88mV99NV+82F#J|qI@ .dl 0Ť[ M29^0' o,{o29igиŻq-F8[?Yv{>1׽wm}HN ͱn#neX!TiMa/(ms [sè#h'ppŊ"(YUUT:qs)IYԥ~_b; F \aU9-VR5WPf庌}"궲5#hȊj&2U 5d-֖$>!LÒ7%wT ,*m9c܄\>AoL|"&sB

QBPV6eU@)La1”12/4 /;("O8|‚g'RC{1!YR}qʨEI<Ҽs҂*F!ϓ7Тw%bTr+(`bƩdQsYSDCt0h2PL[q01M'54Tr(r7EY Ι\jT{v@qe'5LN8{~N9_jSEK>4/Mqp7ŧnF1T%BYO:NVkƇt$ LQjA N0Jo !o5t}u7+S6g>n)uHEJ,P1pSTJT)EZ)Keh+=+Afv=|o6i4;E~>kBCkr^ZעjB0YՁŀ܊ E@4L1RZ%L-JY2MuᮇHΈgcBr3$UrbɉY$'fEӉiqYbA#e"F %Qkp*eu)1ԉ?d&KrҤl\яlݛqls;A=sK廰xdذ$w{YOzrڌl7I뼁O&qyxx!=X~7fX)0%.[~ke5?W _X|%,Sh\9 s>0H]-~b䑫mB9¿h!G_X>kTF?z4uڛ4Y fxKM2z))(ǟi |irsMD4Uԙ$ߦ&;Coy.Ϗ|S|wk_\@.#7yqNr^_UGhu\ZlpTJq+>nJg!6;ne=񬶖܋B0pseIqbUGwJG/QFْ+X2WRG5-rrV֏(]R+HIMO3e*ASKщ1v>'"\M-CsY B{'*Z9UiTT)pB&OYMC6UҨYzk%92D3 }rYi1@)]SC)=;fҔn{Byc9FA&0N5S29|06-˝BQ> $_zg6 a"*p!`gSߵdx(94\H c0[:Y^0H 7LMI#Y$-nk+IdN͵ .i :>tcPD=a b`3Y"iXiiQc!*bH̪ʪ uuuRMCyNTgS݁l+:'wg cQQXԝ51-4Aoa U/wĭifw]wi3Zm>~ҟ/j~~*+ -~I??XBu3fґ?j*G/ZzqVO{0&P& ]i<#96)VyQ2,I>15ه(6 ;V2(N"14$Y+l$QcOQC eYF":9IIi@Jp;)alqu{ +f)ɵ#'HH##m('Y锌,42DznJ턶Hw$X=Ug$vxr4ۈ?x4#~ =-ְ9l՚r?Fd-Rdvv>EI8DT dH\K֚[<7<@4r;pRD}Zuf T`5[ _`m Dgkj1E\C]FIdt&zNn:Pg{|\{܆Ō1M\sy|RGb畕~WMByVWo?O/aLq83[&(꒢g[,60(Q;Ȳ'zR jOL`$YGӷ$:{{'L0ӏ,T%;!{7;WbpV޿N>eqM 7-TJWԯR|tοup ȴqY˻ٽr+KXU|X>ʧ[pHA9b K.8j %?nocq1,Hݎ ~rpu?7Vwj9lK:د֏jX.l|YW&"~qg~VeU/v*&8vrol$erWx ?Ϥ}엲^6GjbٰrC1ҳ%gjջa'~>>QqȤ!&蔨+68ZLGl 4׻oPb*FDʿ-sF ْt… {<5䑖Ja & 7 jDZ=jlt|>@A0)E)!;$7Oo]"aϭK$PFL(F'oV˕E@Dv%C!7rK§'c(XPd- V605煭]q(Q#Fb1HF|Kѻ~ ,8|~A(gDi䴗7L#ߥY)U D]9t9]bр 9au 'X|2g|ﭳwJ*˘ĖIn@.( d SJ.z.xC(u }>?p#bTWґ3BYBJR(r(%GUIg@#.牁+]%sM)enWFgɈe~.(M6Qx=-9]d\\77'26m_ksE=젃o/ }j!o1x%o]]m̟ڳD(uݰ'2gvNG\^,.nwŲ>VdܞPH|cN#2сZ}DRxDiΡ7q)X] |c+oDIA N< Z z84I{#F$pF=W&[&HZ@d}'iPH=QriAasqc`yxX*Aev5C>j7b(Qeuv5}) & Vm&v# [ْK$܄FaYYv PThs) Rp%&ӦN$x;Q/SΔ$x};„O#;ꅲK<:h#w.ZkcNb* qU95Hk0WIL)CĿs*yqЖP&:eHIf+l҆ UҒ)%pGBD8A/7mWf/x깺ϩR#bJq*'@aGB" ijk`?uYb)ԕb@/3O1.3JX7 , J\!d:[iBX%8 kLN:L0] ؘVifQ(H`$sSBp~@fyaXcS0ؑܺǂ"q9ԇK{=k"KwM2s,/b6[iUwuZϚŞ^GlXu"A{a= !f p2' 2AUm+(BA+.&Dg | P}ƽ)纜Mܧ?] Wץ 6} Ljcu'UdD7Ʌ1< Z2b<+#"l1ٖ- !1(Z@=)AK dO a!-ɘ!>Q2Jl-#0^ |c̐R1{IWP{ phFLqR+h#Pj\f|N"c(TR2YK;ȼV4Sf##9Qšcz<`o[Lp1 *a@B)o#\+| AX.He;z,]*!Rn@M6]K7-rhg :J (FL%uC buS$w]򩀷7 x_ Ԉ|xKW#Mes#}g׋9x7Y7z[ +LfhVR%;0_ޅopݼr̀2 + i:_V3ʻyfek|83)-њ)_n%zl=|X<\Hzn kM.yw\XaW֖\xw"Ȑ^(BGѹl?J{çHzkfnįe+,XGHyRO:(G!6g'[ߞD?٫zaS*BjSLhz5/A;CUӮ+;B7bӡK<\I+N1Q8ﲮ3D(#Haf)Y'҂])¬|E7IÕ 6H (D$@YHg:W/?@ky>#Nt!^"BIBJ&$gΓK0i%!AH仑:U4 /7le1TJeJS2uN2b?hR)>s&SP&3C;Lu{munh{yj- ەFK|bR?11%>@Nv]3y%>shgI:ոM޷nRq*!6:UX-hrhgu PNRYgoZBr3ĵ5TD+yaΖG!p餫<T8Ǻ(sɊYĉ`i(ڃn Ti.Jjl"HbanIH1K7)WAS?u2 /X)}AWMz}⍅:' C)*O4ŋ~F`}OŸ!O5?Uא5 I(DF>F˳ز?K֦7oXah6zj$Fr>_XHU3S*\^-~MD@q|nOg;c*Oe~C-".R^P-D g,Җ#dUƃ,զ-Ki\ޮtys5~%|0*LUVě7C-)~gXa=v2,@eX[&QǼ23kҼG_fU08v*L^&5Vlk-; J}$YE$ݲI$`p Irfu[`tV)ewl2+xO3W0DCtЅą`=zs%o#6w45F1P+i\GpTir; z=lLXSX M PR(]aÅ!5H5;.|25R@Nd;+Uw67TzRNjk׿SZ.!^e0aoR{cǓ\[jVUqAK Sc9B*Bft?8 q1jY8)/}$z0 qsX >R! 1=7BH%E\̶ 1Ao9F0Q*A7!=} =踷PQBsj7Ȋ-s<ٚ(Iwv;w ¸=2.|iDrQDMaY[*t1hbJ8I%z*kx^RQDuNcDc랝+%k8UU ˖7NkrU4"h)&c]:Oە:-XyZ0ca4FnS<̖˘m|u%xD>9mA%VJu.8I2{CJiٴwۣ:29Db%O<cRv\d&iuR ^ ĥ>j(}o;! g3wW2LhEl߁JIA>#vdގoGVmaf TtFp,0"" kK*Va25.[%[+4;`!.ZЛƾ䇋6=_U2uO{W)yʳ ~Rc[χQ:Q:~SG'c!HP!+uI :nWֳDgl6W| 5([8I Z. vO n̦AO>0b,҂U?w]]ϧwg?Ϯ9NG/PR`&܏d0wO,o76t?K8Qzq {뿼de[?[~D*AT%hJ%0da0{Kg%z+G^SO2g]@Ej8 U(Z,eyLVIR}&p!|ё4Hp<Ҁt^LV7P!88H1+pdH&9L888%QDuQAG:_+|wЌVdeAϚgy (%)/1pZ*j=^OPRafuEA"( Q>yi0 ~z0߿ "_o>?d<"K`{1xss=}w"E9Q7gB̿'~G9aZQov 8I 01w08]G ,C࡮J OIĉPx>pSřAW#JdQ}pjXB,JrM%*i5R0[E  B!gɫ"P+DtZ (2Ȣ kɾAA| $5D!%tC~Q񈾫ҕ^9JfQە u@SbP.iR.(.] KTi>[[5h:䐰o H=D`JyJ'煬«-^ !y!t< AP:qTT'H>5C8a{DjO!"jA`yp"T'UBDZ9m 'b " 1>p"(1@rvsHY r8Ռ0Tz<Ixo1IG>ː|2Ъ}VPfк Zߺ6Zp h}f[[7(U1QTb,lP5L7y(!<p[-mZI/^AVrq[F^ ~>7x*0x0ǻ.̨$(Q0ba$S(0ﺞf/ݸ8?gGi$7y"pQȃ'&F j72oqŒ~\9T*D!yأg20W+ p!R(Z/8#84x;~.&sZS|>f6 YMqasMg)6 tXYM&h0-(Ӣt"y尗B( fZHʂQ^+Gb9X Rr& RZ"!2݅Vοk7^9uv? oJh)Q`󆵧<[E1􏆕fo*HQwa}ව py v%5=7xr8\A%x!©iCrR;B hۮS4/5#k蛅L'm~"ע]1yD].1m)$.[4rT2Sr1WvV.۳"I}0~LSCjΥ ;?nnw+H[i]T]%o?QV廜fB8aS 墖sK`46h9{& =VtѦ8T2D{0 Elg Kiߦ[wAp!6(ˡJ)N 2ǃ#KwLw+!DUIKY<Ilsr& 2ڲ̙핣1; ,ND_)bgE&$§ UV@">-C`ЫQ49A+ U /`áHX4/*#|l '$4Nݿ'{)j7{ךg,t^EZK;#\LRmP^f\P)r*Ѿ8`٤MB#-qNcBhӳ0H6ǪZ#Czm4UڼrLlUƼj(,q,/X 6xv{1x . >_~x/?\;_7DΥQ$E^xZ˚fTcrdubˢ@8 7wN}^z$aeM 51GDhA!c#B8t֚ɂu 1OW,R': /z(DwyXH{8B*8^.`|g,,^I8]R2<] Bo fO (VjTiݍobwnZhpHa$V[AyMp Ɔ(t@(>״N?˻;8 wu'|l;Kjb6gn2_9K~9'?XBꎻ8'aâW>Pϛ1.f{ЬDXM8b-_nw_gcC,0 J H <(0/8l5u8-8kZ_g{gSU)A矊g1z럹:?9N婶ym}{:thp4%`հ?{䶍|m(+ۛںd|JXiV'ޭ߯Aj$JF; l/4+ģ|;oL&t"c8x(7Trٌd*1隣 A^V<;#+hx-"̺faC;r{= vY V+OϟO(i$G!"^NC|q6Zd0M@`_Boh0r-?nP6'S][;7u*H^H˭>3 f2 soڹL0 J;8JA 3sX'H)6&AɁGOEG 壃t:v0%z: ӑH(ٞ(=1ŁrIRe-%!KX x \0CR)u#I\*W糍Ӯ_v뗄]Z}&A7B|5S9I~ZLٙbL;;/(X==p"6;91Y-y74z0 LL$$| |r<K" 36şW~Aw+x~\N&| 6&vm=x > I[]X5(7E$o45x+vj^:'6!˫2-l9YB0Y^OpnfiGpa.&V?../77./1vw/oE,aAw ݆5/.T:͵S6k^T2lO ;)h\7 ֥)~QbʇiGBEdX3"k8JFlOEr*:¬05KWǒF) O !5DA Z#jjB on*x`6spY ~w鴆㽪KN= $(? ٮGP/2b:~SoFHˍxHOqħ?܆sM2;u_\~/F2ϧQD >s)wQ *.19KAdZ w7#ϗEgn`/p$/aJ-za}!-Vov6flߢF.@ujc=ڊDmilSW lO4\ ߺ#*+X "Mͯ]qP̴_צ܏&uyUiI<4"lu/cPwg@`h12'`Ujb–R"}:r9wP"FL>ْOsE"u%'N(aԁwE%Ǣƨ/Z-LuENXOWi諺Q/z*{Zjo:nj񄅼 3WÄ(#2bļq"K`@!gXD Cd$8Rjot9:ur#[UG"$4gp&V5jIDS@+B4b[0JΑZBbWPaIJTTZ.rxrpWD؉aո&59Ӱƪ;d#!knN9F@{+r7s g\-bt1 HЪW m^KVz4X #]gc{%Rʫ!x9) EZBj 4n 6[tAwd6pX SٝTo\hX{Lw{ڄ2.e1.$EsڐQnDvb=c?׌bzC{io 8urUpŔ3#?pp-r^`%I2d%RqYn0ZgGiYdK2[0@zO6 ߬Z~_ |>iGBﱢ't3,f۽lm l&y2vk`7 llN?qڍ3*rͥZlm]B;RO!h[QLmݵGRu=[>[V9ɮOA%1[5+W"I&0."Z9֦1d!xK1>5w>,SӥF!/% @ϐ E aTWa{Cq1üY±bP4($Wc"E t˫kAG~z-X͏cۛq->cGwEYʪ=͠KB uƒU~R$)) DnE1#m G|_e4[?OpW@h>~I1)' f<`YYq$2sR#oL&tŒ5/G;PvPlKx0;[[1brO.fҢ{c;޾*5V>}6*dh%J1E<1_E3ZZ"~ֈ!"#z3*2g-w\4KOR闭UI2]ָycR |Rtұ.gYRO,+__f!n6V:_r{rR 3|K؜ v(;C.3;CΒaJ+HeJ>Ǩk 8:#t H ;bXNKreBu͞6 ~wU;רQBwmW{gtM=/b&|GgXX*cZxzƸպg(aLyg\G{gd }kQ(֗d8753YWzwZzww jgcl?n1s~A=ػnݣ U#{D<]ΞKHwԬA՝NUxÛ-uY/܅j<:AxzNӢukRH\Pc)$o;{9gVlv6b"K<fZ]@@tك'BzNq1,yvL?3)'ul~~}񇷃<;_A*CpΔ&Qhf\c`'Я4Ya]7 okIpL{3t\c9ɴyR*XnLfF(0eTpa"|Ȕ׭rj}y1tV^ok\#%ޞ!@|y/*z(,(>ז@J]:\Pk 5nv:2yt5k@;Ruͮmķ{$v=uh2Z$[.mxG9Qr(t&ʗN^|i%ԝ\,)5n E9q(u>=#v|7b`F{3Hn5>H3OE ̦=t|L'RmecRT@:e@2C-4_yd@`j~iUdO&h?9S/X6#Jz3 nD;C_8d>Rs6TtjKpз9ߛϴ]"9< R]WDk<6嬛4LHk't9&dt1JǐsĖb Z e+3f\&#t;֖JdMVRFu;THkO)$S$)+knHb`#2Ha;/`) ,Z'7G"ݍKlY_fQix搠=J%\ ~ Qx i9q,dB)?JĤՒk/ bg6[[ dTE_(j4e3/M#.bdX+陗p^uDwo;h3TSHL\y'rķƤpSF˫'*gvųc}`B{Ճ_(C:YF\/cŹRND )hN}"9+Ҝ_'idfhog~͋I%ʹݣ"D6mIrQP.O.GCvT͘^0Uɗ Tg7S5e0C9͜/F5})}K_a]V!׸m) %Mapsu8'evBm/Im #"T(c?>WoI6沤~U+|E^3,ʾ]fF]wM}^z$aeM #"M4VӠ1Fsnq!XpuB-۝KuwKľrMYː_P3)Al?:'LL\NO\Q{9w9>s%!m3ӄ!J |͘K^"&L,J/ɢx!Fmۚ~|"s@Xr&a4BXyz">CiT_@Y:-Wjq=}wsJڞzXjJψ2Io0G[)9P@R j@w6 f4{~tn:[7O$>!L /J/ܛwzk/ )14RsA)T<64y_/4ejG=" pB $.'aƓ"amo1 Bx7Jb<4~s{q|l?{v99=Ӧ/T(tO0'0^7ט.?&ܼ?pu;7,&Q૦Oc`Peفr[d%poQ]>ګ$2W3"dϱ%q R x [J.8؟?I4s\5MxA(ݨ-8#K"ܫ$osnL%鱘ŗV`Je5=e1"yl3Gk ]I T)904[A1l Xea>5XE4J 0<# clEN~PŝjX r ވgL2IoxCq-KE`s }X8ɿ~+΋7˼Vbk<} (.<1(\#J\ _UnR*_n6|=2y^ކM1v.0}͂$vo夗# _'SX9) Vn~%{[%Xvq3~Fu0 O7yo]w=.>|htg˥̞>ҪM̞whj7C`/4J}(] M~ `}p@KjIR-3Քo[ޙ8LӸdN\<&!! N $L٘%d$8VLj ̂*c1t2@cXA".9;-CS:v2cv QǪsXg?'x_=qpu Ⱦ~Jf}G_v& ~|/~%U [q/*|v~iسG\ 4y%PU2^`p,.*P%4z_eP5. #Nxy%ԅ>lȯإx hz/6PUeT/*mWWxcۯp7A& W++,>L%RB^DlA[PQysRTG:&ծ% -<0l$]I]$QwIU FIDJUS䣳#hWhN["&n7"Q`c bM_Xӗ[Űl)q]ls*wbH+EsԋBDĬUJ>B`RҞbJ+m,a8O3N(SFFY$I) NB"2 \Z"ZE "% \iP`@2rU0PI^Q Zs򢯌bfQaIη봥؛|S}+m0۴c{9Ǟd|dA] /Ye.LA;)qc:}>(|'/8~[1 FQ3 c [n0r@;M$3XXRjlkǴYIVPQ]!,]*"\rL8tW4_v)UN')պ=*VlmپK.TN#0nC=zU -8k5:i&0nYK1J5Om |~ܼep'ЊwW0f)-Uk/v򫻝wyC_})F )cVd+6\Ls*Gԯ<ߊ1E (uGH8wj|(O\BUzuE+wE Q1 'u 9r042I]=ӢDIӸsk] tL!mVA\I vh,6P)i03 kaRڹTN=CauJ"ZfrUiTc$}QGZ "fj .JPE= C!1D(T@71ALwRGn2f1yk]Ha*7nIQސ7;v%rj9gT^ROSLNddğ1; h<);)K;T.UDRL5xԂ>3/EbJpfu<)?==>XibK|J,ܵt 8J]y5o:w㠜UҺDJO$ iєZ*-L -gj9@(E0X9L$8rԀj >؃^)`b^MieT"  FvT0"JG^ `<7"URV]D)Ul9mToTJ_!=^.Z㙞uHTg \+2R8Am_ɓkg7xOn:>4`VtL:]((OIA XON5IԨ\,;_p}K?/s%epyN6W+} +$+F2yX}.P5Š4}><]DFѴ[VnmHW.d8׵,!9vkAi})->vk<Ѫڭ ELiB-%uob% nW<*#lcf w3务[) 3SC e6 8 T9֦qէ:T+NBk Q`UF@KTA )pЄ:ODE9#PZ)jQyl4H)T4h"L*<L@ 5jb.x0p1cKy<2wswORZjŕ571$2 D$ݜ?PI((=r1Čt Ki,ĀeڃPȨG0 Ro, 酰 [هjw~t!"@cpLEp%+:0+Q g.rی$j+,jR ɈBXJЅ2t8RBQa`R xSjԶK؛Oׁ\:CV7ϟ m tçr.W Uztѡ N知#DCxX)`y|!Ϩ;YRn u*ӹd׈%rGdI)l`IUʩYvnĔRz $eMK3 Y0&.,KdS6dOOFM y!E[)RfjQ47"Vٿs Ue&EbE?- 25Ȋ"XfxO8L3Fɬ !S#|jg6-ҥ*o,Ԯv.~v bTϸē-8TH!KZJ2cU1XYSn>qu/%QGs~Q۹Pޢ:]O3IנfPB TbED0Oe[ޱĘ][SU޹S{N5FiAAB}dTYZXJ<} "thTbOX"|So8V[?\w3Klpq͘, Sw8eaMv]Yi$o@gRKk[AVY,Ѻ x $U rR>X nP6?zZeԧ̤֗b=3W\jOSJ6FkMv'U";P[#kNrRZ0wO#xyބ1cʁQ;gFXebJdddڤtZ!i[Yթ{qoj!bt5ӑk!iVyhD-V[!ʒ6HUW&2ɉ3"jֹ5!q㸟#'Ij?9N|ry~g/{;{ pNMR)q9$fwIZ+w_J'&5؝!RΙ֠#,dZBgbӞ Fo0n@&ɽedm[Ue"YqaP9DZoEUOF6١AڜFhnOR @3`xw$1 .-n%_b@N Of]x3whhF=S\[1tӀB5`uۀ#ĺm@mmv$xo tSv<eu:0ϟ_0IА\L-GH8!$GoAb4osלݲgV{=Qk4kx&[G7_p.l=kFuU3` Cg`@Gp2{J=gJ: $:wWl kĽ[ 5i..Qxh/ţ;~pțr R<5'l ͜3Ô9 ШY|]m0Z턃LEv 2)rY ;s<<@ͅ`,%}e2j5Ӏ{ Nx;iԽ,6R#Q:veR1ast4"W[ 79 4MFcꑪVxi*AXh=>.*ҽf6U%'觾FOb0 )\2.Z:HGqΖp?2cbhPf1]U3,|FjMV1!`~Bg9}̼Es1"I5H[&:87RC7J)B㘆'**7\~G 4Cz7֚ >=릺W֊L^ %z՚Պ>_Bۆ7 Ϋރ 9y59} \[X:T5gLT5*zYYpP ^~NF}vldB Pc"{KR*ZjO[Nؚ4$Oz.Dn[cݗJdҪhL#R9rp |A;dT RӨ#c,9l%%3`Rd 6EV,JiجEisЬƚ̽:ZXr' ϫeQcoYY.~wU\Q醨O1xhEKu-fXjPvbW{w=s.o)6fo߼68$ehj{! G^10;J`Fmů6TǠk8ZaʊTR堙qQV,j: 4^Q/wYh*Y[dŪNޛNG3xgL3AZk2D|s4 7WU^ f{Dc+i# Qv\΃d9DxMҜNe&r.R {)^)+ r{ApƆxg`# WJn 2 C9Kb@UJҌ3t)nC2RslLr(ucmE] Pu@qITJUNrԼ~Yn22sDZtͣܓj<@ &̋AhL|kP'LˎjA*05=f,x_$j8o]?=|<4FFL,[1eq#g[<.of_ܧ^˧9%jas:Al7 )C+4L0 \+kcК!6Oľ4C ]&'ǣU[OOB[VYU\?yR-IF j{xO6{jypOuը5cu a򽍧޺iHYsOը>δj# 6,uxݭ K16%Vx`~֧*F b2zHn8DbMk8rL'=$i3sd8gQ(l#~ڎZD  uQ3m *-z&%0޹gR}vԖYk$MޢDJ^NR-u\Q}hWrҭ[MiRB'ceoK}s=vRʩ:e Ja/ejuYʼnjGKMv6ELMh JB2$qr4 Q lj UZZIX(Q;%+T#4y@H-:=qS0npZ z?'*1P0LnN IR)``I[v;H)8?JO0%$=HoNx3,ח'Y% F顮WDj4>=+AnLP5 "}ZOüB`C͂GHB~Omd?؋xH2xp~7&:M>X(JqvdUh0PvD`Z2+'[ :; kaj)&u% {FJ N÷zƐB>iÑoj+p7[XbͰv6$AfaZh=~RΥwQ?tJשNhx$D1=||w泻s}߬?&guY+q28Lh'H z!V\*c 8F]0kwI)=NZAA;zwV62[2td+-Dmm)XZe,u E0mcWNR@=LvL荗; plaAz> gQӤ2iHT+fN Pǚq&zܟҟNY j8KY~Ucj{spH6t%i>7s>R,ko[کCuwВl!)e=.vnmW+wm3@ }?KmX,{dмA>-Wfm1Xҟ08JGC-GZU.r9/@K+[({тnhoyGE4k#Z;GwxO@.ds w<ٌ|ׄ#yĚp$\F=1u(ީF],HIxfRtWϥ'u9ŭ$Zi.Γ輢'poM$l!]pw(S}l۾%3ރFq\Z}Hu7՗d95[?E.Z?}Wgr>HV?~e qnWaĤ,9}ݻK1=_?0E}uxXcc$rL/>C[im Akxu}$9訯'ӷԊRe] + l̶ r>17 $?C~wK-9|d_Jf8wi0bʨSBu&a*ld6^nR:ҙLaaҋi$&0 e`~MMYaGD\zw׋uUH[ѳ6]/~됷%tI'fΰ3+XwnHtf-z)fV'o.-FVL\NOK֜Bj^J2?}_5-X4PqcŐ屯fJv=Aafj^R0mFu۴9hlu{ Sd%m4]in/( hpSm"}1g ,4qtK*./R]Z: ^z*;~<AOAy^PO4.;F<;L'꽺H޵q#eT !@,{0s@Ӓ%- Oh1+36 KUf̈9 siYxm4餤x 1HƄ8cɔ`P>bbL@gq^Bx4 Gl› 9 *[Y>јrE$U>X5a49z~l-F +U<4GkRA̳6q.'ɸh^I&CeZNi"x7-;;|p`&{60s)GE|K_`DdQst ;9 -kڳPO6U_C}uاn_o޾;6J%Ieu)#p]HɁ/kE*|3? E@Z;f[n{ZMĔ[}GՃw]6w?hأ}#aBn2|ZSdk`ikqqkS9ww{f?ǼGK *<'9z~Pi{X/;S O{^C GPA~:ғY"CTHs #!Z Of-:]Sgt8sj1S7*>tۜw ιw !㣟kB0&}eķkoqtm=>4jpD=[C#R+bvVw~vz..MJu?֠GWn1C5jmW7D8to6v7m>>Wgzt~mN N' Ppr3% p A)ً: Y@P#!]cU'HLnK&ya8Іm;s}:_̧I6X8@ҋUQhCU{q_E[0NrgCOE:"ԽZ{#R/?h} xHG$׽|WӀ"t􇔒@|x,}ߜ-ir{g}:9^ w TZo{>em-&++FMf/&n*S~7F-Oif"t_.AMe/R;T7UVvs8 #=|Η`ځ~92oL4>(12/=hktR=D ЖbU! $mE e@XR%)CmJmOJvy)_߬dע2uV7#ܗwpqjW+Fɇ-UW?/:#?]K|D RⲔ #Cy|s蕖w~Hqޯ rzM_^09UjbI S Zs{4-=JQhrH)]Q,fXۛn~ צhk?g_k^]\^_}X./^=s<퇷^_^=uiS}g:xV Gbjp-FO1:$m|\IA:d؊ ^,Yc&IU?[9EI 'F$AKrFkhr4ʻRYaC3N+v e[Bbu6K&QH~"8 4xNxGI =J AIψfq66 r{7}T['aUጱҸ6J1 */j%XP<2ž^G 1yhD\xbq ?JN3Gz~-#1yrۮ![[-% ",nsb)|k #0yApD*y}%* aydOe1&ik]||p$PvQoa;k~X\jZ@PlF3yh|;rtý^1I: L)JɨU,9kd G(NIfU Ve åA&u4Y=,2,;D}cIE<@j> 6%:V٬!R  9e-8De (%9B$C>Bf*͒#ڭ^:Ϋ+("EbyGBSPœEr9:5%!@퉍SdsM];dKmۤ=Ljg!L GvƗ6 $X֋;2H*~ pSwȁC5*c\z_<$y[W~s'2I8\#ygIAFgy! o$i, DY"ȞW*ztU˧J{W/ӇDMl%'/Ԇ_?^7풋ܧgӳVo@GJҺ`Q DU# HyB7Vqw'Fp1[axuG{vd{1gȵAޯ "YbJnBq`z؟>?x􇻀fvRž;Ԧ/ӯFD79>?<m G3#77{-G E&r߮i9hxKz"^*Tx16.xm)G8Mu u1ƔHR,Nڥ d^[SP)у'RP\Wfs\ EZ~Cou+騠k PAr4 7 آmBɶKE)h+[3*H*0DF3@jxow^a#ArB8 #f HGyJ)aE%QOBFmJAMVXDh9& Y6@bIaGFw,ےԨl4=_(Vw%h:p4u800}tCfS0f5Ͱq'ءNzA0d*iN"9ewk4B1CKw:WpcD$uiZ>Є#x-G.gc̸+S*^rpO2{]"5&cZ GVIrJO+c.')\|d&apͶ60/eE]"OP ~z>2Z~GPBNRǀU~-'JBNS5%1NM#~m 6C euqY޷Ya\È;; cs#Hrقۈ)Lg4[fnf}mzGP&Ac-}ҧYj {1? m<[0N\C'Hnѹ~N# 2S7}GWTrΨ;IcXmMk]kᴼ6gj!2gZ(v6l5J ˩0>trNxh1F!^M5uԪ7ben>@U 4E70?Yj?#].0aSƠbqR ֔"p&섋)ɐW;2(qo./INqp6aʫ1r硫oGc)X5햇vkJy^Ly2Yղ[4-D4jb KD x}X`z%Z[[;9-`T[[p"~d FuS@g(!ex&}dIQ%dДsP-xJpr8%Hg&(@tT$kYO^2L$sgżlY'ۍ_j$|M , x+ DڲmO/^9Ryi 8 ;b]uW꫉yjO c]ƻ&@N]-1=pa:s QaĻL_FZq]oSGSM~{ejCTI?Ls |L$4);-U;[4)e4}P*})Q"<Ġ)S MdE#++U$I-`_&I/ c{#蔺$vy?l$$:hYC=StZ:=wJ ^1dDIEljZcHlC厙q!)%qdQe =i%bT#PCfUt֗>+_J<.VUy}dio 0Lh;?Pç;DS@JGwWZlRܽyZRRRD[F/RK-(ǀXhyۻe׾dF 7J*4 *ݍC;.N7ֺ =F@1?h?Ϗ…`9zP:4FƤŁwm#Yha˖KR2@.IYE)G6||n2/J5F~kf}R9 @Y:ypq90 싂zuB;#>^q7th߯ѵBM0jvE++tP\\<$+DV(J@~M^$Z롐~JoG2@.z|^hAtLʣ}y~vsmTG$]4 c :JBǩ7dVJ(U+QGXL_n1;!tl45ݚK?ͺbp&Kz7jiܻ;DŽ ЫqIbY2'w!Wڅz()#qTlN%y]t&nO,=m: ~o~K/ ݲE: VXn4*GZQJY2R&iLE\ ݟ(D=&>T 5TQZ\KL4).{,jTKe ܕ)G.NmN~n*dZ)R4Wo+?=,\;/I0_ |1i:~t/+3OTlyXvð ?hfuDPdp]M,2'QVL'&0(In[x1|\]hTO}ųf1Lʽ >MO|߇{EF:4r)\5ULW9R~j7{e2'T K㞜Q'= ۭ`i;!UH n'Rs}.0H^[/]J{"?kVbT5q?(p_9C=ŷ٢ą6c4k9/ue+Mf ħE>^j :j:tSڲw\սT?uOYsU+OoEMr |U|`wTb>[Xmu{]Qjr^Ws+)_!7v*@_>'D5w]ϬTP-+_ %B_w?4iĘp؅.YQB&אxI5l cӞlr9K&e[錡8J4:TP@v\-^;% E9)/IÎԚ+)dR36+lN\GֺTNy_ΚUGXVoݶfx&}'3=ִ c^C^ ߲N&R'Iso^ уWE"L<)Nv0@L&-U -+g/CEvc)HOKPWku4ϩ9Ei:WjFʭvusruYq+j#ASEem.ΆI}cJ 1̤ ̥Z @Ҕ1T F9 >R1̭`0Ye:C'S f%Oʗ]#/RcŌV] S؀d|f)Qk&wkj54 hR2uR{Sp↾V Y»QE/hnKp cK^ˮ\Ttp5?/ixCFKCL.z1NPoG-o_}dD'#jD A`ΉTtwc"h<03bSWzPBӥ]'F|iӝwHŃe[ ^KRJ]a\)vFDjגZR#"2z NX"% M-JbE[:bmG z>ZsԌyjzw!I"JpQj) a!eP}*hP1q+F+xtASc00r3t E!8t3hN$޸`d)|ZJ+dG 1DR`rw^RUL椀ZWEF$D@#hΕJ2"8-(Aa˲Vg.C #Q̲"fLqG =K?5⒦~\r]DNvN`r<.x9Fpi4ɍK&ErzO{/\%Mp&-jQ2,[1YIaSP !v(i$>cFHnjd{6нݕW^W߽z~vVAJd:^-|:ci@}@._-JIO 3@[Wj1hyr6&7Dޭ`*6Xg !H.s1~1#j" v~a/"N}Xl31֊܎Ԓяpd>LCF4XSܫJwP]xIOQAh3} ;CCh6x﹍##5_Lx8&<{z謁b=;2#Yb}n(v ؿY7E9ګ3sN/V8{w&iPwO02 ؉=e-=AT%lŖPmPC_A`k+=8մmb[SQ`kV-9_ڴV5H^:cYm'wh^P`^*d%Nߜ/Jw>}/%Wyrm~,_c惡x:-z(DwӣY4fIX%?)f"OyXɛC7Np[}%>!_ҩVc]VfPwnfvm݆Z14+WA:=e[InnCeP:CǺrAZʽY hG KŐ^@GQ^9ϗdN^ty=*"C4Yg0}{mq7g׷7|ߍ~O{=*:>z=w$'1jxw2^W>K~I+$Mu)^' gj&^Q*V&M4&k?Y9`(RcUU/R o8O락x9%? + 10&$=.q5Aٺi2(Ps0+}M[7a(!f&V"h`y"3@*eFX( /0>f v~;oj{$UG퀒ٛ/?uCLH*oSGN@%v:J FmxRk)Rb< $ץm22p57bi<31&TեowP`Ƥb~Êlg~n"Y~wKu2toi)VkBuZ]I>iRNeZ)k)-}.IK_z%a6T@++M88J2#8Tηs+ڦ-Y# dH4E>Wdezy7z7KuEȍQ*4O ώn>,QHiQjB! 6r),l%+ge]Yb0NLSV_%d!r8] ĤoB租O%z,b|Mr&HGW1#3QIA@R岪0f}^Udu* k8|>_Zb~'%z-eFs ƻv5|BVpRJl\['B9$$wʯeo㧴&7W?JصoŧËN-U`CZ1ZV?wWRQ2Rh9$ܙ=J 4%qPeg1F/v#ְa^Nrc ]B~X2r.clt6gMdOwON4j0wD!sI>r-SͧWK;=D\96Te& cן1ZW"Le*8s0."[Pvj ާPK )R4iB;ZLKUP!poYNP]UOs~i,ZaJ%7du9LXI5/*0^9Y!'հ)s|7΀&.?=OyFFsuH{o҇4o' WnBϫ#ym!BOe]ūY.o3 (Ǵx*Mc !Q} ''X%qg+n2XC&gZ%[e?OYv6">@l +i~slgeݨQo!-#H݈<6a_>"R.jp5x@ &jgy/jIe+v%ӚQ"~.~Z JQ@ٲ'[۽mwci !,P1rGBa94 )Qd.'ZqHVRa\Ik.X ]N+IgqxWzB-=MXTJZ{ˢXnd!kW` dL\>SFk+;۷(l`6uύI~gcG T3fiLCr\1bk_E4|8JzT\ jsW%Y+nY(:9Iq۽(gX m(@;C$-{c9!ߙ8Ju^k'[ Z+NMk?3O+)sEKk|xcohyۛ{q5=CTK˞3iK4yRrPqAl*1O}QX̀I&St1  7H>]Y>\Od*+K/9_g$==/b ^iŧbw>{=)  -W}L)^9>kr>.kF?тG˽M^pjz9y]Ey>a~+RN0,_=X[eɯ.Jw)Bp|\ ¬Yb_[EDB(zx[ӧwuJ]_=#cx}_n]+f:]u0o|m4S֊\=,-u=Y)d!D;_׽nѻuŠt~w;bE:ygޭ y&ɦ=sͻIZNuŠt~w;.BR![jBևr])UhSj]^K8ϰ~9}}]s7Z<8F-nw黫Knw9Pe!h|ږzfYƻZ$.&'ڑ91,cџ.>p^0P$4nglv;rFwl]fhhc5{%I9.~7ioe'?rzxM}x`Cq#dTdLŠtc'97qlLmպZ慰wW#\B5+F}Cvd}c6~\8|XNODY䳻\ mϿ誋'w:-ހp~bq-R 8y|;$Ihx%vE%oض4%ɣ0iSUF$Wm+$h "]eE=KvP*{ T`RɄ҈JrC&ɧ_Lu"*g{JO/%FdPsZHcx2ސi!3: Tv>~!_)LDjN5B8@lpE dM j 4.(Ҿyצ O4Rv/V(n_Zد:Y\'pPo mڂ[P t)f^FԂ2ir2%.Up(&* /LU; 4vO0nh-f@W( .vdd)#2SX#Oˀ6Vٺ~Wj* \h1mhg@Z>G Z$P$V֗/ܢ< ^Y`\CYBV(*p|- w$.WO49D% ܽhΧ [`y;[ i Q7y0n/TRq<쬣Hț'i(aF)a{}+k\9]`J(+z@tmEcpq76tS@9g+W1g郛f1zLlQ]7l: 'ԑ3pZ3mx0sWeG(o,8 h,b>|2&o:H]){VgP -i!y]g/Pȣ'V[kQ"fv-?SWLW JH=~ j3#Bޙ>,䕛hRjNu^n AeQob:ZuڸbpVεZ=X+7N6GM6ۏɈ"Ɏhňz]f*ypF]V=X+76es9Rϻ`Jް Pp9sfV [n8k`(AMTCiodG;hY/"8`icjPHmSٯ ^{x0FVQuhyt%]9Z"K&a6M1a Ƶ2`G#U#ѓR=ȉas+ntD ȼ%T2DUY=GHR BϬtao^5/D~s.- %[žG͡J9L sԊ7+=o+XfI߲VJwIqUj%{ҳRiˬTډ4h%h*rcϟ5՟*,?O^;m#ۺ}΁ *Z)C)]A-ceYk4`D#ߙhȮ2pzz@>ZDbjԖ#^M%@7ȫ8Ϯ["rgǹt ʑ \a}Dr|e9ahk<Z(M6eʵt~жtRkWFUђu(O֜VlCmh5^J![ i X|oZ K&~=lÎp/*DP v^ƴ hgiLw,@cP,(,tQ|BI | jjҘ"XQj0W*xb]Gf@T) <* gݹOӰS}cI| <'Ula^.!wp- px$./%5·6VJ;=΋A-d3 Gݱ NN'! _=]v`[90˙(ߏ<ނꑪbigu?VeL֓vXXH (Y/s-~10,$l/0¶#[$O28ݲj&"[dƖb}bX:ǍR4MՂs]ڴӃwrs6EAnz@\U6??.JoIT +;bY嘡1Ǐ)Gb8Z^m8\zM/y1Τ1啓qRv_]PiR_=I-kN*P@EK*ɇSh7*/lDw ֣%&3MRw.>2E*庆jqj7Q9V:sA=9Q;vnj[qL{2i;,N2'@z~kP(ޠN,eq8"8c93ASDE80XI.`%)N@3M2B>gEg|{z,h4]H`jH?|Wd MqB}2zY tw@B* u` tnISI{l txR:i2=J K%VZ K%oXB)RPBLW4]ES (ŤdJ{6BIVl`焍Ays?6TJ׮?G+m)?OELt%Ƌm|JE*RO_[p&6W|?(fsֱbĭkK Lp"_~+X4 huJYE-.ʾvǢx*DԖ_!f|Tw},eC0S` &9H2 2Ixtk _E Bd(+dAnq< CT+V!d |/{!!tVe]"1o "j/H[b. g-Aw Ɓ duV _7/@ER#Pe*dД l "0qI<c8C6[&WB1&eӼk#{2p?w&B4B FqPH)B$q&[IB %+bvXddyJ'*ZIUB 9PHbf\)''Ye2J+ϼnvkBO[RZ!0" ZVY?=ʞK\o>Ǘ}@c"34+l急M+=aU5柁o>^r+پ1Za&xG>s[ HJWi\}qukaН1P GIk4 s` 0®FǫG뗖J) 8v‡X N zR.κ8!زjI@"g-r#'rL2<)=FN5A ]T~hܹ*wN ,3{5> %ob+Ym˼g39 Ű=aJ$@B*˴4fjqVuTOT!^8B a6iɻgE%N#ǥ1i\hjo/ i&deQv‘X(EXw:fn,N2l#DPbK|rc`)o3jwK)}OV]!>VXQ]4!2:YjE>㭏գ\!ʙsh>nip٨#Z+ٕZE~_ D E M@0<\qү]6$bB A>Wtv}qe$ABEP4")Q8rRHxH=U!-CIp:^큡MBP/t/<6˜ H KD lTtQh`)jxVmtv߹]7dxU 0좙,8O |h$DJr 6a:bx8g{܀EFl+t;d=>/GmL=^ZcV2hQ!0sIɂJ 5m\64TSCSd*@1E,)L` 8ryNdUb#R)U$$1NJ KD $bbX349AvN (AD^nذofzI ocE_MEhuv>^vgT3@.λJ@olV c2,yQ*R$Uh= aL@QDS2AH=V@o+ucĐf4y`ee [!Z鬶 5dҪ 1B=ӳpJ%`}ykm ۷HI}/#,/jd)h,s):GceD@TJ @%!`T6M%,eu/^,h 1q wS2fDB}"`t}t,JXz9h{i:6Fzݔn!gZz~qYjTj-tWŕ凫{OW߂x睦)j2SMA+r3i7ac!&@,)D s1n (}\D _'iI;!bi&wcS,~b;Ph?D0w n_Y"aŷЂuQxf7{r9ϟi~@Έ]O#Bh{%Y{ |=LO^`s Cd8&yd9BMe`/U-66jX.e! ;1n\au{ӝ~2FvATҹY$?9uWK La ϭ+鯳;7b'x p_ c Wwd!y>,Y7>Tsq4lFlN&FӯK՞\LrJw<_'J-79 J}*IjHFS ZX|J_d]#֖:yo6eWaEEL-S}:{ferj̓\ԣ4z2,>cSbƼNUly>-o~pd"&IQxuk>$롯G'>ynBaXT1';T N.evRJ[b0)=FN5^i(mN ٜ4a3|jKAV׆+ (}z; fd6xb.h+GIU)Q.$@%Qܪ {rK Ik,U E{h08 FIuo:7n_Z=l[$ QiF48;!@]A (E%A 7Otfqs~>咋QN VE}ӱ5HMιaK|:2AOFq7CrY֜#޳ rJt qXw-.= nV>Z*+*~.g>9%.* z6h C635".ʾvɔ2M)ݪ^D8͍#3Z; %u" 1eJ"G$zS1P,8s;so&4yN|u읇V0A +Z[nɨ[?{Wȍ/;dV|ڝ;,nd͇؉3ebjӒ%٭VytWCb=UJڔ;~ 8<$l i3 $+Lf'p& R 0L#Ӟ#?nlTws;(k;|&f~c:L<w&Һm _v UҐ&*۲1VJAeL#TE]wNPRGԏ?u rS?wϾmza* f˳sYMqOa?.B]߭sd)#&~ f ([BU*j45Ei붮X܊d XUmS'ʲn^% ŘTgԩbR# !;a-DvpƖTuѬЖTDȖT.X6k$9`ZT $ۃ w-:T5iC_΁gIq2+Pp8Dբ݃C9j99Q8I )'ů1Dh >04u:HGݒfHYqdݼH{J5(2Y倎(o(k21)7u"o^ѽ-uQ̢Ȑ]@*#V'&T)kja'hA+s$z VLȖAR1a5KSr \wtI'B}Ο pDo^ mt5rȜX`tb1k)-滹}:YPuIvӡ(AIۊ˶akkm uMmaD*.P55AEan6GgQa[EqO vT-uQBAU`U %F{CbQ_$x3 6^SE1X6E1Ml^(sV!ahHZWù?%x=KPI B|:/&Z@鬻5Zy8ܒ8Lʵ{p+'n{続ݳ@sڹƩWJIa'ʄD;SASl Me$R 0E cNЮJNpz\%z8Ӊ񬞑( q 2ly) Q8W) 0|Y13, zR7Z(;g`ƱJu1vGA9=ڥ(GqYlK~__f02BbwƜd@]wQ 8dj)kG=풩;9+0Oڻ⹯?{<;OG엯_~嫾_ףMܐo-~pJ<}*;E}X?F[pߟc5lJ}׉stqZKθ`U޼hI rȧ7&Eh</Wm~~q7D6j$ 6R0IJCnwNE9,iSilt}/hP̈:_|R,*$ |:-=BmPɆ<v=fumBs!}f !&FT6EQŭ %QZ ~r#655 $H1FimȖF5T Ej* Y7$d umH JT `-Y$nwHK¸oڽeAl' fBr0ukM(CzVϕ 3vzP!nа$^~y,¼^J` <ԥ3`~`U>(g7mSV6-EcIZwz!+߿H j; Yb U;_}O1EMV3Cm "wB5B J#gMؙ-J|QPIR5϶|k/7;Q|mDq87n#^-ˬs9>?l3&e݁0غ.>!#GX+9R'3&3aPWk'&&w<"`${܉annNlp-NJٕ_yW~]wW}WOhLeX[kєE! ڢqSei[i+D[,@[}K˳wS ,CgYx3)SOn٬=5} yGy>?'1__ YCnlS-zcl9HaC * -U-Pk WJF0S !ljjE- cۺ#rh&0xxVjUV 6gPf&)gEV3S u[㹿p~񥸿)/vyku}R.K6@֏"+5gͶ։"+Zԯ>d=NL $^Fw2xJ`3Z|)iTT#ٌ͙뺧~ z84^׉)^2nk>R.݇=(p>ecJ8eA1BPqDS.ZGۗt?X6oXC*) &k[}4wxO;]<7o/f=A'23y"2'K_":En2' 3I@E~@MS/ݫd*&xhzT)TAi ځ%R^•KM 5hlmrZބMjBMAr7&[p)ON $ZiVWf:LhY<B1V "1(#pmU:rWCVH=.[)ZZ"C͢m2UM[aݨB.i4$@)$֚h9ʚY5j,duU`Jભe٘&UCUS5J])iTf.\{y2sEaBoNQWB]*H<i[Wɾrv7?7KN@: Zï 8wmPz ?;fv6ݡ-Mmjg-Lj+R频,ں'Wrfj?|M'[%R,X J(ЀfPS.(Q!h.7hȞ{qg]ܞK"7#ZKav{]\BiPmX<\[)FaVNx5 y>nD1nytR)f V$̞S$ e(UP#`Y$EKT{+q]l[֭tRTWSG_B_?Y+Н4WNrlY#MEacuM:r&n:'oFI;6ZZ3"~kA]qWmLrWo\(Omxn0{zI[$ࠇCYQaAّx)b_^xE{TTp2Adߏ5WGGSAc`t}n=h )0GtzmJ5,?Y崰?+8>%v}ؗZ=q_i{ h36$4l/>{iaL> Tt{33wM'!y"AS0cX523`F.tmMRVZD_)-*\,f۞>wXb T}82OvH[q$p1J3rPMZ>] (!G⹞Z,6IROދXU实.[c'>)Dmi Ģ,N! "Lzs܆^()̅3AwIzO2B TlmQ45[U k]hlYjueJA$*l82n@\+h[)(ZjV,6.u!h.L1 +Ab&ݔs}UÊSq̡) G߹#p[y 0G5-rXJ}PrAgKH}>(F%!"e|,̻H5WM4*)Aq8qmM3\ /rnؘq+:XRO4o8%l#N0ZzS4[%gzm/ ~=̾ [ܗٷAwbdJA!et>'Q:pxF#mxHA,1"ɽW5ZUT5ץ3hHͤ.qfY~HPawÌڋ/|vcw -ep؝!Cr0&9VƖ =Ec˲Zb-1[b=bX/U&'.T^)h J!I5J 5%NhqtA M=XӨ.dQJكQT=S t@HaSئ$7MhȳZ9%DGdR3;DT\cD2Y3%ت{L:leJ%M 7GMEz-ވ@c|)d`TIH ĵB;JͳܧrpB6=@z">#{Е# pUxNؔf+yJSާsR##ծ/ǧs.Ul{Bpŋ:d,q/bjK}Қܫu5F޷dN?oJYݕѩ=],SQDă ޔN]XN*4fS8|oqz/f: >C[߶Qާp\t/oNl'`s`~qcp}1:C&>ގX&Jފf-)?{;BgNhC/XxcOۣ3I /7Λ ⚐I+e`+QJݎ1e'#NmgM; 10o%!& 1zo C4IE#E!fʩzb[;Wx1I?.`p*?n=!9nqR!Kuy)e~8B_hD3ю[6ɿ0 6 ¸ }ߏ[z&ѿƙLW׳?~0\7(QP/^ lQ[g-5:X!ÍAnkaAcZ $5 |\.H]`u׾[MAAi߁zp Dw͘l}5΃ɇHi2 vL?M_) ÛcÝpt2S ĵRPq`Dj8{_\a%Vv.9@ҙې 5)~0>d@ŔkFWUp^ONQc 41s- &\يȖ^ۼQRD|n '?tb 9 -Yqxge?,l;ݩFn5 FDevjrp1Y WKm.ҏ˧:N}) CNHKCixH紐(i_93^fJ0& 3{?[sC]Hט!H_:寽O?KH[)ߋyf;E:od;pV[bsZ3YG6Y8 D*؜ѳ˵Yr87 oY}T U0]iCPʼn\I:[} %*fp0Gco*x3U,\>eqTOY)&* ГWioK}vO,{yq6=`ԶOǡͣ*2y =`{@T "/M#|Hb"-ށs-4:3;=@ o xܔ-rn;@ Y2>>= ⌵1(9>]շ)ơ:@F š [Ϋ%V-ݯ:0YsٻmfzaAQq}x+ˎAԘP*T]es-e^?8bW`x̻!yKFTKCޤMy wcCv,;lFmQFK<1{\٧bFBr5-9Z#P  x6,NK"yNy1ͫ_{{W$_%ˮK.ŭzTLV@|X"9*뻋* H|^V}ܗf~Y1%r_QoK>{{x0#XHd?U$ ( f,oSdBog7_20wk܎󭯷J[~=ɿt41t}i6zWd?_<Q9ύhwφc>thNuIpp 6zs|ֈ \҇+7B QRg==J.k4?x6 >v QjWי#zͳ/:vskEgTQc/D(ck#OO qI1$"Qkz~p Q$=ՐVl[^]gK iØ?Y2+w?*?+dw?[;H,YZBLzNSAn)Ysii睕OwS)Yt)q?ctV0XLsVqԛUJC[=U?ZUB<.#;bK;<Ļ+?{+zF,G6w+s0wE,[OK9E*={RDbf/gۓHf#E!T ]YEr& N4KN:kʘ+gU1|bG&!u.0<6}T c1%" S_j|HExxwcn;8[!!;wJ#J=΄h IZF˯kͼ0dahĽMI?: aҢZȏV8BktnߵuFQBF|ީU:l!;ηbY821Eu6;oLVNuJylT$\I! y HkftkqZ +W`8/£T"17DeE [cL9G`G)+xYl\-8pP7MLckU- &rD sśi[v=SB۞b_(!۞Zlt`ƛUT)ŌÒ&mO}2}Q 4PBS_O}@ j{ {M`sN3#"{)HiV4P4 +8b_| i.:15Yo%39@ "d0ex]ܛ^f"ۺ\~xwaGa3\ BqalM0Xí+-VH8cC]$&\ZZSmAl1Vc°i"4]]n֢_>q^<U\P[-t>}RyyTU:06WjJ_O^y27 UM\=NX5=Uƨ~䓛]eG.[q4'דŪ1Vʹ4^ZRaGAtE .5OhF@@%5kO)b&;Xbrq6W;KkLf#7=}IATz3E,2nkf|rYxh LK s>8Ԏd!wa.DJH$"i%FF9TRwkc5r6R8Xr([ԏ ~v~=W+7Ov.x62ˏ[J[5 A$S K۞fK7 {gRa5YXeO(cKzv6FPKUU B2n }tCA ^*n#6p>Ƹ\'+7||!#/8.u`)ia{Y)kdMjqIU}o%7"pf3'!mɯzz kw47Ζ3x}˖:H:Ԛ2_ V6j6ʗRp$|` pftԬگ_G6%췮+ULm>ϗd,lȩJ}I?dH`]:y'q;7발X~C>Y ;ڧykFMr$bW1rޑMhPNt)NAE- /2 *"hhGqYyGKHKSq4!)cHS3jdh Eb˘AX T#x/7flf9r M}T_>Z#t'-<ҖrD5CDvj٣w{hsE(xmvU`ՄláUxNZcZH7FL NTi47+#hlɴ@$NLz} dL/7ٻr&hynf?3B𛵬;s㒹 ݬRҌl4\9| :@Rڙkg|K-HQtB"\lN@5)lFm1U_Zn:(L,@>ted(g[YOG+ H ,[P8/x_CTxT*cS >\YRfzwyYަ!JV3b1r YS}Sxhb؅of ʏ[aYq9cBUnJr_`d${Z$S!='sCq.߽Cz)VU1NLv{6Ă1&nt]{O!ow{2 DKE֦":94"բ*Ki1tl^Z=񭺣tg;BDax0#*pv6lVAnm3qP;f\t˙}"ZIm -f@ j+?UM>iqzg<%a=])P[:zEѱD ,E~E ;7n{ـ6D5 YzoBf=æ)i<eāfM9N(m>[˨iF3+gCt9"s>>;\Le+ibdwK%ͻye1Mj5iGG"#{RL7;LxKLTjFDrT 䌈gQG0)ޥD9H3 60fA5i&Te^oaU/W!ӢxH}ٷ%}84cxA΢wxK2$%|ͧ"xū&){y3 "I٬_^] h2y{7I/`է7/,C%@(bpm/7r!!H)x >P-GCA8`:sMKRSJܓV|D5}n[uHHixn(W@JS}IPPiK)qRJ`%$RJhTstn{ĥ8)zB~?՗Ts+ӖR㤔JHspq]kG;sWeAVFA M2,=YQ}j=mWAJ5wڥfrIq)>FEZ,W7͉#_iSJh#6\MucNB/P74?KA8ﮡKtCy Pz9؆4TMY|sc̼Zs+ԏh:M6L`Gji;N1u. k=z _]Q p~ vufID*+Jտqt\Ao:݇o9K^ʁ&nRj,q~5 ٩3;(!"[6dSSCVMq=JF(~'r\Z~fWq%<;Y̫s#N b? $y]X9;iCvzW^dt+$/D5ҧl '\}o &Śqk8.1B]" 3%&:IGڑ|,j 6}K9t=S13STگGw0'=Qm }f)"_Ԏ˗{N(R-%-uҠy'QV{ lC!xyb%z=s/nŴ=~>Fjbek}nF K_8ƣRj6WS\Ⱥ$sCJ 9"@ mICL׍F7B b2#̲4TPs`Ҹr{.^1#<HCaohU|0f@ҷ|p; %Z7+*FF+6@TT԰e<׈-W>xKYgфIQi? HAAT4xԈ⑇,Y+$x*uEyK|*U%D<<,RktL%]=@C)բvع;'@e#vb6=[ vo^r۴*ia좰&*i%)yZ)Ӎ\=A2H LTav*S "UW8ia+s','GҎL&2"ϘRghRLd%MZR\ '2avU FWXU$AQHPyVd|\.Zl(b]ftm}{ڭ̙9T٪T.8j)H3ųi e"v+񆝦_"eY:tFㅰK 2fm9M8fh6T^iɄzZ=e4cFO) Z"3Fjz `m\dvmnR W~Dx&;@Q* `wYk)AR4E T)0Yɱ ,֬NSQrY]6uRҞx ^(U4ߞkwKPT]ƹe>ĮA]Uf2+We:j "FPFr,PM4C$^`w72?Q6n nr =\n@`O63>T;[Y~I_BMgGUz PGE˰*b5[2FAY k"IƬ2ֲ1e1+eaߑk3o(Oq]C<E*cXݭ TU0@pY7T M"QVu)obrecWJn^;Fimىhf..ٷilEԏf~XϬex{_7`o'KL"|^\L<5xV-!  Ul{7T'b!,=\.֋fhoRm<Y(܁ہWav`3M)DMl0Ȇ= R%Uq?3p t51Tc٠B~GM svz)ŶZI%Ͽj?oڟji%T kVNz:_zoRgw3Zn_:c yno kW>M_1m2o&7b߱M5G$/F?XJxz1H­W]Y kl4z:{ɫK +@͋ ry֑r)EFSoh7I([,>v@[ֳvn]H+(W hqnjzjX |D'!^ Axj.$䕋hCLQ 8},+82KowtzpCŗwgo3`oܑבջ.jZ5ʇQ3~!km;šu:x&&Hk.]ߕx㞳Ks:)}&@ju[ORzRʨ2t)qHsT7v')=j) hr9)(G!Q}ޤ')=n)NZ<ăÇ=܇ڏ"ԁ\+Z?_726(y\WV BOkDYus:^=.C~2'//nMb-[)1[eГaI]S)Sq bI, [XhNVI-;JD%]UK%NH-R;V G•5SIi6t8+aDVj?{j݅ .֪bm]ߜ7<n8TԮTP1H4-2\UЪ/ ]/ +C)pL֋5q vFtȠfHG&g؈Z1Rfcp|WEL2Xg8#3w 〒'`np]Yщ*t~bOwcNy?( pCGpD:><` )}!π+ \bj\4$"OM<34Pj]f3%)2A&ڴB ɱɰ^J1*ߥ _֍ȿ}v7=+s󇬺Y&.K|O\*X~#?*ߟIG>hET]߽;S¦Cݿ53PfŊ7[>OJpHY"QgtJI=3&VDBx'.."y6]A͍Mlxv5h\Ep2QLQ.gE16&DaΓC$AWj`sK:PZ1:t ' P^vA`"=*+oB0n~Rlbw-O{(!rwLOy¨5~w䀉TgߛhFS+je0Zhߕpw2Bȕ+vQmT'^T*[Mw/娖OdtR*jRK<,64{)6?$sqR`(Z+.k ju_Aw;n!/%VIuv!'{Fs~Rgɕ:LkNɥ~2?vSG7kc,>bp:򴞡wxcE4J@~n@([,>v8T[||vBB^۷v)Ơb1whCzPO[xj.$䕋hCS0 BҎ+:?Wg[╧KqchXa\k{~p24G ƸM>8;7%>R00+.пR X|úӁ|#C`}u&@C%0PS)` 3bTQE +)d)HRsex&8AUޝH=5g?L4_w6ѳde\164I!>|x BEtEV;_~Z y7(ֵ.ft駱m nOa K4;8KkhA3ñd)oNphcF1!$VF;,gs[cIJi⵪ v:s#lܚq _-!$z~QeOCm'zK˙ knwܒvw:-}J*Wi~9tYE$"RAJ ]7-.n.bRC*޵L SWE'zx>ҫgj_Q^vC/Ux6Le6_fEX;H&kEɔ/(LRM~goXG\hkV,3}-SۄV S,lڠwH2xXL1~\GdkŕЂ(3^;'_|:9w^(wʁuM"8`2 A&w+uyZ#;uO]q֙]k׳eMsFqXvO{zLp:uj{t³TT'p1'ݞe]]w;?JꈂJg/ZSmn)GN8*b̩xMٜ?9b4B\ Q-[ΉG0W.!D!8't&z"*hƖG˯n5t00vNCvCScyu y~gc et*'J|⪠)##B\tN!nCV(dmnSU:OguG7&.SIZՒSBtuvNzԪr`]%hm-HӪ0iNNn֚]L"g*RU9m.vJ]/@*4ݗd.v2kV<,UPMWo͓wk/o~.3Y`3[o1J87fBT1<MT*r݃}\|`3zh||'rRd@`+G9+o_3"W]F?$Ueւ*3to7{S5D2W>c}|%/'wN$T!Xn&~| \Vf]!ckI$F4 :PF@F$gpR^ )u>o.7ف a=__/J/'NGwg5c{g^yCTrӥ7n`|=&';4_7m= FqwQ%Vq^pʥagI+f?SvPLi ~l B-ۂok0Χ[\;M:{w 4Y?auCu ]7[H!8i6C#4t_*dZ2Ί`_^%)ʵsGqc5Y{(sH2e 3uo;اOv3+XPPfc*СJkBG\HJ01 hdVr^\xJ.2Ilc*"T AF+ ǡ 0x8IDcm*IZA̡Jha}̡*|C3Y9twS]Oň3R1f) hO;?`%`%q5&R ԡ&Q(t{;\y2oF"y7bVO6M-t!T^3\,(ZA;ى9yYi$q:C}_x<2FZ0QC  `6 5M ˽*ē>mƓ]0;y)m_DT0{9|bj;..v<(Oc*6 >@+T離ݒY1{tJP .F"]{AS/_;2],U$Bk v .߄ZmzI 3+R׋b2h9h>MΙ3;gwE\5`i C?IbXEІrkthB V4VD!(TRF@ z>v=ϧfM\E׃({']6{읃2w-m11C07qD䖼壳M̱8vwAG^3>>xjvof>WWC~ސ54ۮ|3mXqT+1qm^sUCKpmNJ^J`DHI;7r1!(ƌQrfNhLJrQ{vDZF+PyҚB8铝+^w;WI/\j:]BA'u(7CqayN|ya8jv /5E$+ܲD5=ҭ!5CF^wU}[70^^9rT^:Zjh"rԺ1P"*#1cY?/|~3Ln;MQva೙.'qL7͇:0= scéxvBEz6󃀸/2r kvv~]!YP?ʭgsѦ(C,wnl v3 ՙx:2N}9{ڙޭ h-#ލ/ []NwTn'\Eĩ[ޭ MQO#8_[ gNo&6tֳ }wuzpM*Ffxί&j+1zV5orύTEB*ۤ"$z=b 1ɍ ?5x Aڙ)l=(|7~.Q0\g Cyÿˆ0OL#k薺ʳqmp2Hl:#/crn¹~iDM %kY!]+d߱TeA1Mm, b0l qD$/J[>&4XPO,hӸ7_E +E{o)s>s#,Bkqb3quLvs\$A㓏H6vT4]=yZORk&.s7DЭ唓>lHƃ7~z&.㏏YɉP'd7$ u]}م±G&(r4 Z O_PJu| 'Z\[$s|$[UV|3  ߸+̦P_ g4rfW_WI>~IXZ;Jӝ7Y}8kGyT#S+M "dAٓ[y$ X(h0cD:N$ U*i+XyKiͬS Su<<='z!$%Xsh/9*4f-O6BSlj1 xQX'fHQru~Bj .\գJ.Gs'<,8 c%J$/RT,Hs̿nҴI~uʷ7W'D-8J`!%"(1<(k#:wM.Mr&[d֝7+ ve3զjV)P ,CXv'fogn~?yaGKE/&oO > w3)? D=N+ʁ{R;4~H$E'g Z`g)NQ>͇ Fц,ng'#Q<}F'o>mlB#"(JM0W4,2X2L"]m㸑+$8yȪb\ޗlp 2ؽͧ;,(Kcg_lOڔjAYUd*LE;N+=ٝmՄꌱeMI :W0y &BsuY .Jɬ{V?nMh A vNXjw* c >weY t[YX-,˂Qj I:ArΝĻ.R[<|gӥޗ1om{K-IH;d.pczU倒ʸG._{._$Z W>{ BjmkHږ1Z+H'W<=:裂tkFɆݧFo/ѶyW΢*z=wբHrkp@ DXC)m3p{?Pi,B=5ROknX>m,H::𽪜fzĭ?;n1])q*qDѱҒ`Y]fY+̘X+-P8(I&qFyyRT}Gz\kjKR_KM,.ZƵ$j.B?=btK^d̈ }b"cb"YTy9-!>y4Ꭺps{qXVDBdǿu?uǟmb;t̂9m$hǶl)}c1 oLcWʇS|Qa,r){ut ~'&9+nk(^:i\p͓K_Gg݆>~;E_zs@6jZ=UMzTEi8dmCpIpדyztVSCFi29 G'DcO_bȆSɆɑ --N%[Ll1>٨ O솘9wѩmdtFbJ>>=1:kOʢGl`X&7(^;^6(:7E@w6֠8zrPjv5I`x3ɦ(PJ_a + ba/N[# Do}. yorC6EđkmTȓ<.{"Ty­均n-7w_^YrAUB#!e1Mm., )j1 @Jˤ̀L7jNiÊkbɦ9Z}#|{}VcVޤ2JpK7yzvUvPk#hRW~j/s{ﳓz6gV2z\a,3|{j"'C9Q^vNEu|^&y lˠ2 \f(P y}kP+kyC)7nڰLŽ.jGϘK#rv/8gW[aO޽>au})U_/I?MbN$!jyghz19~^*9[72JH]i/Wx)x GT [>Sו)4LčLp0@v]lHT/neo|8X$T rwYԺDr;\uRPac7Wk͍}`&-퉙'afduNmnBhʴJiFHP"]`aBČJQFcxr!TU+rB _bAy$HVj`5!XPz! YbQޓryp6/%@]$U)RیC@ {jQF*Vg P:=̘=>AER h4gSP2U71{,Lt_hQI0#I }I}A }eB)4UVZ0 y+a5jPDig4Pd19CjOj2ZdsK%^K̭7<7Sz[eVBi-"MKsvQSV﫳.jq$Xn :ŽlMbɾ^FLiyyC:#2ˁl|iBY75XKn_ZjެnE}YQT{YТc RXYUz&ρ"2WU\I"#~9̉So"v-A#>_SSEz>R0pe]jʼnL e T^;HDyap-+Kۭ/@tlR RA^H. %}Qb,=R2(}@ 诳oeI+Nv-a,AyK+X:!PGc8nt{T2 1,l FLgUyTeS]ɭjʏ֍6 qn M(Y{GѠJߖL!m]2PFZ4j3 }a k^LGz1}kU$h&j9Ji-$'F¨HD5+Z"6^m AMew JځF/8 &imJ$x@;]X$2=Oz},or2ז$Y-MNZ>*Uc0+'r,cQ mRm|ÔC/B}`BWXzU(1/B8%Y1A箠"7 ! sea1T3Ԥk7л' :_J?m!Y@q? Q8QFT"*jYcW/d kcc-:c"`pe*bǂ%i0";;ww?:fD}(6e|KBFdž8퉔 ǛC[>:4|љ=,Kt>+T666 2sTOwA>2'. ?~兹Aщ E6>}zl,h:v3V[ԇ͍ͯw~}_pQ߅j[}?q33n Liaߠx{À _5v'_5>]mnևiX>,0;H6Vjȑ\b2\D%krK{7-EWSjo,) UyI=sBW_+|) Rppڠ` M )lj?GtZ k%Ci.UᡐD"]4JR5dvk*yI`= IJZf2 F~a/3 C(SEPhYuU +Zͬ"TKB(`U1=Ҩ>B"޲N~I hhܶ(vbmFP'Ti ʑj6tAY;r#NM Pf,U)y\yjzDc\U-p],no޿`3q~ {'"n& V>p:Nh>Z=VA;h<О/ xYFh32x ߘ2|`׻. ^MTI7]}KG|p(+q'<=:詌FF+J0g1ntF@aGqym% nРsO8028Zd+brx;H3w;*< wE":!br. !TMSK9eEdˎ7QDXļ.5k%u熤KMl@j yepƊ;u _6Ήf+}V86ұ^˨ B5s!gZܸ"nO2> 'w o$0$x^'i߷hzZj[x#g⯊bΖc"vb5q4G#2DHco*?T lp;F N']PU"q~aʭ \:_O&xռ ۛɦ@w;yu;ȶ(>eK$,tNntNr1why%FGj.$nJQhL D'I턧p[D[2D`ala,u,hQ mjޯDVDX_yVv~끓pϗvg_x,^_2R^e&gĸ-zxޝy:N A}5j:èSNydz33_QM?˫;t{bzJ?e--}0,;$.'hŔZ%0%} ;XR G>Kz5%$,r .B}0x SZnk{R)P;"P 0) V4L^(/vesUsyt3|RE:dQi9oiu,2gJ S祓  :I E}}?ˎ=kT.(t׵?yZ;Ӟa= )ш4y30n+r4؁|h˕~<9 䇍d4۪:䏋Wx&hWa%iq?WS\r> r~_NC)%*hC%{'RgH)e.ۈ`LgL^GJ $;  LiPg`aKĈmpY8M r)Ҁ FJ9aiIr SzZ\ %lw-\UϹb pHLr{Ü62=80x[$8C)\4R*PRo=OTKX. p7QZ3"kj[){J O$JZ7Wˏ__A V8rMә'&_2P|B]),_OZk#3`+aÕP]?Y(ilII cio9.j* θ *S?9 pCd'}T_4fuO|8rbKiw+qMWkͻ׃Ԗ|r`\(X=S uyLa}I [Zd8 T"LTiu&TV<_kXiٳ;~r\/gOp؎*~{T̷kA qeSaV3)c\r6Q0,!BH>\4ęvQj"-58^D7X褉@G)WSOWa)eœAnMCgnf"p±wo-$:R ǫ~VCFN@0֧(w^Lp9@ 1I)9q VW + ;+aTL3,zu:,H.iq;1)7Oa,{ot$z*b5V~jבQ &,jρEDD楯㥧Ls;|!xnM Γ/\6n_MH =^{nуsq7]yn8 MLp\/?("-(ԧsNka _E i C"u a! f:9 Bձ1G㩖14֓vD\=j @.rW)LBLdWhPNmȼc1<厁FUE>/TssYg.%pR(tMYJ[J#1,rB{-S& 15)XRJa&y3-pI `aӦ0#o)K$1 ťA@ۺ'k0卿_+>ڻǵO_]>-E|M6 )d|+B:OHjM=_,>N?/M 䊑i+6bRڈuBryVK(:f\ ɂ4,?ZoKģ^7j^0< /njʹ{{îW;oyLj!Wu\OfoڡﭯlZ+Cb[v~Ey31CH+,>~DxFdttQCӶmgi!CQ5ccƌnx&q;gwO{54\w Qtɏ%h?zGgf:!i3](\+-VBQj$F9URRKYrsiu|Q1+{vO'#8H8zʑՂ|}74L7C䬛L6T Y{ So96t{n@ TkVN1~ZO o(-cunw>t%3j W%iDVYL #JmQ E`IvcS[H.BZbrw˅VH⦬X 66rv.< tDZjQ! C+,K2¬AƩr#aǡ%9H\r=X;$"4M@"nd"7V-QfHݲͧ)P #$}Tp1WA8d,"\s,q e-KJK&3)`I;k89I i@GjndPG }RgaưK#,r6qe`mu)0})(2qUE(f`)❕DޗDQ璪C ֮R= E0%hWSbOrjuI;TXNj,xFWh2WnNu 3S2T9zrFj7ÃWX?z`k`m!M1,&mX) E)Hq cil*UN Jv7p n E 񒘲$5RD=(4M/D{/|%p_h.s* CΩlFM*g륍1qǒ@O% sAˑ"T7eM+ٶaGB$EKK |"2Q8,͕\\4bp~Q*q =Cejbݤ^Oֶ]C:AtB_Xg27n3{;)b' wrR+ҽp/[=v*|Ս cu&D$!"p}~o/ؚ4#{RJf!GbF0Cs+4Ϗl> /j 頾 vKo(J1'xGbJ霵N9iH]JnIШBxb(`nhK`641}Eu-B May *%D7u !P"Vo@޼韺&u\`_awcmۥ [~< ?/.73=:BG2_A`a1}ӻbq溪})`D]$KU,Q8%Oa%%=rB:%S6[(EQ^ݰ2rJ[jMPNŖf,vrYX,a$K %}dpq30dXbiws7ҙ Tsn_<# +,ƣE=${6#'wjBZRlF` )Z9ޒV)]/Q~| <\xr)np>ed,1iɔ%PzӛNS˃[~~Y~ .|Kt% SCܕxƦ(a^ 7:g7t7T } z(2|pp|W艻_Y)MפPؖ6 o~6`U {7K "LIE/߿<_,OU/!pLzU,?ywn(|shшKĻ_|HBqcCql>T`]Sa3Ć W!C ogޣhRh{wy }_QM36#Q 4rvV͹0hpWZScË-eKeγGU򇪓<[5/^55dĮA;[# wH X<}z#,y%tN<Ϊ&p*vfO~Ycʝґ|"zLI$s̾4(\)nʄ`(pB9QK<[̣dzuqfKktE>}Jh(Òw's8?<ȕ{"Ha1od&9JrwNsh,ߒѪ^)TE߂azNry2{8@(VO4R%k]2hWo+{!!%c[._Q,l_k2VYUXUdUM.ɮrް-5M{NSp Z*!ՄR5rZ]nus~mؚ<]^? |zG~p%ě>e%/tFY06\&:lB{SVePnPÉs"b8ge)CjW}: CLs{/Wa<N9ɴ;'S,xWEwj֨Sm͜}Hy2e2m)QwJ..ق=KuǞlmŞ-ٓ#(77.0zpn&m8>j~CӃkdQ Uap%[}xhB&P=ۆ۔xnu^db` ɻq;x:s7oE;&2jeV#'.+ :A#ؽ#+uʌW B+&,b}N:8ܦ #l { I jJF5f7_P0(TOɜrh۞FeT9vVfnxt%Yk&Zo; x{qo|btƸ!&4]\bRq Z4 @R`!1M 0A{ &B\b4.f*V lLag_+SX?1X긶?Y62#5{&՞{t,C l=LޟjAuS5ςZ71+,qŀE6F9)6V{A@p(2΍Z)V%< $ h8qRJk:TP5 qZD]-%$xW5D.UʀjݧЖ12 =ޔiNu\߲R'==PP.j\*M9BRs+2\CNZCl 8P[jYs?0?%;r/jvLks *9$'Č$|E02/M:`FKw1R/]MGp:1<}QƛF?xLPVj.*ϸ*Y5ܭR'%e],O5`iJ1nO/IQx ^;XFq$KVv7L\-c_{_Z93bRE.M1MM]~LY y"Z"SN5RF[)9S:F6b y*`ڭxڭ y"Z$SzOT.ִ֩xv>)槓,E&e5[-)D=9™UXLfjUBDa'~T<3=OLS1f1pT^*Q"ﭮc7VW3p)sA0Wh$QʟJ~O@d&+ҁޟ,v NI^-VEX_xE\͎ 4pnu.9"ꩉӵ.E$. *9/^緋p -^!eW"x]XgE~]Cq |ҭ8qWdJラrhj7mZ&׉b)b啕('XE'ǰ 9:@0<AH©:c=!u oTm]{H.):A:wP l^nΚ]*ū]38jT:ռ1YAN@t61yKz_4όyv!!o\DKd 7j˼n v+ GtJhFOjj.$䍋LLgrٔ9^y/0*>TF(]q1Hb`G8gf6b9C踵B[PC^1S֯ĤZ<ito,3WbSSFɴ:?Cv o焩 =q!_gi-cLYJT/da|Ciڳ[{NwA; åѣPK=ZQ+۾s Y-ߐў!N!~3D/<dAd"L_9B+ |-z@ʾN^Md8@+ѹ֕ҦQ94fh(FAjճ|tfH ŻGb5Jٸ|jq O(bWJ>5=WBdvu Ѡ\@Q*`_*:V1jO/81A {HGۚo0秇Z޿ UMr[6zx[d3D%\ jznmWBvñvRW5`ڋGG f!x} 6ajy>(5.GaT?pnz[,T^JUT^JURy\}?[]0^Fy ʸs)税":Ãsj%EVQҢ|e،ం:|sz2irq6$(ODv%0o/^IJơkNg)n`H-e'ImE-v#Oמӵ畆c4RSѡKKd\,Σ2Kj<ԑ3pTc5+<_>]Ȧ}ŧS9vV0+&W;6+C-#؛6R ?} DI笿:MaƉJ˹!>U4|=}_$/g/z VJTGj7jlI"CN33"sg_U>M҇^̈vX\ۮn>/h[zޏw LJG܀/n߿{G9,{4Ļvޥ'RzϏ;]T3%gWG^ 'W{9ȑ#-op0pw?-vu[kzʳ6PF1I/eZ(|Bh7, =?ZFNd =itvb!zA2*':՚IytڭLCZ0}p}(4qPUj4p!u2ɭkEB)P$e 4RL !8 s3h'3-λ/h*qVQvp8|]גry~Z_TQKՋUDioұhůE$ 9YW:aM*h͒\vj/8mX*}z/ETe)77_B^i%%Ka>_\dofd#G7V9m+rMp̊j+motjÍ#݋k]fffq{G~u:a|]+>0_=0}G8n$V:f#g|ׄ+%Ǖ)9.Yx鑧3RGkx{Ro뱶&qﮠFy4ULThO_=ߓ`Jlz5|eV=hXfъ˚ Mq *S&-˔:D\Iw2F7n*"pW+Psfdqq,ұ&>f]o`3wސG>Ǔ٪vӣ:$s9RCTpə:<ܿ:q) 3꠮l{]4]eWEZõ9S=f8ٵFQ'tdϟG)Ag<ϼ%=y>bp'h>H=*,װf@-XYSo v{NK##qe8æލ?V@nߎOfWahc^ZszseOF9!06Ǎٿo|org%X`T{?/sVEX+_c49]svI}X'.]Jpݳv;gu 3*5ս{_wp7u$]*+oC*7v2˵@vaqXajCxG0ziQ"=˳ WjFd>5_Mο|7W%\ .4& [qXL./1C+uyxqxye q!=-dI:q_OY+EZQ8.^qmx/8Á_>ͳ?/CAEԬsx5ռTҁSf4c6#+D '@ڛ,u;ǥ{q-#[Cjw<'N[JS/7Op!-skl0eT` if2&kH1.fU:eܥQ-rY lYfpˀ a25YY^E]5\L".%m4XsX?njv>+Rk+QB N%80͍vK1Lb]kFX/ iaJ#v҄# YII8x&d„ʔ]>kֱ%$[B%;e޿9@q0d (ϯ<ME8$f@;}X*0zQ/8,|+n+b,=Q5 V+{yƛwރ1=,4_py t2Rfr9@/sXon71p")k^j;eC^%פ!T*aHK7&L:TĦ,,աtJOY{a,t[mv_߉XtɳnL_\^4JNMjⰇ)-yK8QRجGt9v$s/Rc:bj#r"EYf^x(*ΘifQX M9U(5r clvYU4Mċl=]% c*s"1ܪ$IzRĨ0b(8VZF3K`fiF ƩQDHmt2jo.o};< a^෼jW;Ia12.LB3л6bVlo禗6 )JsL-f*(%H}v w>2$Tl-"aCڡso{AP:4rmoj"1TC$Q_AI;ʭoCJ0(WXYgD+7.~]:L- Y62 QBZK뗕ؙ3W[9rX %C-ZJ}AНa}mlk{KG%q+ cǿFS|DFxAd$ņ74z>^)Q ։ ôOa5iIFq0 j)!-ݲ&bPt~ʱkR oOH3C(vb2N./IoqjG֜1r2SL[7sWzAa?n3iϙz3CMՀ4Bkݩ-^bY.nZVx_գ{ u\ n Wwn1)_Qg̰:gqSH^:G ֍l0 N`E@"`"Myiq*Puw%_A*P9,0kHem:"♕ jS_GjYлMH0lf.5jvG 򂺩=)tD߿r=M+jş,5Ay\mYa >Pr"d&eD~b=Pb6!zyJ~DZ+*Wӕksm Bbf?o  >cUs*M<\7pKJͽ*=gF19ڳLY_,6Ґ?)z?So[7t[] bT3Xy_ÏY="['WZ:P=-FhL[] bT3X3"S;nGZ&4O:%#uJj+Rs;':{q:̦wChy ULg3/CH!*0|Ydz'#C9.ihv ܓq^ǓAȣm棜ϣ~M@k;ÓmɔYb$R s' 9`b-bң,LL˷2AA.,}lz7bƟyF!J9N1:zWzuql:G!FCU*ҷh&i/e#="j^EeǪ-*qFmd,񔵙 P AFY"&2##.4~4Ee6`5*o_z݇o4}=_ M_p81Xt.l;_SVURsnqrx1y!jPW4 :.No<a*뀀m1.&k /QHKx"jRSXQ"=&IkbF9125 0ư'N4 OIM-j+R|xݔٜ3b*Asj" ݶ |'Wѓu`%?\Ju# 0޼sxm^aEe(]ΛSFHʝT[i(.Y0fD'$uJREihTm&hf+M3!sThvL)d4I%ֳCf1D*٬ZdYHAh}M*ϔk%Ld,Q$g|˘A'75- >FA[?}Zj?7;-.޷?C@ЏU9?z㶸swm5Ff޸ݳ-zMJroqE;_jmTACqJǵqVUH.XL1-s/m|&t)SԒLL&́JyԡYm5$ni:`IUy-WF5g]̵WXޛ-F팻pu1*pg4ᮤ]W.\'WuJ67T*ʵvJWUn412Dcԣd>VJ2TsM颠dd#CF5Ӹ긐ˡ;ڪf Ǵ˯LL08_$`L~]j/˟LaMz)*WC*vX0zvte6vS50g}< ߯ny=x#f>}zkv-J bqQ*\R%|;h)u<-=jRN펔Ӈ%z<E|Oa}K*hpѸ.g]~ԓٗ2Ê|Fi"="*ƹ I) Pq/`׺dcoo*/7a6uzg<|8$ gqH$̶٧1WKx)¿{}oU_vNY9GdQ_p5۷Se܇w7DԶmv%)-R# w gK&c7E,Rh_>cZ` Jc08dJvl]xs)5) lT v-!<D‡Dk0iJ:DJTVik qUłҌikIJ"D1U(6;/q3, yS0 2l ,ZFӲx_>E 0{汬bxy=3iwrTFtZ!!Kno~rrqtAU$$o(㛋A6K:=Gw}l4%HP+oGC")| #t4[CL t1,~@#7'aMofaFt{r1CnL{$r"!:R;b$ΧWIPQ06pfT)3ɹY"(F X'0[DQj ֞[]p'7vsH'I(0&ҕIB O-_& DamE8^ĚqQ4ga6xuC>h9R6r#EbW~}%H6 6ɶxlC73Yb-efU⯊Hb`, ~ަL<coͦĵ W; s4(z2Z8.`K7ceeee]ȫlzjP*%R3ː;L1/PtE)q\j_j"4[%KcZ~?GBIUH)EUQCld'şV+ ݂~~PM(ǰ3ۙ5 en&8$ofV2{ҥ%h]jXb""cGͦ ~Di=zϙ˜K$=*[IUt < b6?UMlz*%I/H ,vDwigt\))S[KK#5T 9u+&ڔZNPK՜a %A0t| ZrqqwMM#jxl W2svNxM"9 M$|/&DRQb[iHݦA0(A|Bxn¤Qm`ƅӥƒ$A>{ (۲^>:FZ\_}>v =D*=ZԋP0%=bq&r?|V﮽Y\O757~bcwc0ߕ}aW%Zjzj`. hᾑeOwQ36KXY;WPHq?bYD 9B4۲A§+|5MSx) 2 >#VTs|nlI 尠>ZC;W3t_X$vրZ-2$i2Q]QhsF+TN3zk?{8.a(תﶖ-xvyeqeX5zPO5L֔QDn\=iS'#-Z^/\4EDj輓ߠv7u׊(DPKc-ؼ?q2߽)!krHh9 >p_Ss^&TN$n1BY\׵vMJ?ŪRNa*X.*eL*PsJ` 5.c,l>g7D 4zƉk#x[/^=/fƖxI焢ZXWo x 쟗u9k5y{s0V,!2漽VI`+f`G "@1@[Sd(e5yԈFc$)$N0|#7YnO3映<6D}oTH X:[ 1[N5 ϖB\ Wh,Kiib;3͢_y v)%RNiѳ$Ǥǜ 7)偕~fb9lqAĪ()K^1#ڎ- V;j jO4s÷o (xKp^pUD _$4"t_7 ܗl7/^G{dWjdY.&}6f"!RDFGNOᣣjfA8֌؋8E: #uxőCNDR<`3X5#Jv<hLFG$J9HTu>hPO>(9(,6u%?̅G-'a#{a`jܘ(^24-+ ˔.iǘSLĖV¬QFa و}ΒNe?PniŸ,٬#}XxkBgԗ\ؒM-t"alrcyk0wܑRLy%x Sz ^X&5)+,=YbRZ &j0vuٸGAڏ 39;Fb؃ 4)epDLɜ,8h.,@!h+|Uy`.~\ϠuSܲb毦˸يfqq}aqxެW (rj $YU==ykZ\ Z/ ?N> ,+Wcn{F}e][OE{>v 1WU{%4 6Y;Ltpϕr |<;-h6S?aZVr\ܹ/9L愤5E#˞A;Cxn `w4BS4YH#Ȧ (cJy`se 3_Uq&ݕ 6N|#@Kl:` i/$%W LQ%B1 A'i_@,`@anQ]Dدщo]1z52@|s2.[G2ېM?UطcÌ #xʹq r"VbkWi؃lXNjö,,lm[~c<Ҧ(.;8_G_,4DYof6ZZnqGgf~,黲 x Ohֻ@e^4+rЈ'}Gړ q8 8q$p,JJ,1YL8cJ=-CJ`CíT _WM-fv<ֵw|Z(νsH(\&s982,!%_Q zèq)&hv(B": B+VG6'HznVK,I,pC[ƴ@VZ찜FN4LVi``U=<7=BVzSd9A7Y꽢-V=>܋sSʽU=j͌K WޚD_.\dtR$Q@VAFJb*s4g ?&hk$awµ}.lʸ<[{ShYCCi,MKqLㅂEQ2Pcmp2hw-T3&Q˂ Tsu(@BXiyqrvT ֟K5'5AOySXQ94Z%7'HCl{vo7d(b- A =1̓GXR&tww~^ bDiFm(ta/7IK]^ ݾl{NXaq5Q.O:Ґ\Estb~>zߺ |-1v+ȴo_iݺА\EtJ8@uA#źs!Yp֭ UtGxNplu~T3R' o)Z z%p-VNDj3BcU ƙR!bc@z_A=̀hX? II$ 24v0 |#:6SX&x!>A6wU~dͅ9N| jV1ܵsQ"i^Y' m6 2nt{kG~LHo6>Zw޼G0}v:\nbJ&J]4Q Jن%jJ-ZRYxBp3e{[4^ZӿJT*g !G9KheasYҊ?!CΓvyɧfK4󬘂&k7W;[ 0\ٝu/#Y ݩ.B~~j+118\,K Oo6TsZ1hn^*X>Z4Q\e4e=WI/eߞ1$17dJٺQjVvc ƈV9n<9Djd&uv!޹8Va Nh2^,d͏.2?),^EyoysE_=f7wabIiޛbqjDX3,\1YE *ΔQP[b&'N̹uZռwc(Ƴ+b$ѽw0h"B)m\ORl:IQJ#m0t!f+?uQDvUr{?a򦨤y7]*PaD`0kc 3ppQK'1s+rEef/*[\>.9"\T[T8}rj~➠!D~+*P~3b`!L]Vpր1u<\+0 ޛ8@>RuYٝlŀ%rѻx/˯95bGTۀ`|gᨬkE:X6O?:°}&7ٔNR޶t4_rlYN]ˇo]kcކԢEp4dSl'[TaJ힍b{jrp]5-`K:m;7yBG&ZVy괣y*$AC[M( #)\"EdݏGd#UdW3F )c,d(ERs&<'A;4*J[d/jn:h"iRM/4(MYRYSIYsG8 {*6 o0.fq\繶nFF"PQ`/n& Ӧ"`{:OQ/Mgm#ق_DzXKJIae %0 T<as “9ZR}jk8st#0,8.W^r2uK— GuqI)Ǯ.d'ΛI(~}w__aOo.pqw 6ӷ;_0j hgHEghoaxEJhV.Ĺ"~4"c}R1x> PɃ>w٠R; %bu%ъ^ {VS/ݪ",ݪyt{lUi qM˅ޙ|`|"7|g"?,HS]~&{ TТN\M{nQw;׃Q_wÓ 4 D ֖|ڻ_NmmrkVo{Wr2֩I*jT1!k.&ȇ4∼x ;#NmϾf‘6^N! cq]J$Kl}j3VL w#a_q[{;1_{>ycm͜X)"4)^j)01?) `UL,'" +vcKQW*6#cNa}Ld>hf /נt6MΞ׊}$ 6w);|3 by+,̷'EWiyHD_P-g6J=~\䖩^%RXiGWkE9UEjT F`qn[OUjI-u\DiáT%uT̴z ZANRBorJMTScK0soZ TpK *BȤe&s]!8/*UhvNq/;bze|]U `0LfVE6& Y~pL(A3$[Zc+rɼ7ڱzI.jDVI2=:V\[%E+f6Gju,rqۘ%  &Tcd @2@sux b| Y<Ԙm0&f&|_5Ff4 AsHq%FhN⠞*#! ,GJx45 +W) AE_ѐM ӻb}FbG`]2avvD\{ß% AHV}>b PqF@`H)rmS<d5P `۫ҔQ*=^FЪ`- 5̄c@.qHC2GsEn*߈&v \͚sl[l $ޡ*sa8$Vy?ˌqKCfȀ&`LLN_ 3uL:. W4hGъ)U@,M?7fHa{ McIhʌzT; Ff-`{Zı *OXKu)8\ɎՎ 'iHNc7v*^0X`m< ~~B(ϸF\c"(S>W:ʂp0-v;y2<&euPL@>"Ƙ<ŶcS8ȗYR96*#4eP0CvQܿFE5SfxaGBN`mn&l5*pK<Ĵ]P(h`1,6`xLN r,4w{F{!V jF(?Z"E%2!w`"V>#Xx$ Y !ơ@suUn ce %h٥ ]Ե"Jg9G?(Ř V["i4 ^e *%Iゼ08&j!6C\ָ,% 5W<YMY0Zf<|P,0B CK1Hx e6.M(pAZZ@e.E 3{3{6Q,wh3ZXsN"(ry2re"/-cr;Ofcg2dFaeB9E8V#ℂJ/qȦ`"IcOkm#GEv)~ A gNfV,lY~Yr$nZo&Gl~Ud $ZD`L:4 j%55:56>ꓷm}'0e1J wYHH K.}Cc/|4:5>_]\|) t@I@7@VPb$01^ "KQq˪ng)jR@Kh*/NTR@CqsV`5ceOAҀC*<:n}xTm1ۻ:5m1? pzyQBsfp$\h$H.z=60acsݡCՙzHayx4J0!{K_צ49[9 MRL'eI0@̀]GP=tIc`#F5M ]5m[k AasY6)l߰W)Jl`-߫\C# x&O=8lq`[P"A!)1,!X2'bghf%[h=$P>+"!9GHJVwm-mGn>~ዀ( IIiDFew7R[C|'_G3ЊS}@]02E:E\aL%X0 Ϭ?%!3 4W] K|838`f,U:Ij*@qᐐ.̫X#k$/Y\ݲ# zuB6/a'!O{\wx}3,@Q8us"Jy-㈋|sDYUm}`0EkFVCcxʎoJ>Y̐JhG;Wg^iZ [$XBd _ yrߐ WB+hZ?hZ]':V3薑u2:/ H ϡj S-DG@BJ%ҽс|({d9"Zw 1iU(ajA4ï\W5E~MLL <*a̒}\2J) 10pNhcavffi>Rlpn\5Z, ?5"2ǂFsh=ՇFTmG 4Fa<hBCI?IJ*Ic"CE&qOK1v{ƴ~z]j2Ŧ輼B)5QQ)R27r]'+}x<8AJ}ֽCD/ʟ4hnt5\}讵|Y) /Zĉ'b[6&%}l%Ⱦ.eH!I#B1 Eã#U[4$#m]=sըB lM`Br9& ,V|LPGkeAJ,`逛PdY)F\D\$`1p6)Ho=[kF1#F~Z5Es$WRݕ[(٬i3&>taF]¦'r,\j~ΑƇM_`X:2s%r>TvTMglwusW7[&kfT0{_oZ~bGw}0s߾!V;ʼd^ lZ4Kf`8qڌ)E;bZetX3i(<yy Ǖr2WFq1|Rkejr<(oƓg\)mBYN`&W;rǔ9#B#6,h^z 6Rd8ǝ-߬u-;~<+Cw8wNte _Ggy:F0~7OݳXߌ*eVs8~[Ф8m=-42΋ 2կPF*|V&fRh4. {~#'tv}sUp6)tzէzSc9ż^r!< XqqtYݢ04W,/5)X`o%b)Im||WzY㳯#͆-Jn9:XFs3xSTp3@|](z.F"z'Tk{?6Z?UCHù,W˯ܑ,ۗ聁3AbFZڼi?zoߵCrO=Xqvd@[hF-ar~bJ.@ޘOwf#L$:x4 6unPV ֊HeN ##2O9Oc5DwiXⓔ2^xEď5'KXlQt&^&xVṣmsBIΘpL3M+59"|ƻtg'j3%#X OD26]SetEGU)@S$LwV$"Dp)(aybMW,_?7w@?_@ֱ6. su  K!8g* QGXDR^dv͠z;7!z'Ʒ"Be7I.:9wz[7 ߮hx =קucj_ #iaq,xGG"$% '%J#{D2D&f/dT 4^6 cӺ 4'8"\"A~xBfwj*mkv:=GKӭJ`WATWZIUv9 )(iGl&)bEj;K^BT2s:W3r" rDňdNThE\G_}(6>B28>>@`qC$|E̳ɍE^ e0Z-\&-]w㬙6閅SD6oON|_ B4`%K Aks,$_TJbG9ZYaaCN^+bi, JDTAS4Tqvj21hK0B~RT"eԡE$(lXƩȮRqbZ unB8H%5 Ѩ"9O\~;iksGTM/El$9q$EsMHQǔ 'ʄ*E}_5qp mV.~f1vH|vWӨf{TX!PX{PeM0(AFb9%SaALOLh@FFf4(.!H9b;a~|fߧZW]ОR%߻:_ ];?1E p%l޿PL%GφCvɆ0IIZwF'5\0)vvZP1B#g@Tt;%1+K+pkP5'rlF tE]:f-&=cb"n-yOi]%EV7A=aH}Bk^Hpl&To€Y+DMc%nhl#N6wY#8}ۻ]?>b^&RUd/t'S-Zz_7_=,ze<ل[oԚ`\!.^@]0o/:د-()AD})ƺ;}w?Sy0k:jSMw"4&{'}l* ղ~v` )6?EEssE#I+\Ic/Udq<=X`o8tcYV$9b%_"VFv&mYjWU,aOv+ەa{pcp*RQ_ ;mlP{T^#`M}\QOa3?/o,ĚZ3_?/{YcvqV7g3gmná|,vtjӇYKcO7qղ2y#2h!X ^\['Xpw<ߔ%˓S8?=Y(ZL(aݣI=F"||gmJ+*׋hYOnf`7>v;T}L؇諝cKzA6#ٓA9;a?/Kc<&' a!?.U*L[O/VeJd.Xˏr`|q?s&]gO;ks 3_q\ZrZy9]n~WQm}Gav;9\ ˁOT{`)1KɁ;Bq=";L;a|f xS'.*h2UPDEkAVR _$w b4)D5d1xd5 NC8PJ db۳? x;n|6oEWAPSWSQ^VB:á4i:ZpnSDB'>9mdvZ7H!}Rf6[9xN͑/to;`b #A@4Jryb,9z[Pؽq^/n7Ykt1{gk-sx SBX>l] r~3LH1#+MzY_E {B%( Cs4e&="ϻl^Fa΋LV Ƙ@Cڂ<_,7WX*Z .킓UJҦOz"ta~کǖފ^j\(r7x GmL~rM.Vl$.2U"_Ub! _;Ѳie>a5ںOW]&)8eKjDMb:з 7w>{Zcf9BI`e׺T0v=o2| ߲%;^:>l:})(PY^mN5NZӴ0yW̃ڋ9GS1E]qUરG JLZ;pޭ؆uhe"Gv~6I6jZǜU*?-lEo{#0VKѷ2ANi]5rBZVWFh'y~ej8wgˁ v8m 0@x;NT!17_CwX<ژ_}wtةAbimt9?doG~ȲF]O(8[T%^\]HOF4yМEM1CaoPђrsk%3rLӋYg ^f턜pغCWh?zhpf˲xmXSao=R%& 4k7h[zCi }+-"Fgw˕B"U 'iTA .3LG Vݾ^EIzt~uuq+x$ǸtY?+* (J|C=ӈǛr6Z[ [;L𥳐$Zˀ @ֹ`Tz(Je>ފr}ǴC x2:bv~>**bM*V ]l ~r͖)1|>ȈbGijTBԐ 8n2=`ߟDI HBI?FB9( \._eSE(zz@#Ruu>WsyW6*z9ζAFJ`f}}P\F1eSTls?0! Mjkd`튮@N}s 4tфV?u(l=% 0s߆HjCT?*E? T#)K[=Vkh' Qȡ@+NDs?&9:"9N'Z]FVSVRVeN3+ R[ISq7;;ICGbݽi9JůTZvLP<()9b CջvH+~귫>XRq%lE,DOA*븏ڡ'x1Ei8tbVUSd 4c) 0 1D U?ƞpi@{vH'VD[jzHCOGt|F2j@,irXRLjul-<ש+Z* \j[v1L.h31)d{ B3%d7'g Fz]M fEvkϽd$,GK4Je0!FH,52VOSJ>(SzZhIm"Ś$ԤAFXEHk Hh P?\;~ia& K8:|ۮX ٜ]IOƚ%|p5ZcuvHF!`P%Y}ū B3x;Jz/rΡ~  ]?E#߸  %Hd5'2r&n3@6Rs Cր+lu>i-*["fe JR߱(1dX"tϳFR>-b/A(k~Y + ێc/Yh-ɸ_cdr\j[/eϛs.RQaF-|=S _./-H4?^2tx[aN.GKς9ǜ69-̨x̊3rUOj$xĐ>\_N77.DxDi)jl;,eAC$B4<^c="3Lh vjDxIi~knGTjZR#2Fýgښ6_AeO/CҞJ$vq^)03H8V9@i@Y]t=}( yMXj?z+Y9ʇMydFO~jI$`jXOM.u' w3% K,6c'rm/TQI4IW=ٝŚuۧBw\mj7qdRᢏ |'W:3~u~fԀ?h0sɢuc奏%)5Mx;Ia}D/n aM{wM}G gό4tڥZp,]KS nݖw|SHV6ҿ?=/K}8L Jctso@ݍZ۟_ D i_,LC`ê^>QAWؼ]B W}gb /}]tM(Z_ʇq2pu.$-:x Bm޿f"GnSQѣ7We@(o4W;h!wЁ3?MK5m}EMw:b]4~wt~vB+ \A%X*yn΅s-G<_u%+a۫5y=d~. Gr6ccӍ֘gmͺjq_>CuF > v'TI@q%k кb-yu)mGmlkZX{ m+,ٌUİ.Lt5TW K_A|A6W+ۀ1r/mC `͎yM~(E=(p*lD'ӂpX~qryXԒ{;IcgJV.QBw7p{ɻ?,omR>_}k:Ý bL+{B i+rwa<+" JjQтN˲IpyDJ^_ H'믴_5^ܺ';Šh GxbEq-\iƄ`nW&ubѥGKqRsr*-/R3r͸rW- )IZ7֐ns8|`G8b{=GG>qmm֠25=fJ%?r*YjȖR$lՏox#9Q$1ɞ" 317&\Z mYfE*?bHkŒa&"cZE5vXIgfL=mG=+LKr Ij9p2.a,Pf2n,dNrUj̙ iťQ &KC2-|g囁KFd4{M1vfexb?;[k@2a2D#K|V_Cy.2SD%-6uh&2xn$S6#8@YaPR=mjU) XWa\px|x Z6z)"w |PV1%+[8cˊ 1Ӈj-`oC0:˞8ȥ]U? `鋀೻y n4ۿrJ)H؋a(kw yC&Nz<|vVj|f3Ld|VgeQ98 q~GaԆ̖߀&ϧ޺;ʻ~hd[ UXTWW{TJ=*L@Yi7uAz\Wn60ݥZU>y1؟t>$t: #,XjK%m\rI:Xj1>h[ +Tm{>'cKvŘuP,qF|G;7uS;S4uU\k@g*r$Y$Y,xk*k`UPGH4ЄBTM߮IZ5K;:A6 (z_giQ7&Z!tzNG=&h(;\9+\=VQj!J&|PeoRVnW7Ip"$zi8:^ޮ zji8u1~'#X-xrcxfv݅G`>Fx{Q "G]F1X8&IlȠ;p=)n1j9jcn9/Ϊ?L>zN#76TLǵU'_w=f;W:=~Hd8xZ[c{D#&_>@>8)j?T8@Gy?šnC`݇D&0B屪r'0m9u4cDarMGT;jzql18坿[p_GVr.&5m?Hk2] f;DDe@cXKcq@RAOW<`Rzj23尰&O~$n2)_:ٟK))jo 4LYFqJŭR* ԘhXʏvRS#SظOSkDߨ5Zs,'w;_:Y{XunԋaDf emۯv])XY;w,y=8 tڿv [ '6BB6E$T@qtH6-j xb!$&! iq(D1 iqj+Vg P\Βc+)@ vv!!,ѓ)v,-2!A1f&_BS*bAI $4 xc"|e5Gv /N]GF!LF8[EQA у(øzu~Œ *W.8'ǓςIFB6)RRr K)ej+eɌk(wbﵕCbSw4fMXx,|8xEfLSAtIn5J*Vs+PlFJSɈvV7SϔIֹ#xb8 Ź sL"%rGǀqV+w5=@Tym$Nln+څ3Ji(jtM4MP1gg1F %=+ryd\{ v&( vGkEhB;I)bHI$r.9w7ۇ`pCVwO p|FޯYH묯'5?ҮP0oכ.E 1c>>w"re{_/8X}y6;c7|O*zƿ?1ʹӓ3D 5Lj><%DQsp)?8?V//ؗ|?1/X?&5Ժug-O"LPe}K.GQrp-!uNSKy<9VĎv9>Kʳ9ř!lt;':L0.{(;hp!\Q`.14uj;:nhLb\ /4?UDW E]%FtM"M>h¯-jl?<7r–V{tMG ïL ؎ 2J0RkMD ^ZƪzmO nMkNڭmMnxͭ%TjL¶E@sRuR"unDl+1Su!)O ;04ke{r/¹̷|_\ck˔t\U(lS A ԥR8 1@7WHR \-o,Bp_5tD{5)Qz8t.8q;z䜖 }19, Eȵ3n{RA̵QGjv]ˬ{i8Z92K#e};?}T:ΣV訌,4KiN+^X=ERZw*sN􂥴:)j g4 n{Rai, gc4lJGct*{yJP[eFoNErq20t,Ec)( )Y ҀZJFX*ztU:ΣV2?t ,qlf\var/home/core/zuul-output/logs/kubelet.log0000644000000000000000007073562715140135614017716 0ustar rootrootFeb 02 12:09:27 crc systemd[1]: Starting Kubernetes Kubelet... Feb 02 12:09:27 crc restorecon[4746]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:27 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:28 crc restorecon[4746]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 12:09:28 crc restorecon[4746]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 02 12:09:29 crc kubenswrapper[4846]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 12:09:29 crc kubenswrapper[4846]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 02 12:09:29 crc kubenswrapper[4846]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 12:09:29 crc kubenswrapper[4846]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 12:09:29 crc kubenswrapper[4846]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 02 12:09:29 crc kubenswrapper[4846]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.188785 4846 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194505 4846 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194568 4846 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194581 4846 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194593 4846 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194606 4846 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194649 4846 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194660 4846 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194669 4846 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194678 4846 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194686 4846 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194694 4846 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194702 4846 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194710 4846 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194717 4846 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194725 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194733 4846 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194741 4846 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194750 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194758 4846 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194769 4846 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194780 4846 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194791 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194800 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194809 4846 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194818 4846 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194826 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194835 4846 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194843 4846 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194851 4846 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194858 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194866 4846 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194875 4846 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194886 4846 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194896 4846 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194906 4846 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194916 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194926 4846 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194940 4846 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194949 4846 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194956 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194967 4846 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194976 4846 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194985 4846 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.194993 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195003 4846 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195011 4846 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195019 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195027 4846 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195034 4846 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195042 4846 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195050 4846 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195058 4846 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195066 4846 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195075 4846 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195083 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195091 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195099 4846 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195106 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195114 4846 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195121 4846 feature_gate.go:330] unrecognized feature gate: Example Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195128 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195137 4846 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195145 4846 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195155 4846 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195163 4846 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195170 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195178 4846 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195185 4846 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195193 4846 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195201 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.195209 4846 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195396 4846 flags.go:64] FLAG: --address="0.0.0.0" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195416 4846 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195432 4846 flags.go:64] FLAG: --anonymous-auth="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195444 4846 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195458 4846 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195468 4846 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195481 4846 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195493 4846 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195503 4846 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195513 4846 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195523 4846 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195562 4846 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.195572 4846 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196475 4846 flags.go:64] FLAG: --cgroup-root="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196499 4846 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196509 4846 flags.go:64] FLAG: --client-ca-file="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196521 4846 flags.go:64] FLAG: --cloud-config="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196533 4846 flags.go:64] FLAG: --cloud-provider="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196558 4846 flags.go:64] FLAG: --cluster-dns="[]" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196594 4846 flags.go:64] FLAG: --cluster-domain="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196606 4846 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196653 4846 flags.go:64] FLAG: --config-dir="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196664 4846 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196675 4846 flags.go:64] FLAG: --container-log-max-files="5" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196688 4846 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196697 4846 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196706 4846 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196716 4846 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196725 4846 flags.go:64] FLAG: --contention-profiling="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196735 4846 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196744 4846 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196754 4846 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196764 4846 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196776 4846 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196786 4846 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196795 4846 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196804 4846 flags.go:64] FLAG: --enable-load-reader="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196813 4846 flags.go:64] FLAG: --enable-server="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196822 4846 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196842 4846 flags.go:64] FLAG: --event-burst="100" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196851 4846 flags.go:64] FLAG: --event-qps="50" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196860 4846 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196869 4846 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196877 4846 flags.go:64] FLAG: --eviction-hard="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196889 4846 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196898 4846 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196906 4846 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196928 4846 flags.go:64] FLAG: --eviction-soft="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196938 4846 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196949 4846 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196961 4846 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196973 4846 flags.go:64] FLAG: --experimental-mounter-path="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196986 4846 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.196998 4846 flags.go:64] FLAG: --fail-swap-on="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197010 4846 flags.go:64] FLAG: --feature-gates="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197021 4846 flags.go:64] FLAG: --file-check-frequency="20s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197030 4846 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197040 4846 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197049 4846 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197058 4846 flags.go:64] FLAG: --healthz-port="10248" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197068 4846 flags.go:64] FLAG: --help="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197077 4846 flags.go:64] FLAG: --hostname-override="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197086 4846 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197095 4846 flags.go:64] FLAG: --http-check-frequency="20s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197104 4846 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197112 4846 flags.go:64] FLAG: --image-credential-provider-config="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197121 4846 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197130 4846 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197139 4846 flags.go:64] FLAG: --image-service-endpoint="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197147 4846 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197157 4846 flags.go:64] FLAG: --kube-api-burst="100" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197169 4846 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197181 4846 flags.go:64] FLAG: --kube-api-qps="50" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197192 4846 flags.go:64] FLAG: --kube-reserved="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197204 4846 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197215 4846 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197226 4846 flags.go:64] FLAG: --kubelet-cgroups="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197235 4846 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197244 4846 flags.go:64] FLAG: --lock-file="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197253 4846 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197262 4846 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197271 4846 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197286 4846 flags.go:64] FLAG: --log-json-split-stream="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197311 4846 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197320 4846 flags.go:64] FLAG: --log-text-split-stream="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197332 4846 flags.go:64] FLAG: --logging-format="text" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197341 4846 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197351 4846 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197360 4846 flags.go:64] FLAG: --manifest-url="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197368 4846 flags.go:64] FLAG: --manifest-url-header="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197382 4846 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197392 4846 flags.go:64] FLAG: --max-open-files="1000000" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197403 4846 flags.go:64] FLAG: --max-pods="110" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197412 4846 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197421 4846 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197429 4846 flags.go:64] FLAG: --memory-manager-policy="None" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197439 4846 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197449 4846 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197458 4846 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197467 4846 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197499 4846 flags.go:64] FLAG: --node-status-max-images="50" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197508 4846 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197517 4846 flags.go:64] FLAG: --oom-score-adj="-999" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197526 4846 flags.go:64] FLAG: --pod-cidr="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197535 4846 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197553 4846 flags.go:64] FLAG: --pod-manifest-path="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197563 4846 flags.go:64] FLAG: --pod-max-pids="-1" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197574 4846 flags.go:64] FLAG: --pods-per-core="0" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197583 4846 flags.go:64] FLAG: --port="10250" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197593 4846 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197602 4846 flags.go:64] FLAG: --provider-id="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197611 4846 flags.go:64] FLAG: --qos-reserved="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197650 4846 flags.go:64] FLAG: --read-only-port="10255" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197660 4846 flags.go:64] FLAG: --register-node="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197669 4846 flags.go:64] FLAG: --register-schedulable="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197679 4846 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197694 4846 flags.go:64] FLAG: --registry-burst="10" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197704 4846 flags.go:64] FLAG: --registry-qps="5" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197713 4846 flags.go:64] FLAG: --reserved-cpus="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197735 4846 flags.go:64] FLAG: --reserved-memory="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197748 4846 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197757 4846 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197767 4846 flags.go:64] FLAG: --rotate-certificates="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197776 4846 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197784 4846 flags.go:64] FLAG: --runonce="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197793 4846 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197802 4846 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197812 4846 flags.go:64] FLAG: --seccomp-default="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197822 4846 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197833 4846 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197845 4846 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197857 4846 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197869 4846 flags.go:64] FLAG: --storage-driver-password="root" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197878 4846 flags.go:64] FLAG: --storage-driver-secure="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197887 4846 flags.go:64] FLAG: --storage-driver-table="stats" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197895 4846 flags.go:64] FLAG: --storage-driver-user="root" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197904 4846 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197914 4846 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197923 4846 flags.go:64] FLAG: --system-cgroups="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197932 4846 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197949 4846 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197958 4846 flags.go:64] FLAG: --tls-cert-file="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197967 4846 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.197990 4846 flags.go:64] FLAG: --tls-min-version="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.198002 4846 flags.go:64] FLAG: --tls-private-key-file="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.198013 4846 flags.go:64] FLAG: --topology-manager-policy="none" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.198024 4846 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.198036 4846 flags.go:64] FLAG: --topology-manager-scope="container" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.198046 4846 flags.go:64] FLAG: --v="2" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.198058 4846 flags.go:64] FLAG: --version="false" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.198072 4846 flags.go:64] FLAG: --vmodule="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.198082 4846 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.198092 4846 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198375 4846 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198389 4846 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198417 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198426 4846 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198435 4846 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198483 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198493 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198501 4846 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198509 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198518 4846 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198526 4846 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198534 4846 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198545 4846 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198554 4846 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198562 4846 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198570 4846 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198577 4846 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198588 4846 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198598 4846 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198606 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198643 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198654 4846 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198663 4846 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198671 4846 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198678 4846 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198687 4846 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198695 4846 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198704 4846 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198711 4846 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198719 4846 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198727 4846 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198737 4846 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198746 4846 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198755 4846 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198764 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198772 4846 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198781 4846 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.198790 4846 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199008 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199020 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199031 4846 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199041 4846 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199052 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199061 4846 feature_gate.go:330] unrecognized feature gate: Example Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199072 4846 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199081 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199091 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199141 4846 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199149 4846 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199163 4846 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199176 4846 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199186 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199199 4846 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199231 4846 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199242 4846 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199252 4846 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199261 4846 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199270 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199279 4846 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199288 4846 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199297 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199306 4846 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199314 4846 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199323 4846 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199330 4846 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199338 4846 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199347 4846 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199356 4846 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199366 4846 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199376 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.199399 4846 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.199416 4846 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.209162 4846 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.209229 4846 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209393 4846 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209410 4846 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209415 4846 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209422 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209428 4846 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209434 4846 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209440 4846 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209446 4846 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209452 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209457 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209463 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209467 4846 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209473 4846 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209480 4846 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209487 4846 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209492 4846 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209498 4846 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209504 4846 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209510 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209515 4846 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209521 4846 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209526 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209531 4846 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209536 4846 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209541 4846 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209546 4846 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209551 4846 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209556 4846 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209560 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209566 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209571 4846 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209576 4846 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209581 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209585 4846 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209592 4846 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209597 4846 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209602 4846 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209607 4846 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209612 4846 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209618 4846 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209660 4846 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209665 4846 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209670 4846 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209675 4846 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209680 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209685 4846 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209690 4846 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209696 4846 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209700 4846 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209705 4846 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209710 4846 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209718 4846 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209729 4846 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209735 4846 feature_gate.go:330] unrecognized feature gate: Example Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209742 4846 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209749 4846 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209755 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209760 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209766 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209771 4846 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209777 4846 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209783 4846 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209789 4846 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209794 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209799 4846 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209805 4846 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209810 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209815 4846 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209820 4846 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209825 4846 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.209831 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.209841 4846 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210021 4846 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210032 4846 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210037 4846 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210043 4846 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210048 4846 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210053 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210058 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210064 4846 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210069 4846 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210075 4846 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210080 4846 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210085 4846 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210091 4846 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210096 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210101 4846 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210107 4846 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210112 4846 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210117 4846 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210122 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210127 4846 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210132 4846 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210137 4846 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210141 4846 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210146 4846 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210151 4846 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210156 4846 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210163 4846 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210169 4846 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210176 4846 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210182 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210189 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210195 4846 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210201 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210209 4846 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210217 4846 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210223 4846 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210228 4846 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210234 4846 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210240 4846 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210246 4846 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210251 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210256 4846 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210261 4846 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210267 4846 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210272 4846 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210281 4846 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210287 4846 feature_gate.go:330] unrecognized feature gate: Example Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210293 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210298 4846 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210304 4846 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210311 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210318 4846 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210325 4846 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210331 4846 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210336 4846 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210342 4846 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210347 4846 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210353 4846 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210358 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210398 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210404 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210411 4846 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210417 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210423 4846 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210429 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210435 4846 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210440 4846 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210446 4846 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210451 4846 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210457 4846 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.210464 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.210473 4846 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.210766 4846 server.go:940] "Client rotation is on, will bootstrap in background" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.215571 4846 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.215740 4846 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.217311 4846 server.go:997] "Starting client certificate rotation" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.217348 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.217735 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-04 09:30:58.335011734 +0000 UTC Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.217865 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.243893 4846 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.246251 4846 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.249498 4846 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.271397 4846 log.go:25] "Validated CRI v1 runtime API" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.309469 4846 log.go:25] "Validated CRI v1 image API" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.311550 4846 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.315666 4846 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-02-12-05-01-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.315703 4846 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.335576 4846 manager.go:217] Machine: {Timestamp:2026-02-02 12:09:29.332872826 +0000 UTC m=+0.561459709 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:292d225e-9d96-4eab-9003-afa615d76032 BootID:90d0bd8e-8fff-4676-a040-40a8eb821a29 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:1d:c3:0b Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:1d:c3:0b Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:b9:f3:07 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:a2:a6:a5 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:a0:7b:16 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:af:cb:54 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:cc:c4:ef Speed:-1 Mtu:1496} {Name:eth10 MacAddress:5e:35:3c:31:97:b6 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ce:fe:6b:77:e2:52 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.335976 4846 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.336217 4846 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.336607 4846 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.336856 4846 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.336900 4846 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.338548 4846 topology_manager.go:138] "Creating topology manager with none policy" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.338573 4846 container_manager_linux.go:303] "Creating device plugin manager" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.339026 4846 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.339057 4846 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.339355 4846 state_mem.go:36] "Initialized new in-memory state store" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.339466 4846 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.344470 4846 kubelet.go:418] "Attempting to sync node with API server" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.344515 4846 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.344534 4846 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.344550 4846 kubelet.go:324] "Adding apiserver pod source" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.344563 4846 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.348897 4846 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.349921 4846 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.352052 4846 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353724 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353758 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353775 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353786 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353803 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353811 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353818 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353834 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353849 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353859 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353891 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.353902 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.353907 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.354034 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.353908 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.354096 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.354960 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.355452 4846 server.go:1280] "Started kubelet" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.357069 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.357580 4846 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.357590 4846 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.358017 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.358049 4846 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.358121 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 12:11:14.421267229 +0000 UTC Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.358218 4846 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.358240 4846 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.358263 4846 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.358222 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.358322 4846 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 02 12:09:29 crc systemd[1]: Started Kubernetes Kubelet. Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.359088 4846 factory.go:55] Registering systemd factory Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.359112 4846 factory.go:221] Registration of the systemd container factory successfully Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.367328 4846 factory.go:153] Registering CRI-O factory Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.367397 4846 factory.go:221] Registration of the crio container factory successfully Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.369296 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="200ms" Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.370808 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.371810 4846 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.371903 4846 factory.go:103] Registering Raw factory Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.371826 4846 server.go:460] "Adding debug handlers to kubelet server" Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.375280 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.375360 4846 manager.go:1196] Started watching for new ooms in manager Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.376054 4846 manager.go:319] Starting recovery of all containers Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.375296 4846 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.223:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18906cb2f0bbb87f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 12:09:29.355417727 +0000 UTC m=+0.584004590,LastTimestamp:2026-02-02 12:09:29.355417727 +0000 UTC m=+0.584004590,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381276 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381379 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381402 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381421 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381437 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381452 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381469 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381486 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381504 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381520 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381536 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381554 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381569 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381587 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381602 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381648 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381671 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381687 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381703 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381720 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381739 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381755 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381771 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381787 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381802 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381816 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381835 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381852 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381866 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381882 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381895 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381913 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381928 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381945 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.381989 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382005 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382022 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382037 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382053 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382068 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382084 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382098 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382113 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382130 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382146 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382161 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382180 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382200 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382217 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382233 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382248 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382265 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382285 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382301 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382316 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382332 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382348 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382360 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382373 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382385 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382397 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382409 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382424 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382453 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382470 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382482 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382494 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382508 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382520 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382531 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382544 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382559 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382571 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382584 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382595 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382609 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382664 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382677 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382689 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382707 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382722 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382735 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382749 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382761 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382773 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382784 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382796 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382809 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382823 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382834 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382845 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382857 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382877 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382888 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382901 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382914 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382927 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382938 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382974 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.382989 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383006 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383026 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383037 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383049 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383072 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383085 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383097 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383110 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383125 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383138 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383150 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383164 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383175 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383186 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383197 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383207 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383219 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383233 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383245 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383258 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383272 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383284 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383295 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383306 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383325 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383339 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383353 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383368 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383381 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383391 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383401 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383413 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383426 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383438 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383454 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383465 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383479 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383492 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383507 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383520 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383532 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383543 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383554 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383566 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383578 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383590 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383606 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383637 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383649 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383663 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383676 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383689 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383701 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383715 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383727 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383741 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383758 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383772 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383787 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383801 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383816 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383829 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383842 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383855 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383871 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383884 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383901 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383914 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383927 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383941 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383956 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383971 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383982 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.383997 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.385946 4846 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.385976 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.385989 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386000 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386010 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386028 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386039 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386050 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386060 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386071 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386081 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386090 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386100 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386114 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386124 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386134 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386144 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386165 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386185 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386198 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386213 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386225 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386235 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386249 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386261 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386271 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386288 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386302 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386314 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386326 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386336 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386346 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386357 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386368 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386376 4846 reconstruct.go:97] "Volume reconstruction finished" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.386383 4846 reconciler.go:26] "Reconciler: start to sync state" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.393242 4846 manager.go:324] Recovery completed Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.408092 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.409887 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.409943 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.409957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.410883 4846 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.410901 4846 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.410945 4846 state_mem.go:36] "Initialized new in-memory state store" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.420283 4846 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.422141 4846 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.422194 4846 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.422224 4846 kubelet.go:2335] "Starting kubelet main sync loop" Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.422273 4846 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.422856 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.422914 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.428459 4846 policy_none.go:49] "None policy: Start" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.430126 4846 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.430164 4846 state_mem.go:35] "Initializing new in-memory state store" Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.458487 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.479839 4846 manager.go:334] "Starting Device Plugin manager" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.480079 4846 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.480103 4846 server.go:79] "Starting device plugin registration server" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.480820 4846 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.480841 4846 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.481141 4846 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.481367 4846 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.481383 4846 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.496011 4846 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.523235 4846 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.523344 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.524826 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.524893 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.524909 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.525100 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.526309 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.526349 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.526359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.526368 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.526420 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.526513 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.526812 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.526904 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528053 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528092 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528109 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528233 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528263 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528274 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528349 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528570 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528597 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528599 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528630 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.528639 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.529599 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.529637 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.529645 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.529705 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.529722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.529732 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.529772 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.529881 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.529906 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.531208 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.531223 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.531231 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.531319 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.531342 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.531355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.531533 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.531569 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.533006 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.533039 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.533051 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.570484 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="400ms" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.581548 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.582848 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.582890 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.582907 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.582938 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.583578 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.223:6443: connect: connection refused" node="crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590335 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590437 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590476 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590573 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590653 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590731 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590758 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590817 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590863 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590889 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590955 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.590981 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.591000 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.591043 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692450 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692551 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692573 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692596 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692642 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692666 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692690 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692709 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692726 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692741 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692758 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692774 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692788 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692807 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692793 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692831 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692907 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692978 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692989 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.693019 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692957 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692991 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.693066 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.693037 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.693087 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.693066 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.693056 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.693089 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.692948 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.693115 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.784573 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.786243 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.786297 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.786315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.786348 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.787274 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.223:6443: connect: connection refused" node="crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.865560 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.885521 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.913150 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.921278 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-653a38c95a6ff5df49bdc2731bf842302522ef0979c3ed5374b8fd37cae65934 WatchSource:0}: Error finding container 653a38c95a6ff5df49bdc2731bf842302522ef0979c3ed5374b8fd37cae65934: Status 404 returned error can't find the container with id 653a38c95a6ff5df49bdc2731bf842302522ef0979c3ed5374b8fd37cae65934 Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.932253 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.932959 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-d2f825f751c703c8a7721ff11c3c38e56e997e88d004ed133b529ab033594edb WatchSource:0}: Error finding container d2f825f751c703c8a7721ff11c3c38e56e997e88d004ed133b529ab033594edb: Status 404 returned error can't find the container with id d2f825f751c703c8a7721ff11c3c38e56e997e88d004ed133b529ab033594edb Feb 02 12:09:29 crc kubenswrapper[4846]: I0202 12:09:29.939993 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.947786 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-efd0a48e5a19dbe3bb25971da65b8a7618de607ca3973384db48c4e4e7ea3086 WatchSource:0}: Error finding container efd0a48e5a19dbe3bb25971da65b8a7618de607ca3973384db48c4e4e7ea3086: Status 404 returned error can't find the container with id efd0a48e5a19dbe3bb25971da65b8a7618de607ca3973384db48c4e4e7ea3086 Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.952604 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-27feb736f4ad9553012cddbd8f1440bdce91f6e4942b2ae8486f0d6b1393d170 WatchSource:0}: Error finding container 27feb736f4ad9553012cddbd8f1440bdce91f6e4942b2ae8486f0d6b1393d170: Status 404 returned error can't find the container with id 27feb736f4ad9553012cddbd8f1440bdce91f6e4942b2ae8486f0d6b1393d170 Feb 02 12:09:29 crc kubenswrapper[4846]: W0202 12:09:29.958497 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a0fa76013e3eac4763fb4c3c76680202868880e55fb44f2f337a9ad459040c23 WatchSource:0}: Error finding container a0fa76013e3eac4763fb4c3c76680202868880e55fb44f2f337a9ad459040c23: Status 404 returned error can't find the container with id a0fa76013e3eac4763fb4c3c76680202868880e55fb44f2f337a9ad459040c23 Feb 02 12:09:29 crc kubenswrapper[4846]: E0202 12:09:29.971931 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="800ms" Feb 02 12:09:30 crc kubenswrapper[4846]: W0202 12:09:30.168231 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:30 crc kubenswrapper[4846]: E0202 12:09:30.168319 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.187675 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.189129 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.189165 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.189175 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.189200 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 12:09:30 crc kubenswrapper[4846]: E0202 12:09:30.189638 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.223:6443: connect: connection refused" node="crc" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.358113 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.358219 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 07:57:29.880695822 +0000 UTC Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.429406 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"653a38c95a6ff5df49bdc2731bf842302522ef0979c3ed5374b8fd37cae65934"} Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.430852 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a0fa76013e3eac4763fb4c3c76680202868880e55fb44f2f337a9ad459040c23"} Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.432194 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"27feb736f4ad9553012cddbd8f1440bdce91f6e4942b2ae8486f0d6b1393d170"} Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.434392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"efd0a48e5a19dbe3bb25971da65b8a7618de607ca3973384db48c4e4e7ea3086"} Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.435677 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d2f825f751c703c8a7721ff11c3c38e56e997e88d004ed133b529ab033594edb"} Feb 02 12:09:30 crc kubenswrapper[4846]: W0202 12:09:30.482772 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:30 crc kubenswrapper[4846]: E0202 12:09:30.482879 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:30 crc kubenswrapper[4846]: W0202 12:09:30.646837 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:30 crc kubenswrapper[4846]: E0202 12:09:30.646919 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:30 crc kubenswrapper[4846]: E0202 12:09:30.772642 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="1.6s" Feb 02 12:09:30 crc kubenswrapper[4846]: W0202 12:09:30.857972 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:30 crc kubenswrapper[4846]: E0202 12:09:30.858101 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.990087 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.992067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.992123 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.992134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:30 crc kubenswrapper[4846]: I0202 12:09:30.992156 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 12:09:30 crc kubenswrapper[4846]: E0202 12:09:30.992854 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.223:6443: connect: connection refused" node="crc" Feb 02 12:09:31 crc kubenswrapper[4846]: E0202 12:09:31.145429 4846 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.223:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18906cb2f0bbb87f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 12:09:29.355417727 +0000 UTC m=+0.584004590,LastTimestamp:2026-02-02 12:09:29.355417727 +0000 UTC m=+0.584004590,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.331131 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 02 12:09:31 crc kubenswrapper[4846]: E0202 12:09:31.332963 4846 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.358542 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 22:29:50.303749872 +0000 UTC Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.358612 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.440066 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9" exitCode=0 Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.440173 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9"} Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.440233 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.441994 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.442046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.442069 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.443604 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544"} Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.443692 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac"} Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.443709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e"} Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.443728 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406"} Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.443643 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.444211 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.444669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.444690 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.444701 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.445067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.445094 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.445107 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.446644 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde" exitCode=0 Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.446708 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde"} Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.446781 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.447667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.447704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.447716 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.450028 4846 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752" exitCode=0 Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.450056 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752"} Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.450182 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.451323 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.451364 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.451379 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.453211 4846 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6" exitCode=0 Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.453265 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6"} Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.453356 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.454463 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.454526 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:31 crc kubenswrapper[4846]: I0202 12:09:31.454544 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.358552 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.365240 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 23:27:51.452734498 +0000 UTC Feb 02 12:09:32 crc kubenswrapper[4846]: E0202 12:09:32.374189 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="3.2s" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.460356 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89"} Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.460408 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a"} Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.460411 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.460420 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1"} Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.461508 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.461544 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.461555 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.464938 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ba1e5f01aa5d5e17829be20d5a74da3508b73a3c502b9d6ef89d6362cb01bd8c"} Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.464973 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.464981 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75"} Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.464997 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979"} Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.465012 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a"} Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.465024 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138"} Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.465811 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.465840 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.465850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.467880 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b835d20acf3590c96063ca2b91a8f02cd5a374811972692fef36d29619697d20"} Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.467965 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.468897 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.468921 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.468930 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.471183 4846 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58" exitCode=0 Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.471266 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.471661 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.471939 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58"} Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.472436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.472464 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.472475 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.472993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.473014 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.473022 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.593572 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.594976 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.595173 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.595222 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:32 crc kubenswrapper[4846]: I0202 12:09:32.595253 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 12:09:32 crc kubenswrapper[4846]: E0202 12:09:32.596308 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.223:6443: connect: connection refused" node="crc" Feb 02 12:09:32 crc kubenswrapper[4846]: W0202 12:09:32.739524 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:32 crc kubenswrapper[4846]: E0202 12:09:32.739597 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:32 crc kubenswrapper[4846]: W0202 12:09:32.981387 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:32 crc kubenswrapper[4846]: E0202 12:09:32.981526 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:33 crc kubenswrapper[4846]: W0202 12:09:33.202582 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.223:6443: connect: connection refused Feb 02 12:09:33 crc kubenswrapper[4846]: E0202 12:09:33.202714 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.223:6443: connect: connection refused" logger="UnhandledError" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.366210 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 05:11:54.878482812 +0000 UTC Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.481599 4846 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2" exitCode=0 Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.481694 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2"} Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.481739 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.482804 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.482849 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.482862 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.485548 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.488436 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ba1e5f01aa5d5e17829be20d5a74da3508b73a3c502b9d6ef89d6362cb01bd8c" exitCode=255 Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.488529 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ba1e5f01aa5d5e17829be20d5a74da3508b73a3c502b9d6ef89d6362cb01bd8c"} Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.488581 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.488609 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.488654 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.489554 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.489983 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.490002 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.490022 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.490028 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.490038 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.490045 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.490778 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.490918 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.491020 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.491717 4846 scope.go:117] "RemoveContainer" containerID="ba1e5f01aa5d5e17829be20d5a74da3508b73a3c502b9d6ef89d6362cb01bd8c" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.754611 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.754782 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.755781 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.755821 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:33 crc kubenswrapper[4846]: I0202 12:09:33.755829 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.367185 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 18:05:39.410062409 +0000 UTC Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.496741 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067"} Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.496785 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85"} Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.496796 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36"} Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.498667 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.500743 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7"} Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.500872 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.500926 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.500875 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.502117 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.502145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.502156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.502176 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.502191 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:34 crc kubenswrapper[4846]: I0202 12:09:34.502203 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.368331 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 22:54:06.282773985 +0000 UTC Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.507152 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.507181 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f"} Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.507217 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.507230 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86"} Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.507152 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.508051 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.508090 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.508104 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.508727 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.508757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.508768 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.709773 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.742213 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.742397 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.743501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.743538 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.743548 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.797134 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.798366 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.798406 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.798414 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:35 crc kubenswrapper[4846]: I0202 12:09:35.798447 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.334784 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.369485 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 01:39:02.110631501 +0000 UTC Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.509245 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.509245 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.510394 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.510502 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.510614 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.510744 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.510669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.510961 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.559211 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.559711 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.561098 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.561147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.561163 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:36 crc kubenswrapper[4846]: I0202 12:09:36.568013 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.200252 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.370368 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 09:39:38.127259151 +0000 UTC Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.510896 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.510924 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.510996 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.511845 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.511876 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.511889 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.512012 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.512030 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:37 crc kubenswrapper[4846]: I0202 12:09:37.512042 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.176276 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.176472 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.177699 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.177743 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.177757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.371027 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 21:56:33.513962373 +0000 UTC Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.514054 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.514054 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.515261 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.515296 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.515307 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.515756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.515811 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.515828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.743438 4846 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 12:09:38 crc kubenswrapper[4846]: I0202 12:09:38.743523 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 12:09:39 crc kubenswrapper[4846]: I0202 12:09:39.371515 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 15:55:27.583229556 +0000 UTC Feb 02 12:09:39 crc kubenswrapper[4846]: E0202 12:09:39.496220 4846 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 02 12:09:40 crc kubenswrapper[4846]: I0202 12:09:40.371818 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 00:09:28.66370934 +0000 UTC Feb 02 12:09:40 crc kubenswrapper[4846]: I0202 12:09:40.916203 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:40 crc kubenswrapper[4846]: I0202 12:09:40.916401 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:40 crc kubenswrapper[4846]: I0202 12:09:40.917952 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:40 crc kubenswrapper[4846]: I0202 12:09:40.918037 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:40 crc kubenswrapper[4846]: I0202 12:09:40.918052 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:41 crc kubenswrapper[4846]: I0202 12:09:41.042344 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 02 12:09:41 crc kubenswrapper[4846]: I0202 12:09:41.042597 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:41 crc kubenswrapper[4846]: I0202 12:09:41.044190 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:41 crc kubenswrapper[4846]: I0202 12:09:41.044222 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:41 crc kubenswrapper[4846]: I0202 12:09:41.044232 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:41 crc kubenswrapper[4846]: I0202 12:09:41.372201 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 04:37:16.464277842 +0000 UTC Feb 02 12:09:42 crc kubenswrapper[4846]: I0202 12:09:42.372730 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 15:48:25.184786476 +0000 UTC Feb 02 12:09:43 crc kubenswrapper[4846]: I0202 12:09:43.359830 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 02 12:09:43 crc kubenswrapper[4846]: I0202 12:09:43.372952 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 04:20:09.897981059 +0000 UTC Feb 02 12:09:43 crc kubenswrapper[4846]: I0202 12:09:43.427103 4846 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 02 12:09:43 crc kubenswrapper[4846]: I0202 12:09:43.427170 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 02 12:09:43 crc kubenswrapper[4846]: I0202 12:09:43.431912 4846 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 02 12:09:43 crc kubenswrapper[4846]: I0202 12:09:43.431986 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 02 12:09:44 crc kubenswrapper[4846]: I0202 12:09:44.373721 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 11:30:33.002876802 +0000 UTC Feb 02 12:09:45 crc kubenswrapper[4846]: I0202 12:09:45.374323 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 21:30:26.093002416 +0000 UTC Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.220663 4846 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.221083 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.342519 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.342790 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.343283 4846 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.343359 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.344389 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.344456 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.344479 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.347972 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.374931 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 07:22:32.856219205 +0000 UTC Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.532800 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.533408 4846 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.533483 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.534092 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.534283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:46 crc kubenswrapper[4846]: I0202 12:09:46.534506 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:47 crc kubenswrapper[4846]: I0202 12:09:47.375602 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 11:00:02.475042088 +0000 UTC Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.375810 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 04:58:46.018037165 +0000 UTC Feb 02 12:09:48 crc kubenswrapper[4846]: E0202 12:09:48.413274 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.414224 4846 trace.go:236] Trace[1982962068]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 12:09:33.497) (total time: 14916ms): Feb 02 12:09:48 crc kubenswrapper[4846]: Trace[1982962068]: ---"Objects listed" error: 14916ms (12:09:48.414) Feb 02 12:09:48 crc kubenswrapper[4846]: Trace[1982962068]: [14.916252873s] [14.916252873s] END Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.414249 4846 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.414444 4846 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.414931 4846 trace.go:236] Trace[985609934]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 12:09:37.993) (total time: 10421ms): Feb 02 12:09:48 crc kubenswrapper[4846]: Trace[985609934]: ---"Objects listed" error: 10421ms (12:09:48.414) Feb 02 12:09:48 crc kubenswrapper[4846]: Trace[985609934]: [10.421091972s] [10.421091972s] END Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.414976 4846 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.415065 4846 trace.go:236] Trace[401208868]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 12:09:37.851) (total time: 10563ms): Feb 02 12:09:48 crc kubenswrapper[4846]: Trace[401208868]: ---"Objects listed" error: 10563ms (12:09:48.414) Feb 02 12:09:48 crc kubenswrapper[4846]: Trace[401208868]: [10.563552902s] [10.563552902s] END Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.415103 4846 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 02 12:09:48 crc kubenswrapper[4846]: E0202 12:09:48.417226 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.418101 4846 trace.go:236] Trace[990183916]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 12:09:38.399) (total time: 10018ms): Feb 02 12:09:48 crc kubenswrapper[4846]: Trace[990183916]: ---"Objects listed" error: 10018ms (12:09:48.417) Feb 02 12:09:48 crc kubenswrapper[4846]: Trace[990183916]: [10.018852074s] [10.018852074s] END Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.418136 4846 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.426141 4846 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.478274 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:48 crc kubenswrapper[4846]: I0202 12:09:48.485012 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:48 crc kubenswrapper[4846]: E0202 12:09:48.542956 4846 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.357121 4846 apiserver.go:52] "Watching apiserver" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.359593 4846 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.359921 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.360294 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.360397 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.360492 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.360540 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.360584 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.360820 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.360862 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.360887 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.360923 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.362107 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.362454 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.363705 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.363998 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.364239 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.364394 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.364491 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.364919 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.369197 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.376669 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 09:56:21.826453254 +0000 UTC Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.402588 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.417450 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.429427 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.437528 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.445782 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.453089 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.459409 4846 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.461438 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.470444 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.477464 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.485111 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.494584 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.505871 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.515366 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.522912 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.522970 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.522997 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523051 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523089 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523128 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523163 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523207 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523246 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523282 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523314 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523343 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523374 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523404 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523434 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523464 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523493 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523484 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523542 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523532 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523668 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523711 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523732 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523748 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523765 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523781 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523797 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523841 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523878 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523894 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523909 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523925 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523941 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523957 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523970 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.523979 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524043 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524078 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524082 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524136 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524161 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524186 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524210 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524235 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524256 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524366 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524392 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524412 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524434 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524456 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524956 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526593 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526643 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526677 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526710 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526736 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526769 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526797 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526822 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526920 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527019 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527048 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527135 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527171 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527201 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527269 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527298 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527329 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527355 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527391 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527423 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527449 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527852 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527888 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527919 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527941 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527972 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528000 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528031 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528060 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528135 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528169 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528202 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528232 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528263 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528297 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528327 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528386 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528412 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528442 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528512 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528544 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528575 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528610 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528553 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528648 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528680 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528767 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528797 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528823 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528913 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528945 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528971 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529002 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529030 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529058 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529084 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529114 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529143 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529170 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529202 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524090 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524322 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529977 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524434 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524496 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524817 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.524979 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.525015 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.525167 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.525199 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.525237 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.525416 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530158 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.525508 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.525993 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526130 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526364 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526401 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526395 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526555 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526816 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526831 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.526877 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527020 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527114 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527059 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527396 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527455 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528005 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530533 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.527022 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528030 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528177 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528191 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528468 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528692 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528733 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.528785 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529120 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.529239 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:09:50.029213707 +0000 UTC m=+21.257800570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530698 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530731 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530855 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530877 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530840 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530899 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531014 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531027 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531117 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531097 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531274 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531316 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531355 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531386 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531419 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531453 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531481 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531512 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531540 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531571 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.532104 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531266 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531716 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529298 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529476 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529599 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529837 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529890 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529934 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530052 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.530524 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531751 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.533560 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.532165 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531813 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.533646 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.531964 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.532073 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.532650 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.532696 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.532757 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.529286 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.533180 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.533765 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.533946 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.534104 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.534106 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.534235 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.534250 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.534371 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.534404 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.534967 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.534975 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.534998 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.534987 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.535478 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.535749 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.535794 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.536369 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.536509 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.536379 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.536706 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.536860 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.536866 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.536923 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537560 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537606 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537653 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537680 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537702 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537737 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537758 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537779 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537800 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537824 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537854 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537876 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537895 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537928 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537951 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537973 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538000 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538021 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538042 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538064 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538107 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538129 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538160 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538180 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538201 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538222 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538255 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538276 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538299 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538322 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538343 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538364 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538400 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538433 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538456 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538490 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538517 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538538 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538580 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538606 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539461 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539497 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539521 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539543 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539569 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539644 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539668 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539690 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539714 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539738 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539760 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537649 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539786 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537695 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537827 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537892 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.537985 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538038 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538195 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539869 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538273 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538291 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538308 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538525 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538555 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538685 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538810 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538893 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538953 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.538959 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539361 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539425 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540246 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539811 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539551 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539945 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.539968 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540085 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540109 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540266 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540336 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540389 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540401 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540427 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540448 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540470 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540493 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540514 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540537 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540555 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540575 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540617 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540684 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540707 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540727 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540747 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540768 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540816 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540852 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540876 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540897 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540951 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541007 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541034 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541059 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541081 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541104 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541125 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541148 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541170 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541192 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541288 4846 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541307 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541320 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541333 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541344 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541356 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541676 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541700 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541714 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541727 4846 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541738 4846 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541750 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541762 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541777 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541790 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541804 4846 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541816 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541828 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541841 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541854 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541865 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541877 4846 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541890 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541901 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541913 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541925 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541938 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541950 4846 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541962 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541974 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541986 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541999 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542013 4846 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542025 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542036 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542048 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542057 4846 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542089 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542099 4846 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542109 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542120 4846 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542132 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542142 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542152 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542161 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542169 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542179 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542188 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542196 4846 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542205 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542216 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542228 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542239 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542466 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542486 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542498 4846 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542533 4846 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542545 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542557 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542569 4846 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542581 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542594 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542605 4846 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542617 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542644 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542656 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542668 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542684 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542696 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542708 4846 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542720 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542732 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542744 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542757 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542769 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542781 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542795 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542808 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542821 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542833 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542845 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542857 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542870 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542912 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542924 4846 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542936 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542995 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543009 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543023 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543035 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543046 4846 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543057 4846 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543069 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543082 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543092 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540400 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540488 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540564 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540779 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540904 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540948 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541479 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541480 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541562 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.541691 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542032 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542051 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542271 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542392 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542582 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542589 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542641 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.542710 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543093 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543308 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543440 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543970 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.544039 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.544045 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.544464 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.544488 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.544498 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.544803 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543880 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.544853 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.545317 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.545641 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.545687 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.545805 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.545497 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.546003 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.546120 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.546470 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.546481 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.546594 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.546709 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.546783 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.546951 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.547041 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.547307 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.547390 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.547977 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.548003 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.548598 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.549423 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.550338 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.550550 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.550785 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.551226 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.551280 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.551907 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.551933 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.551991 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.552048 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.552071 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:50.052042687 +0000 UTC m=+21.280629620 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.552095 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:50.052083148 +0000 UTC m=+21.280670091 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.540106 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.552569 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.544815 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.552934 4846 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.553168 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.553538 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.553692 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.553933 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.554049 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.554314 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.554425 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.554815 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.554845 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.543102 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.554898 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.554903 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.554918 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.554968 4846 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.554993 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555011 4846 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555029 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555053 4846 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555069 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555088 4846 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555104 4846 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555119 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555134 4846 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555149 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555252 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555437 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555459 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555165 4846 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555494 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555514 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555536 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.555841 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.557132 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.557556 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.557576 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.557589 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.557663 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:50.057641045 +0000 UTC m=+21.286227908 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.560222 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.560774 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.561532 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.562926 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.562946 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.562957 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.563045 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:50.063031898 +0000 UTC m=+21.291618761 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.563592 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.565249 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7" exitCode=255 Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.565555 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7"} Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.565604 4846 scope.go:117] "RemoveContainer" containerID="ba1e5f01aa5d5e17829be20d5a74da3508b73a3c502b9d6ef89d6362cb01bd8c" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.567117 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.567245 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.573666 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.577423 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.577761 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.581066 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.581075 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.581568 4846 scope.go:117] "RemoveContainer" containerID="ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7" Feb 02 12:09:49 crc kubenswrapper[4846]: E0202 12:09:49.581792 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.582335 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.586301 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.592123 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.600430 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.601430 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.604703 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.610742 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.618559 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.625557 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.634399 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba1e5f01aa5d5e17829be20d5a74da3508b73a3c502b9d6ef89d6362cb01bd8c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"message\\\":\\\"W0202 12:09:32.518882 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 12:09:32.519236 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770034172 cert, and key in /tmp/serving-cert-2894499281/serving-signer.crt, /tmp/serving-cert-2894499281/serving-signer.key\\\\nI0202 12:09:32.924167 1 observer_polling.go:159] Starting file observer\\\\nW0202 12:09:32.930003 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 12:09:32.930335 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:32.933092 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2894499281/tls.crt::/tmp/serving-cert-2894499281/tls.key\\\\\\\"\\\\nF0202 12:09:33.112902 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.642646 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.651004 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656268 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656398 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656498 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656515 4846 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656537 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656562 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656571 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656582 4846 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656593 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656602 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656610 4846 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656638 4846 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656647 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656656 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656665 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656673 4846 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656682 4846 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656690 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656699 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656708 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656717 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656725 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656735 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656743 4846 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656751 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656759 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656768 4846 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656776 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656784 4846 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656792 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656801 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656808 4846 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656817 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656825 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656833 4846 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656841 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656849 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656857 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656865 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656872 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656881 4846 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656889 4846 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656896 4846 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656904 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656912 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656920 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656928 4846 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656935 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656943 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656952 4846 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656960 4846 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656969 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656990 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.656998 4846 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657006 4846 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657013 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657023 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657032 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657041 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657050 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657058 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657066 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657074 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657083 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657091 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657099 4846 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657107 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657115 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657123 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657131 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657138 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657147 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657156 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657163 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657172 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657179 4846 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657187 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657195 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657203 4846 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657211 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657218 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657226 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657234 4846 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657241 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.657249 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.686486 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.692497 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 12:09:49 crc kubenswrapper[4846]: W0202 12:09:49.696256 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-b98df30cd0b61945532aa660e85f4e9655b38e6b8cb7c1d46c043ca528dd08a1 WatchSource:0}: Error finding container b98df30cd0b61945532aa660e85f4e9655b38e6b8cb7c1d46c043ca528dd08a1: Status 404 returned error can't find the container with id b98df30cd0b61945532aa660e85f4e9655b38e6b8cb7c1d46c043ca528dd08a1 Feb 02 12:09:49 crc kubenswrapper[4846]: I0202 12:09:49.698562 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 12:09:49 crc kubenswrapper[4846]: W0202 12:09:49.712791 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-83f1c839a69012a6a409454d549b4fa76d5863fc5f1d06144980c187ed9e465e WatchSource:0}: Error finding container 83f1c839a69012a6a409454d549b4fa76d5863fc5f1d06144980c187ed9e465e: Status 404 returned error can't find the container with id 83f1c839a69012a6a409454d549b4fa76d5863fc5f1d06144980c187ed9e465e Feb 02 12:09:49 crc kubenswrapper[4846]: W0202 12:09:49.715268 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-c53b99e924948a27fd3a74e50f884e8ca7312fa5f1c608c0aab7870a2cc34994 WatchSource:0}: Error finding container c53b99e924948a27fd3a74e50f884e8ca7312fa5f1c608c0aab7870a2cc34994: Status 404 returned error can't find the container with id c53b99e924948a27fd3a74e50f884e8ca7312fa5f1c608c0aab7870a2cc34994 Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.060395 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.060674 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:09:51.060639173 +0000 UTC m=+22.289226046 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.060846 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.060901 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.060961 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.061060 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.061071 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.061129 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:51.061118904 +0000 UTC m=+22.289705767 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.061148 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:51.061140914 +0000 UTC m=+22.289727777 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.061180 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.061205 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.061219 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.061337 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:51.061282658 +0000 UTC m=+22.289869541 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.162272 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.162451 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.162486 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.162497 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.162564 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:51.162538457 +0000 UTC m=+22.391125330 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.378518 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 22:12:38.618560874 +0000 UTC Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.571964 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d"} Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.572018 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"83f1c839a69012a6a409454d549b4fa76d5863fc5f1d06144980c187ed9e465e"} Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.577835 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430"} Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.577904 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9"} Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.577921 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b98df30cd0b61945532aa660e85f4e9655b38e6b8cb7c1d46c043ca528dd08a1"} Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.581462 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.584578 4846 scope.go:117] "RemoveContainer" containerID="ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7" Feb 02 12:09:50 crc kubenswrapper[4846]: E0202 12:09:50.584839 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.584875 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c53b99e924948a27fd3a74e50f884e8ca7312fa5f1c608c0aab7870a2cc34994"} Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.594476 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba1e5f01aa5d5e17829be20d5a74da3508b73a3c502b9d6ef89d6362cb01bd8c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"message\\\":\\\"W0202 12:09:32.518882 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0202 12:09:32.519236 1 crypto.go:601] Generating new CA for check-endpoints-signer@1770034172 cert, and key in /tmp/serving-cert-2894499281/serving-signer.crt, /tmp/serving-cert-2894499281/serving-signer.key\\\\nI0202 12:09:32.924167 1 observer_polling.go:159] Starting file observer\\\\nW0202 12:09:32.930003 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0202 12:09:32.930335 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:32.933092 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2894499281/tls.crt::/tmp/serving-cert-2894499281/tls.key\\\\\\\"\\\\nF0202 12:09:33.112902 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.614802 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.630494 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.643382 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.656171 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.669380 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.681277 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.691574 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.703216 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.718019 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.735034 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.748268 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.759504 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.773194 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.793478 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:50 crc kubenswrapper[4846]: I0202 12:09:50.809804 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.071401 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.071477 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.071507 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.071570 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.071592 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:09:53.07155451 +0000 UTC m=+24.300141403 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.071639 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.071642 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:53.071626272 +0000 UTC m=+24.300213265 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.071728 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:53.071710154 +0000 UTC m=+24.300297117 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.071751 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.071841 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.071855 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.071867 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.071896 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:53.071887817 +0000 UTC m=+24.300474680 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.076582 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.091659 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.095843 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.104888 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.117759 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.131801 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.148127 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.160796 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.161538 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.172049 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.172194 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.172215 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.172227 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.172291 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:53.172273358 +0000 UTC m=+24.400860231 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.175318 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.188840 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.203630 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.214159 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.223662 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.244140 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.257342 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.270416 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.291343 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.315056 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.332401 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.379434 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 13:03:09.374185222 +0000 UTC Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.422992 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.423017 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.423008 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.423112 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.423180 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:09:51 crc kubenswrapper[4846]: E0202 12:09:51.423254 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.426960 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.427590 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.428413 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.429208 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.429890 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.430537 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.431233 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.431916 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.432699 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.433357 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.434017 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.434839 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.435335 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.435879 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.436489 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.437162 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.438784 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.439350 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.440042 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.440777 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.441303 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.441860 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.442269 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.442886 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.443259 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.443832 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.444404 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.444861 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.445401 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.445891 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.446317 4846 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.446408 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.447637 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.450769 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.451173 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.452610 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.453599 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.454088 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.455251 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.456322 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.456990 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.457781 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.458564 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.459316 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.459902 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.460525 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.461145 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.462065 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.462707 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.463288 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.463880 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.464487 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.465192 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 02 12:09:51 crc kubenswrapper[4846]: I0202 12:09:51.467156 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.380162 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 03:59:55.799801141 +0000 UTC Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.592416 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89"} Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.608978 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.626910 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.640983 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.667394 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.685006 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.702265 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.716070 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.728677 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:52 crc kubenswrapper[4846]: I0202 12:09:52.744581 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.088575 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.088719 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.088777 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.088802 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.088881 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.088945 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:57.088925242 +0000 UTC m=+28.317512105 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.088989 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.089027 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.089028 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.089048 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.089120 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:09:57.089093425 +0000 UTC m=+28.317680328 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.089159 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:57.089145197 +0000 UTC m=+28.317732150 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.089181 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:57.089169677 +0000 UTC m=+28.317756670 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.189546 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.189724 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.189758 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.189771 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.189834 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 12:09:57.189815223 +0000 UTC m=+28.418402146 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.381300 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 07:03:53.812533378 +0000 UTC Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.423388 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.423399 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.423523 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.423651 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.423736 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:09:53 crc kubenswrapper[4846]: E0202 12:09:53.423805 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.695283 4846 csr.go:261] certificate signing request csr-dwbgx is approved, waiting to be issued Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.741986 4846 csr.go:257] certificate signing request csr-dwbgx is issued Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.819288 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-44zpn"] Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.819766 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-44zpn" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.821891 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.822064 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.824342 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.836065 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:53Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.847406 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:53Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.861170 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:53Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.872105 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:53Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.891534 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:53Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.895183 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/40f86b56-873f-4c8a-81e2-abc5643f4869-hosts-file\") pod \"node-resolver-44zpn\" (UID: \"40f86b56-873f-4c8a-81e2-abc5643f4869\") " pod="openshift-dns/node-resolver-44zpn" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.895253 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whqc6\" (UniqueName: \"kubernetes.io/projected/40f86b56-873f-4c8a-81e2-abc5643f4869-kube-api-access-whqc6\") pod \"node-resolver-44zpn\" (UID: \"40f86b56-873f-4c8a-81e2-abc5643f4869\") " pod="openshift-dns/node-resolver-44zpn" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.904738 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:53Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.917522 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:53Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.930933 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:53Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.944010 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:53Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.955292 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:53Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.995884 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whqc6\" (UniqueName: \"kubernetes.io/projected/40f86b56-873f-4c8a-81e2-abc5643f4869-kube-api-access-whqc6\") pod \"node-resolver-44zpn\" (UID: \"40f86b56-873f-4c8a-81e2-abc5643f4869\") " pod="openshift-dns/node-resolver-44zpn" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.995980 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/40f86b56-873f-4c8a-81e2-abc5643f4869-hosts-file\") pod \"node-resolver-44zpn\" (UID: \"40f86b56-873f-4c8a-81e2-abc5643f4869\") " pod="openshift-dns/node-resolver-44zpn" Feb 02 12:09:53 crc kubenswrapper[4846]: I0202 12:09:53.996077 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/40f86b56-873f-4c8a-81e2-abc5643f4869-hosts-file\") pod \"node-resolver-44zpn\" (UID: \"40f86b56-873f-4c8a-81e2-abc5643f4869\") " pod="openshift-dns/node-resolver-44zpn" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.015266 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whqc6\" (UniqueName: \"kubernetes.io/projected/40f86b56-873f-4c8a-81e2-abc5643f4869-kube-api-access-whqc6\") pod \"node-resolver-44zpn\" (UID: \"40f86b56-873f-4c8a-81e2-abc5643f4869\") " pod="openshift-dns/node-resolver-44zpn" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.134401 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-44zpn" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.302041 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-7jshv"] Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.302415 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.305061 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.305168 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.305200 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.305397 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.305434 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.317415 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.330773 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.343066 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.346975 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-dhzp7"] Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.347239 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-6jrkw"] Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.347326 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.348022 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.351464 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.351468 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.351680 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.351796 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.352053 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.352696 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.356345 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.356462 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.366959 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.380564 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.381531 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 15:55:47.563064401 +0000 UTC Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.398945 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05ed703a-a310-418e-8c5d-aeddb32977cf-mcd-auth-proxy-config\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.398984 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9fb5\" (UniqueName: \"kubernetes.io/projected/05ed703a-a310-418e-8c5d-aeddb32977cf-kube-api-access-h9fb5\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399003 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/24fc18fa-9739-480c-8f43-ab30989164c2-multus-daemon-config\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399020 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-run-netns\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399080 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jns97\" (UniqueName: \"kubernetes.io/projected/24fc18fa-9739-480c-8f43-ab30989164c2-kube-api-access-jns97\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399166 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/24fc18fa-9739-480c-8f43-ab30989164c2-cni-binary-copy\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399194 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvl6z\" (UniqueName: \"kubernetes.io/projected/ebd8eb79-1502-48f4-98e1-45c365ac5638-kube-api-access-xvl6z\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-system-cni-dir\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399278 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-run-multus-certs\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399388 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-system-cni-dir\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399447 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-cnibin\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399380 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399492 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-multus-socket-dir-parent\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399523 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-run-k8s-cni-cncf-io\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399540 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-var-lib-cni-multus\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399564 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-var-lib-kubelet\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399585 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-os-release\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399603 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-tuning-conf-dir\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399653 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/05ed703a-a310-418e-8c5d-aeddb32977cf-rootfs\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399724 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-os-release\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399763 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-hostroot\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399797 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ebd8eb79-1502-48f4-98e1-45c365ac5638-cni-binary-copy\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399820 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05ed703a-a310-418e-8c5d-aeddb32977cf-proxy-tls\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399847 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-multus-cni-dir\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399879 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-cnibin\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399900 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-var-lib-cni-bin\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399940 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ebd8eb79-1502-48f4-98e1-45c365ac5638-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399967 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-multus-conf-dir\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.399986 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-etc-kubernetes\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.416792 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.435295 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.452168 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.467410 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.486534 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501214 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/24fc18fa-9739-480c-8f43-ab30989164c2-multus-daemon-config\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501260 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-run-netns\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501283 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jns97\" (UniqueName: \"kubernetes.io/projected/24fc18fa-9739-480c-8f43-ab30989164c2-kube-api-access-jns97\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501305 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/24fc18fa-9739-480c-8f43-ab30989164c2-cni-binary-copy\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501325 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvl6z\" (UniqueName: \"kubernetes.io/projected/ebd8eb79-1502-48f4-98e1-45c365ac5638-kube-api-access-xvl6z\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501357 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-system-cni-dir\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501376 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-run-multus-certs\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501407 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-system-cni-dir\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501429 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-cnibin\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501450 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-tuning-conf-dir\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501469 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/05ed703a-a310-418e-8c5d-aeddb32977cf-rootfs\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501491 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-multus-socket-dir-parent\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501510 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-run-k8s-cni-cncf-io\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501529 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-var-lib-cni-multus\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501549 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-var-lib-kubelet\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501572 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-os-release\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-os-release\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501629 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-hostroot\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501676 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ebd8eb79-1502-48f4-98e1-45c365ac5638-cni-binary-copy\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501696 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05ed703a-a310-418e-8c5d-aeddb32977cf-proxy-tls\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501731 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-multus-cni-dir\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-cnibin\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501771 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-var-lib-cni-bin\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501800 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ebd8eb79-1502-48f4-98e1-45c365ac5638-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501822 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-multus-conf-dir\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501840 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-etc-kubernetes\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501875 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05ed703a-a310-418e-8c5d-aeddb32977cf-mcd-auth-proxy-config\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.501903 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9fb5\" (UniqueName: \"kubernetes.io/projected/05ed703a-a310-418e-8c5d-aeddb32977cf-kube-api-access-h9fb5\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.503274 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/24fc18fa-9739-480c-8f43-ab30989164c2-multus-daemon-config\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.503334 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-run-netns\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.503892 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/24fc18fa-9739-480c-8f43-ab30989164c2-cni-binary-copy\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504075 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-system-cni-dir\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504114 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-run-multus-certs\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504143 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-system-cni-dir\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504173 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-cnibin\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-var-lib-cni-bin\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504439 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-var-lib-cni-multus\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504513 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-run-k8s-cni-cncf-io\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504539 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-multus-conf-dir\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504606 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-cnibin\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504653 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/05ed703a-a310-418e-8c5d-aeddb32977cf-rootfs\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504664 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-multus-socket-dir-parent\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504690 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-etc-kubernetes\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504727 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-host-var-lib-kubelet\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504733 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-hostroot\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504838 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-multus-cni-dir\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504842 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-tuning-conf-dir\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504958 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/24fc18fa-9739-480c-8f43-ab30989164c2-os-release\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.504967 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ebd8eb79-1502-48f4-98e1-45c365ac5638-os-release\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.505424 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ebd8eb79-1502-48f4-98e1-45c365ac5638-cni-binary-copy\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.505588 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05ed703a-a310-418e-8c5d-aeddb32977cf-mcd-auth-proxy-config\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.508536 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ebd8eb79-1502-48f4-98e1-45c365ac5638-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.508601 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.510219 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05ed703a-a310-418e-8c5d-aeddb32977cf-proxy-tls\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.526404 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9fb5\" (UniqueName: \"kubernetes.io/projected/05ed703a-a310-418e-8c5d-aeddb32977cf-kube-api-access-h9fb5\") pod \"machine-config-daemon-dhzp7\" (UID: \"05ed703a-a310-418e-8c5d-aeddb32977cf\") " pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.528251 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvl6z\" (UniqueName: \"kubernetes.io/projected/ebd8eb79-1502-48f4-98e1-45c365ac5638-kube-api-access-xvl6z\") pod \"multus-additional-cni-plugins-6jrkw\" (UID: \"ebd8eb79-1502-48f4-98e1-45c365ac5638\") " pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.528317 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jns97\" (UniqueName: \"kubernetes.io/projected/24fc18fa-9739-480c-8f43-ab30989164c2-kube-api-access-jns97\") pod \"multus-7jshv\" (UID: \"24fc18fa-9739-480c-8f43-ab30989164c2\") " pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.529775 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.541240 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.552702 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.564663 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.575716 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.594174 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.598290 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-44zpn" event={"ID":"40f86b56-873f-4c8a-81e2-abc5643f4869","Type":"ContainerStarted","Data":"b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019"} Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.598346 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-44zpn" event={"ID":"40f86b56-873f-4c8a-81e2-abc5643f4869","Type":"ContainerStarted","Data":"513a1d168f8d57d7b12eefe63b40fbf8e92ba49372210629362bbb69fb8f94d6"} Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.608177 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.627394 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.647071 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-7jshv" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.647141 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.658797 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.668461 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.688440 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: W0202 12:09:54.694053 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebd8eb79_1502_48f4_98e1_45c365ac5638.slice/crio-9e357b6d10d54ec96832696710a157c0f58c0357e636ee9a6e84a8597217a45f WatchSource:0}: Error finding container 9e357b6d10d54ec96832696710a157c0f58c0357e636ee9a6e84a8597217a45f: Status 404 returned error can't find the container with id 9e357b6d10d54ec96832696710a157c0f58c0357e636ee9a6e84a8597217a45f Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.719541 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.738048 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ftvcw"] Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.739026 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.742913 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-02 12:04:53 +0000 UTC, rotation deadline is 2026-11-19 04:47:49.921757702 +0000 UTC Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.742964 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6952h37m55.178796369s for next certificate rotation Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.744456 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.744671 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.744699 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.744851 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.745941 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.746158 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.746692 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.765003 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.800400 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804181 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-systemd\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804216 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-slash\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804232 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-netns\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804248 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-env-overrides\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804263 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-var-lib-openvswitch\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804277 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-ovn\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804299 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-ovn-kubernetes\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804314 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovn-node-metrics-cert\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804399 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-bin\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804453 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804481 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-script-lib\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804510 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-node-log\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804555 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-etc-openvswitch\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804605 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-systemd-units\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804662 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87zlr\" (UniqueName: \"kubernetes.io/projected/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-kube-api-access-87zlr\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804693 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-openvswitch\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804714 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-config\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804743 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-log-socket\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804765 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-kubelet\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.804794 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-netd\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.817776 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.819306 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.819759 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.819789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.819803 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.819908 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.839442 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.839734 4846 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.839988 4846 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.842281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.842325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.842336 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.842350 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.842359 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:54Z","lastTransitionTime":"2026-02-02T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.857468 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: E0202 12:09:54.860834 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.872494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.872536 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.872545 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.872559 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.872571 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:54Z","lastTransitionTime":"2026-02-02T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.878085 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: E0202 12:09:54.894807 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.899450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.899481 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.899492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.899507 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.899515 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:54Z","lastTransitionTime":"2026-02-02T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905243 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87zlr\" (UniqueName: \"kubernetes.io/projected/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-kube-api-access-87zlr\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905278 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-config\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905301 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-openvswitch\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905318 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-log-socket\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905333 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-kubelet\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905348 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-netd\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905370 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-slash\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905385 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-netns\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905398 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-systemd\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905413 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-env-overrides\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905428 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-var-lib-openvswitch\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905441 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-ovn\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905458 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-ovn-kubernetes\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905474 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovn-node-metrics-cert\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905493 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-bin\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905511 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905521 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-log-socket\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905531 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-slash\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905596 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-netns\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905601 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-openvswitch\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905609 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-ovn\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905551 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-script-lib\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905688 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-ovn-kubernetes\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905742 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-systemd\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905751 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-node-log\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905774 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-node-log\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905797 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-kubelet\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905799 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-bin\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.905832 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-netd\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.906203 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-config\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.906249 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-var-lib-openvswitch\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.906306 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-etc-openvswitch\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.906327 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-env-overrides\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.906363 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-etc-openvswitch\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.906447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-systemd-units\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.906515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-systemd-units\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.906767 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.906846 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-script-lib\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.910981 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovn-node-metrics-cert\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: E0202 12:09:54.912151 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.919788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.919827 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.919844 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.919862 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.919875 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:54Z","lastTransitionTime":"2026-02-02T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.930399 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87zlr\" (UniqueName: \"kubernetes.io/projected/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-kube-api-access-87zlr\") pod \"ovnkube-node-ftvcw\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.931738 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: E0202 12:09:54.935968 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.940745 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.940771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.940779 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.940794 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.940802 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:54Z","lastTransitionTime":"2026-02-02T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.946306 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: E0202 12:09:54.954867 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: E0202 12:09:54.955036 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.956914 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.956943 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.956954 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.956970 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.956981 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:54Z","lastTransitionTime":"2026-02-02T12:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.961368 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.971211 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:54 crc kubenswrapper[4846]: I0202 12:09:54.981485 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.000271 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:54Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.011255 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.033934 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.046293 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.049063 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.058078 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.058837 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.058858 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.058866 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.058880 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.058890 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:55Z","lastTransitionTime":"2026-02-02T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:55 crc kubenswrapper[4846]: W0202 12:09:55.062971 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ea83b34_bfee_43e5_8b72_5a59fcf0fc8e.slice/crio-f2ae9005ccde81fc6c8ba26dfc3337492496c8bae1840dfbdd1186fbde5cec48 WatchSource:0}: Error finding container f2ae9005ccde81fc6c8ba26dfc3337492496c8bae1840dfbdd1186fbde5cec48: Status 404 returned error can't find the container with id f2ae9005ccde81fc6c8ba26dfc3337492496c8bae1840dfbdd1186fbde5cec48 Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.072384 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.087952 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.099770 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.117111 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.130917 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.143918 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.154038 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.161186 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.161230 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.161241 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.161257 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.161268 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:55Z","lastTransitionTime":"2026-02-02T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.166809 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.177579 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.191522 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.263678 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.263710 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.263718 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.263732 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.263741 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:55Z","lastTransitionTime":"2026-02-02T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.366191 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.366238 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.366248 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.366262 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.366271 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:55Z","lastTransitionTime":"2026-02-02T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.382644 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 22:01:57.004992678 +0000 UTC Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.423449 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:55 crc kubenswrapper[4846]: E0202 12:09:55.423564 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.423900 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:55 crc kubenswrapper[4846]: E0202 12:09:55.423950 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.424064 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:55 crc kubenswrapper[4846]: E0202 12:09:55.424122 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.469116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.469152 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.469162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.469178 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.469188 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:55Z","lastTransitionTime":"2026-02-02T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.571768 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.571808 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.571817 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.571832 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.571842 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:55Z","lastTransitionTime":"2026-02-02T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.602968 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c" exitCode=0 Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.603039 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.603088 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"f2ae9005ccde81fc6c8ba26dfc3337492496c8bae1840dfbdd1186fbde5cec48"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.604754 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.605200 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.605240 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"97188d1fde1ef003ecd610dafce30ba4f2e613ea183c19d12c6cc4cb4a3219f7"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.605969 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7jshv" event={"ID":"24fc18fa-9739-480c-8f43-ab30989164c2","Type":"ContainerStarted","Data":"de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.606003 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7jshv" event={"ID":"24fc18fa-9739-480c-8f43-ab30989164c2","Type":"ContainerStarted","Data":"d3d817fc04d9039a49cd46e8efa99699d97f312cbd5e51ce9a22fafcd444fae4"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.607831 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebd8eb79-1502-48f4-98e1-45c365ac5638" containerID="cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c" exitCode=0 Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.607859 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" event={"ID":"ebd8eb79-1502-48f4-98e1-45c365ac5638","Type":"ContainerDied","Data":"cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.607873 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" event={"ID":"ebd8eb79-1502-48f4-98e1-45c365ac5638","Type":"ContainerStarted","Data":"9e357b6d10d54ec96832696710a157c0f58c0357e636ee9a6e84a8597217a45f"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.619984 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.639814 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.658572 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.671916 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.674454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.674478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.674487 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.674499 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.674510 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:55Z","lastTransitionTime":"2026-02-02T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.686725 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.705363 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.720458 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.737267 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.749181 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.760773 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.774084 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.776598 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.776624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.776661 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.776675 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.776683 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:55Z","lastTransitionTime":"2026-02-02T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.788655 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.801443 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.821974 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.834494 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.853920 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.868002 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.878453 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.878496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.878507 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.878526 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.878537 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:55Z","lastTransitionTime":"2026-02-02T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.880653 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.902848 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.921495 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.936474 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.948651 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.962707 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.981391 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.983098 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.983130 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.983141 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.983157 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.983168 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:55Z","lastTransitionTime":"2026-02-02T12:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:55 crc kubenswrapper[4846]: I0202 12:09:55.995180 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:55Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.008199 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.028826 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.042167 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.085423 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.085478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.085494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.085515 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.085527 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:56Z","lastTransitionTime":"2026-02-02T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.187521 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.187556 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.187565 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.187579 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.187590 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:56Z","lastTransitionTime":"2026-02-02T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.220260 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.221348 4846 scope.go:117] "RemoveContainer" containerID="ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7" Feb 02 12:09:56 crc kubenswrapper[4846]: E0202 12:09:56.221658 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.242875 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-9vvp2"] Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.243407 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.245229 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.245265 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.245230 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.245870 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.266671 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.286394 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.290211 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.290246 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.290257 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.290273 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.290285 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:56Z","lastTransitionTime":"2026-02-02T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.318031 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.322448 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt27x\" (UniqueName: \"kubernetes.io/projected/d5aa2f93-ed54-4835-ad19-8b63707975d9-kube-api-access-xt27x\") pod \"node-ca-9vvp2\" (UID: \"d5aa2f93-ed54-4835-ad19-8b63707975d9\") " pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.322480 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d5aa2f93-ed54-4835-ad19-8b63707975d9-host\") pod \"node-ca-9vvp2\" (UID: \"d5aa2f93-ed54-4835-ad19-8b63707975d9\") " pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.322504 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d5aa2f93-ed54-4835-ad19-8b63707975d9-serviceca\") pod \"node-ca-9vvp2\" (UID: \"d5aa2f93-ed54-4835-ad19-8b63707975d9\") " pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.331278 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.342003 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.353609 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.363358 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.376080 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.382877 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 01:07:58.826339992 +0000 UTC Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.392493 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.392541 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.392554 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.392572 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.392583 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:56Z","lastTransitionTime":"2026-02-02T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.414114 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.423809 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt27x\" (UniqueName: \"kubernetes.io/projected/d5aa2f93-ed54-4835-ad19-8b63707975d9-kube-api-access-xt27x\") pod \"node-ca-9vvp2\" (UID: \"d5aa2f93-ed54-4835-ad19-8b63707975d9\") " pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.423846 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d5aa2f93-ed54-4835-ad19-8b63707975d9-host\") pod \"node-ca-9vvp2\" (UID: \"d5aa2f93-ed54-4835-ad19-8b63707975d9\") " pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.423879 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d5aa2f93-ed54-4835-ad19-8b63707975d9-serviceca\") pod \"node-ca-9vvp2\" (UID: \"d5aa2f93-ed54-4835-ad19-8b63707975d9\") " pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.424092 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d5aa2f93-ed54-4835-ad19-8b63707975d9-host\") pod \"node-ca-9vvp2\" (UID: \"d5aa2f93-ed54-4835-ad19-8b63707975d9\") " pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.425085 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d5aa2f93-ed54-4835-ad19-8b63707975d9-serviceca\") pod \"node-ca-9vvp2\" (UID: \"d5aa2f93-ed54-4835-ad19-8b63707975d9\") " pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.463595 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt27x\" (UniqueName: \"kubernetes.io/projected/d5aa2f93-ed54-4835-ad19-8b63707975d9-kube-api-access-xt27x\") pod \"node-ca-9vvp2\" (UID: \"d5aa2f93-ed54-4835-ad19-8b63707975d9\") " pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.473565 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.495890 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.496035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.496046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.496068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.496079 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:56Z","lastTransitionTime":"2026-02-02T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.512743 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.554628 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.597806 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.597966 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.597987 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.597995 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.598009 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.598018 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:56Z","lastTransitionTime":"2026-02-02T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.614280 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.614321 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.614331 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.614340 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.614352 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.615998 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebd8eb79-1502-48f4-98e1-45c365ac5638" containerID="5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3" exitCode=0 Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.616025 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" event={"ID":"ebd8eb79-1502-48f4-98e1-45c365ac5638","Type":"ContainerDied","Data":"5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.637813 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.661492 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-9vvp2" Feb 02 12:09:56 crc kubenswrapper[4846]: W0202 12:09:56.680084 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5aa2f93_ed54_4835_ad19_8b63707975d9.slice/crio-8673c91b86750468882a8bc3017e4d2e5c8e2c3e31dd47d1717af9fdba34454d WatchSource:0}: Error finding container 8673c91b86750468882a8bc3017e4d2e5c8e2c3e31dd47d1717af9fdba34454d: Status 404 returned error can't find the container with id 8673c91b86750468882a8bc3017e4d2e5c8e2c3e31dd47d1717af9fdba34454d Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.680781 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.700021 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.700047 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.700054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.700068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.700078 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:56Z","lastTransitionTime":"2026-02-02T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.711613 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.751088 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.800466 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.802398 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.802425 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.802433 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.802446 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.802459 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:56Z","lastTransitionTime":"2026-02-02T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.834035 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.872131 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.905782 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.905811 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.905819 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.905832 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.905841 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:56Z","lastTransitionTime":"2026-02-02T12:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.915487 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.954078 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:56 crc kubenswrapper[4846]: I0202 12:09:56.993373 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.007730 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.007756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.007764 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.007778 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.007791 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:57Z","lastTransitionTime":"2026-02-02T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.043303 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.073332 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.109737 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.109782 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.109795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.109815 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.109828 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:57Z","lastTransitionTime":"2026-02-02T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.115847 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.129188 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.129273 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.129306 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.129326 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.129390 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.129436 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:05.129424545 +0000 UTC m=+36.358011408 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.129875 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:10:05.129833365 +0000 UTC m=+36.358420238 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.129939 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.130010 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:05.129985999 +0000 UTC m=+36.358572862 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.130256 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.130309 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.130330 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.130406 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:05.130381887 +0000 UTC m=+36.358968750 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.151529 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.193749 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.212097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.212133 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.212141 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.212158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.212170 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:57Z","lastTransitionTime":"2026-02-02T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.230118 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.230288 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.230315 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.230327 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.230386 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:05.230369048 +0000 UTC m=+36.458955911 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.231692 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.277025 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.314586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.314624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.314655 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.314673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.314684 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:57Z","lastTransitionTime":"2026-02-02T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.383585 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 07:25:34.277587178 +0000 UTC Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.417669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.417706 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.417715 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.417729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.417739 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:57Z","lastTransitionTime":"2026-02-02T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.423323 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.423364 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.423438 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.423522 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.423654 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:09:57 crc kubenswrapper[4846]: E0202 12:09:57.423782 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.520401 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.520453 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.520474 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.520506 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.520529 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:57Z","lastTransitionTime":"2026-02-02T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.621007 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-9vvp2" event={"ID":"d5aa2f93-ed54-4835-ad19-8b63707975d9","Type":"ContainerStarted","Data":"f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.621085 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-9vvp2" event={"ID":"d5aa2f93-ed54-4835-ad19-8b63707975d9","Type":"ContainerStarted","Data":"8673c91b86750468882a8bc3017e4d2e5c8e2c3e31dd47d1717af9fdba34454d"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.622026 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.622056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.622065 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.622081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.622091 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:57Z","lastTransitionTime":"2026-02-02T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.625071 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.626894 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebd8eb79-1502-48f4-98e1-45c365ac5638" containerID="a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021" exitCode=0 Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.626929 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" event={"ID":"ebd8eb79-1502-48f4-98e1-45c365ac5638","Type":"ContainerDied","Data":"a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.635686 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.654592 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.668273 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.680079 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.690169 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.704157 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.715847 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.726387 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.726427 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.726436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.726425 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.726452 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.726568 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:57Z","lastTransitionTime":"2026-02-02T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.745335 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.757399 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.768762 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.787120 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.800485 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.829101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.829145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.829153 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.829168 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.829180 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:57Z","lastTransitionTime":"2026-02-02T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.833672 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.872786 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.919199 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.932023 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.932065 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.932078 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.932097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.932110 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:57Z","lastTransitionTime":"2026-02-02T12:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.954346 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:57 crc kubenswrapper[4846]: I0202 12:09:57.994880 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:57Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.034282 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.035443 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.035476 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.035484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.035499 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.035509 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:58Z","lastTransitionTime":"2026-02-02T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.073919 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.113900 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.137456 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.137491 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.137500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.137515 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.137523 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:58Z","lastTransitionTime":"2026-02-02T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.158729 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.195111 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.233705 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.240088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.240117 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.240127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.240140 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.240149 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:58Z","lastTransitionTime":"2026-02-02T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.270269 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.314193 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.342662 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.342700 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.342714 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.342732 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.342744 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:58Z","lastTransitionTime":"2026-02-02T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.351238 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.383891 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 15:28:56.303031905 +0000 UTC Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.396925 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.434618 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.445718 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.445758 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.445766 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.445780 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.445790 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:58Z","lastTransitionTime":"2026-02-02T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.470165 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.547877 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.547917 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.547926 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.547940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.547953 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:58Z","lastTransitionTime":"2026-02-02T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.632685 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebd8eb79-1502-48f4-98e1-45c365ac5638" containerID="f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8" exitCode=0 Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.632726 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" event={"ID":"ebd8eb79-1502-48f4-98e1-45c365ac5638","Type":"ContainerDied","Data":"f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.649760 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.649808 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.649823 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.649843 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.649858 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:58Z","lastTransitionTime":"2026-02-02T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.654022 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.671561 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.685192 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.726429 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.755516 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.756585 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.756613 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.756633 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.756647 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.756656 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:58Z","lastTransitionTime":"2026-02-02T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.774750 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.784971 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.822156 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.833734 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.858721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.858756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.858766 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.858781 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.858793 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:58Z","lastTransitionTime":"2026-02-02T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.876179 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.916933 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.952202 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.961412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.961468 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.961488 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.961553 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.961574 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:58Z","lastTransitionTime":"2026-02-02T12:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:58 crc kubenswrapper[4846]: I0202 12:09:58.997900 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:58Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.032753 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.064132 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.064171 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.064182 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.064197 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.064209 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:59Z","lastTransitionTime":"2026-02-02T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.072465 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.166871 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.166915 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.166932 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.166953 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.166969 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:59Z","lastTransitionTime":"2026-02-02T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.217564 4846 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.269592 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.269678 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.269693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.269721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.269738 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:59Z","lastTransitionTime":"2026-02-02T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.372945 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.372978 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.372986 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.372999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.373007 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:59Z","lastTransitionTime":"2026-02-02T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.384476 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 07:02:50.296554615 +0000 UTC Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.423008 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:09:59 crc kubenswrapper[4846]: E0202 12:09:59.423131 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.423422 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:09:59 crc kubenswrapper[4846]: E0202 12:09:59.423485 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.423527 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:09:59 crc kubenswrapper[4846]: E0202 12:09:59.423567 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.439832 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.453276 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.463861 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.474891 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.474958 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.474968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.474984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.474995 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:59Z","lastTransitionTime":"2026-02-02T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.479557 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.492795 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.510018 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.520586 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.529095 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.548866 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.560794 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.572120 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.577299 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.577340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.577396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.577442 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.577455 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:59Z","lastTransitionTime":"2026-02-02T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.582434 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.592305 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.633117 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.639515 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.642429 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebd8eb79-1502-48f4-98e1-45c365ac5638" containerID="41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701" exitCode=0 Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.642465 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" event={"ID":"ebd8eb79-1502-48f4-98e1-45c365ac5638","Type":"ContainerDied","Data":"41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.678282 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.679553 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.679602 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.679612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.679650 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.679659 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:59Z","lastTransitionTime":"2026-02-02T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.712908 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.750660 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.781780 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.781804 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.781814 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.781828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.781859 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:59Z","lastTransitionTime":"2026-02-02T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.792552 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.832238 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.880069 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.886030 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.886059 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.886067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.886081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.886090 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:59Z","lastTransitionTime":"2026-02-02T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.912836 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.951826 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.988515 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.988591 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.988616 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.988691 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.988716 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:09:59Z","lastTransitionTime":"2026-02-02T12:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:09:59 crc kubenswrapper[4846]: I0202 12:09:59.998295 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:09:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.031919 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.070310 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.092262 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.092302 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.092317 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.092337 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.092352 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:00Z","lastTransitionTime":"2026-02-02T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.115366 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.154369 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.195857 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.195908 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.195923 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.195944 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.195961 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:00Z","lastTransitionTime":"2026-02-02T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.198965 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.239363 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.277010 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.298581 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.298671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.298681 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.298694 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.298968 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:00Z","lastTransitionTime":"2026-02-02T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.385229 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 02:01:02.47451516 +0000 UTC Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.403050 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.403095 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.403108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.403125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.403137 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:00Z","lastTransitionTime":"2026-02-02T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.509046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.509096 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.509108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.509126 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.509137 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:00Z","lastTransitionTime":"2026-02-02T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.612244 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.612559 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.612572 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.612596 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.612625 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:00Z","lastTransitionTime":"2026-02-02T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.650128 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebd8eb79-1502-48f4-98e1-45c365ac5638" containerID="aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87" exitCode=0 Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.650181 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" event={"ID":"ebd8eb79-1502-48f4-98e1-45c365ac5638","Type":"ContainerDied","Data":"aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87"} Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.664048 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.678873 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.699440 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.714664 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.715177 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.715246 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.715258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.715274 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.715287 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:00Z","lastTransitionTime":"2026-02-02T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.726002 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.739335 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.752102 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.763585 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.782049 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.797565 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.810980 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.817099 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.817129 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.817138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.817154 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.817165 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:00Z","lastTransitionTime":"2026-02-02T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.823416 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.836543 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.848553 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.874950 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:00Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.919794 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.919825 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.919834 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.919850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:00 crc kubenswrapper[4846]: I0202 12:10:00.919859 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:00Z","lastTransitionTime":"2026-02-02T12:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.021969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.022018 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.022034 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.022054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.022069 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:01Z","lastTransitionTime":"2026-02-02T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.123621 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.123667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.123680 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.123696 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.123706 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:01Z","lastTransitionTime":"2026-02-02T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.225603 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.225685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.225702 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.226127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.226184 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:01Z","lastTransitionTime":"2026-02-02T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.331119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.331159 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.331179 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.331203 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.331219 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:01Z","lastTransitionTime":"2026-02-02T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.386794 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 00:43:56.56249269 +0000 UTC Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.423416 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.423437 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:01 crc kubenswrapper[4846]: E0202 12:10:01.423596 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.423693 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:01 crc kubenswrapper[4846]: E0202 12:10:01.423786 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:01 crc kubenswrapper[4846]: E0202 12:10:01.423883 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.435666 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.435700 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.435709 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.435721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.435730 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:01Z","lastTransitionTime":"2026-02-02T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.538253 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.538294 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.538305 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.538323 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.538334 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:01Z","lastTransitionTime":"2026-02-02T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.640835 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.641231 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.641252 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.641283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.641304 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:01Z","lastTransitionTime":"2026-02-02T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.656693 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.656939 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.661610 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" event={"ID":"ebd8eb79-1502-48f4-98e1-45c365ac5638","Type":"ContainerStarted","Data":"a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.676678 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.681169 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.688838 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.701659 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.713825 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.727330 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.739432 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.743218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.743261 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.743278 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.743298 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.743312 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:01Z","lastTransitionTime":"2026-02-02T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.761201 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.778501 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.793362 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.804569 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.819065 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.831649 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.846239 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.846297 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.846314 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.846334 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.846346 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:01Z","lastTransitionTime":"2026-02-02T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.849870 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.862251 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.870499 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.879763 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.887806 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.904648 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.914374 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.923389 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.939435 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.949009 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.949038 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.949049 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.949063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.949072 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:01Z","lastTransitionTime":"2026-02-02T12:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.951076 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.965272 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.977592 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:01 crc kubenswrapper[4846]: I0202 12:10:01.988735 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:01Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.003197 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.015076 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.026210 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.034512 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.051293 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.051327 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.051338 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.051355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.051366 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:02Z","lastTransitionTime":"2026-02-02T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.079593 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.153586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.153639 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.153650 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.153665 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.153676 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:02Z","lastTransitionTime":"2026-02-02T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.256382 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.256419 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.256429 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.256445 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.256456 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:02Z","lastTransitionTime":"2026-02-02T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.359280 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.359340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.359355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.359380 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.359395 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:02Z","lastTransitionTime":"2026-02-02T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.387232 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 16:49:57.368485647 +0000 UTC Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.462684 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.462737 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.462755 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.462780 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.462801 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:02Z","lastTransitionTime":"2026-02-02T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.565234 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.565281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.565290 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.565304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.565314 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:02Z","lastTransitionTime":"2026-02-02T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.663883 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.664352 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.667306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.667334 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.667342 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.667355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.667365 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:02Z","lastTransitionTime":"2026-02-02T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.687073 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.698407 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.712919 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.723394 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.737088 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.749909 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.761152 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.769928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.769974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.769986 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.770004 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.770015 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:02Z","lastTransitionTime":"2026-02-02T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.773581 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.782924 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.794008 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.804933 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.823457 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.834867 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.846113 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.862579 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.873301 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.873335 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.873344 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.873360 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.873369 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:02Z","lastTransitionTime":"2026-02-02T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.877001 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:02Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.975362 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.975410 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.975418 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.975434 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:02 crc kubenswrapper[4846]: I0202 12:10:02.975444 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:02Z","lastTransitionTime":"2026-02-02T12:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.077771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.077805 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.077813 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.077827 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.077838 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:03Z","lastTransitionTime":"2026-02-02T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.180783 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.180832 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.180981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.181003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.181013 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:03Z","lastTransitionTime":"2026-02-02T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.283143 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.283183 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.283195 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.283212 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.283224 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:03Z","lastTransitionTime":"2026-02-02T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.387886 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 10:02:19.022881603 +0000 UTC Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.412359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.412400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.412411 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.412427 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.412439 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:03Z","lastTransitionTime":"2026-02-02T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.425465 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:03 crc kubenswrapper[4846]: E0202 12:10:03.425559 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.425927 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:03 crc kubenswrapper[4846]: E0202 12:10:03.426000 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.426051 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:03 crc kubenswrapper[4846]: E0202 12:10:03.426120 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.514937 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.514969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.514980 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.514996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.515006 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:03Z","lastTransitionTime":"2026-02-02T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.616996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.617026 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.617034 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.617047 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.617056 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:03Z","lastTransitionTime":"2026-02-02T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.672180 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.719546 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.719637 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.719648 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.719667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.719679 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:03Z","lastTransitionTime":"2026-02-02T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.822649 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.822692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.822703 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.822718 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.822727 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:03Z","lastTransitionTime":"2026-02-02T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.925223 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.925256 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.925264 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.925279 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:03 crc kubenswrapper[4846]: I0202 12:10:03.925287 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:03Z","lastTransitionTime":"2026-02-02T12:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.028838 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.028923 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.028944 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.028967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.028983 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:04Z","lastTransitionTime":"2026-02-02T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.131969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.132010 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.132025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.132044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.132055 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:04Z","lastTransitionTime":"2026-02-02T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.234814 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.234872 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.234887 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.234906 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.234917 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:04Z","lastTransitionTime":"2026-02-02T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.337717 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.337765 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.337776 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.337792 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.337803 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:04Z","lastTransitionTime":"2026-02-02T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.388797 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 04:54:19.231317436 +0000 UTC Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.465483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.465509 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.465517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.465555 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.465566 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:04Z","lastTransitionTime":"2026-02-02T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.567886 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.567940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.567955 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.567975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.567989 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:04Z","lastTransitionTime":"2026-02-02T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.670773 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.670851 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.670877 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.670907 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.670930 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:04Z","lastTransitionTime":"2026-02-02T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.678326 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/0.log" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.681959 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef" exitCode=1 Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.682024 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.683410 4846 scope.go:117] "RemoveContainer" containerID="03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.705296 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.741287 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.755464 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.773228 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.773267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.773275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.773290 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.773299 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:04Z","lastTransitionTime":"2026-02-02T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.774354 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:03Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 12:10:03.865926 6139 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:03.865975 6139 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0202 12:10:03.865990 6139 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0202 12:10:03.865995 6139 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0202 12:10:03.866005 6139 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:03.866017 6139 factory.go:656] Stopping watch factory\\\\nI0202 12:10:03.866033 6139 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:03.866040 6139 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0202 12:10:03.866046 6139 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:03.866051 6139 handler.go:208] Removed *v1.Node event handler 2\\\\nI0202 12:10:03.866079 6139 handler.go:208] Removed *v1.Node event handler 7\\\\nI0202 12:10:03.866164 6139 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.790121 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.804198 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.815739 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.830091 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.844248 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.860380 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.874014 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.878354 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.878409 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.878440 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.878468 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.878487 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:04Z","lastTransitionTime":"2026-02-02T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.884733 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.903179 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.914164 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.925542 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:04Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.981279 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.981347 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.981365 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.981392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:04 crc kubenswrapper[4846]: I0202 12:10:04.981408 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:04Z","lastTransitionTime":"2026-02-02T12:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.083922 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.084152 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.084160 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.084174 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.084183 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.093243 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.093286 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.093298 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.093316 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.093329 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.115495 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.119191 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.119239 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.119254 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.119277 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.119290 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.138758 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.141712 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.141753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.141762 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.141781 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.141789 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.153877 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.157438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.157503 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.157514 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.157532 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.157542 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.172761 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.172851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.172886 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.172946 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:10:21.172920078 +0000 UTC m=+52.401506971 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.172980 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.172994 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.173004 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.172978 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.173044 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:21.17303105 +0000 UTC m=+52.401617913 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.173103 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:21.173079901 +0000 UTC m=+52.401666854 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.173113 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.173166 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:21.173157413 +0000 UTC m=+52.401744356 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.173028 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.174963 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.179521 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.179554 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.179564 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.179578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.179587 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.192532 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.192676 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.194074 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.194101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.194112 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.194125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.194135 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.274169 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.274308 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.274325 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.274338 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.274394 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:21.274379053 +0000 UTC m=+52.502965916 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.296636 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.296671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.296682 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.296697 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.296708 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.389208 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 13:32:07.481678078 +0000 UTC Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.398967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.398999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.399007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.399021 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.399045 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.423399 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.423471 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.423512 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.423529 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.423664 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.423995 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.501614 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.501691 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.501706 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.501749 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.501760 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.604007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.604035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.604042 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.604056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.604066 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.685437 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/1.log" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.685882 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/0.log" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.687986 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046" exitCode=1 Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.688018 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.688048 4846 scope.go:117] "RemoveContainer" containerID="03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.688677 4846 scope.go:117] "RemoveContainer" containerID="67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046" Feb 02 12:10:05 crc kubenswrapper[4846]: E0202 12:10:05.688824 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.706957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.706988 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.706997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.707013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.707024 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.709610 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.718052 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.729766 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.741999 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.754214 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.779616 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.792213 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.808599 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.809549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.809595 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.809612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.809676 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.809698 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.823194 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.837006 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.863098 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.877915 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.890847 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.908571 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03294af6598d1d556027ddf6fbac977544a626bd45af8c4c7fdf0d3b050db7ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:03Z\\\",\\\"message\\\":\\\"] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 12:10:03.865926 6139 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:03.865975 6139 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0202 12:10:03.865990 6139 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0202 12:10:03.865995 6139 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0202 12:10:03.866005 6139 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:03.866017 6139 factory.go:656] Stopping watch factory\\\\nI0202 12:10:03.866033 6139 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:03.866040 6139 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0202 12:10:03.866046 6139 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:03.866051 6139 handler.go:208] Removed *v1.Node event handler 2\\\\nI0202 12:10:03.866079 6139 handler.go:208] Removed *v1.Node event handler 7\\\\nI0202 12:10:03.866164 6139 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.911822 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.911850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.911857 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.911887 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.911896 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:05Z","lastTransitionTime":"2026-02-02T12:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:05 crc kubenswrapper[4846]: I0202 12:10:05.920434 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:05Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.013871 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.013926 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.013936 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.013951 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.013964 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:06Z","lastTransitionTime":"2026-02-02T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.117056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.117116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.117133 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.117157 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.117173 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:06Z","lastTransitionTime":"2026-02-02T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.220912 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.220964 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.220977 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.220993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.221006 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:06Z","lastTransitionTime":"2026-02-02T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.323569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.323613 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.323646 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.323663 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.323673 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:06Z","lastTransitionTime":"2026-02-02T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.390204 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 07:42:43.990054654 +0000 UTC Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.426177 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.426237 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.426255 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.426278 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.426294 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:06Z","lastTransitionTime":"2026-02-02T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.529068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.529140 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.529157 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.529182 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.529199 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:06Z","lastTransitionTime":"2026-02-02T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.631814 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.631870 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.631889 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.631910 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.631924 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:06Z","lastTransitionTime":"2026-02-02T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.692428 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/1.log" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.695876 4846 scope.go:117] "RemoveContainer" containerID="67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046" Feb 02 12:10:06 crc kubenswrapper[4846]: E0202 12:10:06.696057 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.715570 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.730448 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.734353 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.734386 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.734431 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.734450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.734462 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:06Z","lastTransitionTime":"2026-02-02T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.745577 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.759843 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.774659 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.786693 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.800407 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.811201 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.837205 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.837269 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.837288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.837312 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.837330 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:06Z","lastTransitionTime":"2026-02-02T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.837254 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.852611 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.866780 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.879782 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.891644 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.903267 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.919587 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:06Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.939612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.939649 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.939836 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.939854 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:06 crc kubenswrapper[4846]: I0202 12:10:06.939864 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:06Z","lastTransitionTime":"2026-02-02T12:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.042373 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.042419 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.042431 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.042446 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.042457 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:07Z","lastTransitionTime":"2026-02-02T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.146106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.146183 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.146215 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.146245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.146269 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:07Z","lastTransitionTime":"2026-02-02T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.249174 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.249217 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.249225 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.249241 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.249251 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:07Z","lastTransitionTime":"2026-02-02T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.330546 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz"] Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.331301 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.334418 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.334561 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.351052 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.351081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.351089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.351102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.351111 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:07Z","lastTransitionTime":"2026-02-02T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.356467 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.367879 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.381115 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.390346 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 03:50:19.6966009 +0000 UTC Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.393165 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.394863 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhr2f\" (UniqueName: \"kubernetes.io/projected/e681c0d1-0580-45d2-a381-540fded41685-kube-api-access-dhr2f\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.394892 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e681c0d1-0580-45d2-a381-540fded41685-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.394914 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e681c0d1-0580-45d2-a381-540fded41685-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.394934 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e681c0d1-0580-45d2-a381-540fded41685-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.403281 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.418410 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.423286 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:07 crc kubenswrapper[4846]: E0202 12:10:07.423432 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.423509 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:07 crc kubenswrapper[4846]: E0202 12:10:07.423572 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.424279 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:07 crc kubenswrapper[4846]: E0202 12:10:07.424365 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.424510 4846 scope.go:117] "RemoveContainer" containerID="ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.432326 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.445089 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.453909 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.453936 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.453944 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.453957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.453965 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:07Z","lastTransitionTime":"2026-02-02T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.457828 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.470885 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.481141 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.492318 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.495248 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e681c0d1-0580-45d2-a381-540fded41685-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.495284 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e681c0d1-0580-45d2-a381-540fded41685-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.495352 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhr2f\" (UniqueName: \"kubernetes.io/projected/e681c0d1-0580-45d2-a381-540fded41685-kube-api-access-dhr2f\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.495376 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e681c0d1-0580-45d2-a381-540fded41685-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.496065 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e681c0d1-0580-45d2-a381-540fded41685-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.496169 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e681c0d1-0580-45d2-a381-540fded41685-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.504114 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.504607 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e681c0d1-0580-45d2-a381-540fded41685-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.512344 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhr2f\" (UniqueName: \"kubernetes.io/projected/e681c0d1-0580-45d2-a381-540fded41685-kube-api-access-dhr2f\") pod \"ovnkube-control-plane-749d76644c-cscmz\" (UID: \"e681c0d1-0580-45d2-a381-540fded41685\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.522346 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.535107 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.546555 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.557040 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.557258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.557269 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.557288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.557300 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:07Z","lastTransitionTime":"2026-02-02T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.646712 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.660000 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.660029 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.660037 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.660050 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.660059 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:07Z","lastTransitionTime":"2026-02-02T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:07 crc kubenswrapper[4846]: W0202 12:10:07.661028 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode681c0d1_0580_45d2_a381_540fded41685.slice/crio-9e8eb6be2f5f1ee459148706bcb7539e51cd96708425a0382829cc48ae93a067 WatchSource:0}: Error finding container 9e8eb6be2f5f1ee459148706bcb7539e51cd96708425a0382829cc48ae93a067: Status 404 returned error can't find the container with id 9e8eb6be2f5f1ee459148706bcb7539e51cd96708425a0382829cc48ae93a067 Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.704850 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.706594 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.707029 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.707482 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" event={"ID":"e681c0d1-0580-45d2-a381-540fded41685","Type":"ContainerStarted","Data":"9e8eb6be2f5f1ee459148706bcb7539e51cd96708425a0382829cc48ae93a067"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.731909 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.744310 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.756880 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.762658 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.762700 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.762709 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.762724 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.762733 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:07Z","lastTransitionTime":"2026-02-02T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.770335 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.790961 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.803871 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.817165 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.836382 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.849462 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.863698 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.865061 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.865093 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.865106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.865122 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.865132 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:07Z","lastTransitionTime":"2026-02-02T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.878475 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.891093 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.903461 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.918703 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.929646 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.941945 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:07Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.967454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.967503 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.967512 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.967527 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:07 crc kubenswrapper[4846]: I0202 12:10:07.967538 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:07Z","lastTransitionTime":"2026-02-02T12:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.069708 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.069747 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.069757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.069775 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.069786 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:08Z","lastTransitionTime":"2026-02-02T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.172214 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.172266 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.172283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.172303 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.172316 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:08Z","lastTransitionTime":"2026-02-02T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.274160 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.274211 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.274219 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.274233 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.274242 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:08Z","lastTransitionTime":"2026-02-02T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.377261 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.377311 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.377322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.377339 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.377351 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:08Z","lastTransitionTime":"2026-02-02T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.391461 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 22:31:04.017550552 +0000 UTC Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.449073 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-j5m9k"] Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.449552 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:08 crc kubenswrapper[4846]: E0202 12:10:08.449656 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.463563 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.471990 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.480047 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.480094 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.480108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.480126 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.480148 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:08Z","lastTransitionTime":"2026-02-02T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.485037 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.494615 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.504231 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhqv6\" (UniqueName: \"kubernetes.io/projected/fe216e68-8814-468d-a6f3-fe8721d94eba-kube-api-access-zhqv6\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.504292 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.509535 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.521244 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.529843 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.539953 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.550999 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.561559 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.570507 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.578674 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.583274 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.583334 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.583346 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.583382 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.583401 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:08Z","lastTransitionTime":"2026-02-02T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.597202 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.604902 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.604954 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhqv6\" (UniqueName: \"kubernetes.io/projected/fe216e68-8814-468d-a6f3-fe8721d94eba-kube-api-access-zhqv6\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:08 crc kubenswrapper[4846]: E0202 12:10:08.605085 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:08 crc kubenswrapper[4846]: E0202 12:10:08.605149 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs podName:fe216e68-8814-468d-a6f3-fe8721d94eba nodeName:}" failed. No retries permitted until 2026-02-02 12:10:09.105133192 +0000 UTC m=+40.333720055 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs") pod "network-metrics-daemon-j5m9k" (UID: "fe216e68-8814-468d-a6f3-fe8721d94eba") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.609122 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.620504 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.622587 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhqv6\" (UniqueName: \"kubernetes.io/projected/fe216e68-8814-468d-a6f3-fe8721d94eba-kube-api-access-zhqv6\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.639281 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.650258 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.685878 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.685914 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.685923 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.685935 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.685944 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:08Z","lastTransitionTime":"2026-02-02T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.711439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" event={"ID":"e681c0d1-0580-45d2-a381-540fded41685","Type":"ContainerStarted","Data":"3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.711742 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" event={"ID":"e681c0d1-0580-45d2-a381-540fded41685","Type":"ContainerStarted","Data":"11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.742611 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.755876 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.768959 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.779176 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.788524 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.788562 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.788571 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.788585 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.788597 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:08Z","lastTransitionTime":"2026-02-02T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.789230 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.800889 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.814061 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.827269 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.845340 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.857602 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.871672 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.881923 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.891352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.891391 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.891400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.891438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.891449 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:08Z","lastTransitionTime":"2026-02-02T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.894931 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.905571 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.921864 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.935896 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.948514 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:08Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.993901 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.993957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.993967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.993981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:08 crc kubenswrapper[4846]: I0202 12:10:08.993991 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:08Z","lastTransitionTime":"2026-02-02T12:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.096361 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.096392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.096401 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.096429 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.096438 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:09Z","lastTransitionTime":"2026-02-02T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.109169 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:09 crc kubenswrapper[4846]: E0202 12:10:09.109312 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:09 crc kubenswrapper[4846]: E0202 12:10:09.109367 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs podName:fe216e68-8814-468d-a6f3-fe8721d94eba nodeName:}" failed. No retries permitted until 2026-02-02 12:10:10.109350678 +0000 UTC m=+41.337937541 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs") pod "network-metrics-daemon-j5m9k" (UID: "fe216e68-8814-468d-a6f3-fe8721d94eba") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.198432 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.198483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.198498 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.198516 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.198528 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:09Z","lastTransitionTime":"2026-02-02T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.300986 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.301023 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.301033 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.301049 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.301063 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:09Z","lastTransitionTime":"2026-02-02T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.392597 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 21:44:01.417669646 +0000 UTC Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.406270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.406313 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.406327 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.406346 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.406358 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:09Z","lastTransitionTime":"2026-02-02T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.422721 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.422774 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:09 crc kubenswrapper[4846]: E0202 12:10:09.422876 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:09 crc kubenswrapper[4846]: E0202 12:10:09.423004 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.423456 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:09 crc kubenswrapper[4846]: E0202 12:10:09.423708 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.436758 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.448347 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.461422 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.470991 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.480939 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.502733 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.508557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.508729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.508796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.508862 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.508983 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:09Z","lastTransitionTime":"2026-02-02T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.516838 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.530323 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.556191 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.567199 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.583808 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.596604 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.606994 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.610657 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.610690 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.610703 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.610719 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.610729 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:09Z","lastTransitionTime":"2026-02-02T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.621809 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.638428 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.653274 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.663843 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:09Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.713003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.713072 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.713087 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.713104 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.713115 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:09Z","lastTransitionTime":"2026-02-02T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.815546 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.815574 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.815582 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.815595 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.815603 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:09Z","lastTransitionTime":"2026-02-02T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.917556 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.917590 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.917599 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.917613 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:09 crc kubenswrapper[4846]: I0202 12:10:09.917637 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:09Z","lastTransitionTime":"2026-02-02T12:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.020410 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.020437 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.020445 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.020459 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.020468 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:10Z","lastTransitionTime":"2026-02-02T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.118811 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:10 crc kubenswrapper[4846]: E0202 12:10:10.118953 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:10 crc kubenswrapper[4846]: E0202 12:10:10.119008 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs podName:fe216e68-8814-468d-a6f3-fe8721d94eba nodeName:}" failed. No retries permitted until 2026-02-02 12:10:12.118991425 +0000 UTC m=+43.347578288 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs") pod "network-metrics-daemon-j5m9k" (UID: "fe216e68-8814-468d-a6f3-fe8721d94eba") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.124430 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.124461 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.124469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.124481 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.124490 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:10Z","lastTransitionTime":"2026-02-02T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.227432 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.227467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.227476 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.227490 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.227499 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:10Z","lastTransitionTime":"2026-02-02T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.330107 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.330140 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.330148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.330162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.330171 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:10Z","lastTransitionTime":"2026-02-02T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.393269 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 16:49:45.197289778 +0000 UTC Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.422456 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:10 crc kubenswrapper[4846]: E0202 12:10:10.422794 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.432376 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.432440 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.432465 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.432493 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.432517 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:10Z","lastTransitionTime":"2026-02-02T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.535328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.535370 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.535381 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.535398 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.535409 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:10Z","lastTransitionTime":"2026-02-02T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.638016 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.638072 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.638081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.638095 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.638105 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:10Z","lastTransitionTime":"2026-02-02T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.740678 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.740716 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.740724 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.740738 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.740750 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:10Z","lastTransitionTime":"2026-02-02T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.843422 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.843464 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.843472 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.843486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.843497 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:10Z","lastTransitionTime":"2026-02-02T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.946097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.946139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.946159 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.946181 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:10 crc kubenswrapper[4846]: I0202 12:10:10.946190 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:10Z","lastTransitionTime":"2026-02-02T12:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.048924 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.048960 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.048968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.048982 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.048990 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:11Z","lastTransitionTime":"2026-02-02T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.150974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.151017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.151025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.151039 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.151049 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:11Z","lastTransitionTime":"2026-02-02T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.254102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.254167 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.254181 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.254203 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.254219 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:11Z","lastTransitionTime":"2026-02-02T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.356922 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.356969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.356985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.357005 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.357019 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:11Z","lastTransitionTime":"2026-02-02T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.394267 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 07:25:43.603442856 +0000 UTC Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.423057 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:11 crc kubenswrapper[4846]: E0202 12:10:11.423383 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.423150 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:11 crc kubenswrapper[4846]: E0202 12:10:11.423577 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.423097 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:11 crc kubenswrapper[4846]: E0202 12:10:11.423837 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.460074 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.460347 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.460508 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.460605 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.460704 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:11Z","lastTransitionTime":"2026-02-02T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.564108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.564143 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.564151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.564165 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.564174 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:11Z","lastTransitionTime":"2026-02-02T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.667438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.667496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.667518 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.667551 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.667575 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:11Z","lastTransitionTime":"2026-02-02T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.770409 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.770446 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.770458 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.770477 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.770488 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:11Z","lastTransitionTime":"2026-02-02T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.876666 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.876736 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.876769 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.876790 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.876801 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:11Z","lastTransitionTime":"2026-02-02T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.978701 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.978736 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.978744 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.978756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:11 crc kubenswrapper[4846]: I0202 12:10:11.978765 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:11Z","lastTransitionTime":"2026-02-02T12:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.081046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.081088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.081099 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.081114 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.081125 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:12Z","lastTransitionTime":"2026-02-02T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.138436 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:12 crc kubenswrapper[4846]: E0202 12:10:12.138557 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:12 crc kubenswrapper[4846]: E0202 12:10:12.138644 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs podName:fe216e68-8814-468d-a6f3-fe8721d94eba nodeName:}" failed. No retries permitted until 2026-02-02 12:10:16.138603838 +0000 UTC m=+47.367190701 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs") pod "network-metrics-daemon-j5m9k" (UID: "fe216e68-8814-468d-a6f3-fe8721d94eba") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.182697 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.182765 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.182787 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.182815 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.182841 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:12Z","lastTransitionTime":"2026-02-02T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.285415 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.285457 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.285468 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.285484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.285494 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:12Z","lastTransitionTime":"2026-02-02T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.388334 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.388367 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.388375 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.388390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.388407 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:12Z","lastTransitionTime":"2026-02-02T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.394662 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 21:00:30.017436234 +0000 UTC Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.423031 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:12 crc kubenswrapper[4846]: E0202 12:10:12.423217 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.491272 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.491319 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.491333 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.491350 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.491362 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:12Z","lastTransitionTime":"2026-02-02T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.594120 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.594159 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.594171 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.594188 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.594199 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:12Z","lastTransitionTime":"2026-02-02T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.696528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.696570 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.696578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.696591 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.696601 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:12Z","lastTransitionTime":"2026-02-02T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.799843 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.799909 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.799928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.799959 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.799977 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:12Z","lastTransitionTime":"2026-02-02T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.903556 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.903604 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.903645 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.903671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:12 crc kubenswrapper[4846]: I0202 12:10:12.903685 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:12Z","lastTransitionTime":"2026-02-02T12:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.006895 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.007051 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.007081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.007110 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.007132 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:13Z","lastTransitionTime":"2026-02-02T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.110122 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.110209 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.110233 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.110266 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.110289 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:13Z","lastTransitionTime":"2026-02-02T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.212753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.212788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.212796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.212809 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.212819 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:13Z","lastTransitionTime":"2026-02-02T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.316086 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.316161 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.316267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.316327 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.316354 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:13Z","lastTransitionTime":"2026-02-02T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.395298 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 07:10:52.691533851 +0000 UTC Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.420506 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.420590 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.420615 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.420693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.420720 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:13Z","lastTransitionTime":"2026-02-02T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.422890 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.422983 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:13 crc kubenswrapper[4846]: E0202 12:10:13.423048 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:13 crc kubenswrapper[4846]: E0202 12:10:13.423186 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.422984 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:13 crc kubenswrapper[4846]: E0202 12:10:13.423473 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.523487 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.523524 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.523532 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.523546 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.523555 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:13Z","lastTransitionTime":"2026-02-02T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.625782 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.625868 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.625887 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.625954 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.625974 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:13Z","lastTransitionTime":"2026-02-02T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.730798 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.730858 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.730875 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.730891 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.730901 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:13Z","lastTransitionTime":"2026-02-02T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.833654 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.833696 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.833708 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.833725 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.833739 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:13Z","lastTransitionTime":"2026-02-02T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.935898 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.935936 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.935946 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.935962 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:13 crc kubenswrapper[4846]: I0202 12:10:13.935974 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:13Z","lastTransitionTime":"2026-02-02T12:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.038992 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.039031 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.039042 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.039059 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.039070 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:14Z","lastTransitionTime":"2026-02-02T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.140498 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.140557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.140573 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.140593 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.140607 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:14Z","lastTransitionTime":"2026-02-02T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.243009 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.243045 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.243054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.243070 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.243080 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:14Z","lastTransitionTime":"2026-02-02T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.345609 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.345667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.345680 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.345696 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.345708 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:14Z","lastTransitionTime":"2026-02-02T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.396103 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 11:37:54.045780067 +0000 UTC Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.422528 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:14 crc kubenswrapper[4846]: E0202 12:10:14.422762 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.448326 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.448390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.448410 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.448437 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.448460 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:14Z","lastTransitionTime":"2026-02-02T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.551106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.551187 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.551212 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.551238 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.551256 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:14Z","lastTransitionTime":"2026-02-02T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.653650 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.653685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.653694 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.653709 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.653720 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:14Z","lastTransitionTime":"2026-02-02T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.755964 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.755996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.756007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.756026 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.756036 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:14Z","lastTransitionTime":"2026-02-02T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.857889 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.857928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.857940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.857955 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.857966 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:14Z","lastTransitionTime":"2026-02-02T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.961561 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.961985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.962011 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.962041 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:14 crc kubenswrapper[4846]: I0202 12:10:14.962062 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:14Z","lastTransitionTime":"2026-02-02T12:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.065703 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.065761 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.065774 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.065795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.065810 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.167766 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.167814 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.167831 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.167847 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.167856 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.270522 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.270559 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.270567 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.270580 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.270588 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.372924 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.372966 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.372975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.372990 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.372999 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.397140 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 02:21:09.826791264 +0000 UTC Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.422868 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:15 crc kubenswrapper[4846]: E0202 12:10:15.423404 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.423883 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:15 crc kubenswrapper[4846]: E0202 12:10:15.424065 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.424083 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:15 crc kubenswrapper[4846]: E0202 12:10:15.424353 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.475919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.476146 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.476232 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.476312 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.476426 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.514502 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.514556 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.514568 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.514585 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.514598 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: E0202 12:10:15.526435 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:15Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.536160 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.536207 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.536219 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.536237 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.536249 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: E0202 12:10:15.549921 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:15Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.554086 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.554139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.554150 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.554168 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.554178 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: E0202 12:10:15.567660 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:15Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.571202 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.571250 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.571262 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.571275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.571284 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: E0202 12:10:15.584116 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:15Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.588017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.588046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.588054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.588068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.588079 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: E0202 12:10:15.600745 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:15Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:15 crc kubenswrapper[4846]: E0202 12:10:15.600890 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.602575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.602724 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.602807 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.602897 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.603011 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.705019 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.705076 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.705091 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.705112 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.705126 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.807909 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.807990 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.808008 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.808033 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.808052 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.910957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.911024 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.911036 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.911054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:15 crc kubenswrapper[4846]: I0202 12:10:15.911065 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:15Z","lastTransitionTime":"2026-02-02T12:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.013925 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.013952 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.013960 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.013975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.013983 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:16Z","lastTransitionTime":"2026-02-02T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.116253 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.116296 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.116306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.116325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.116336 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:16Z","lastTransitionTime":"2026-02-02T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.178021 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:16 crc kubenswrapper[4846]: E0202 12:10:16.178198 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:16 crc kubenswrapper[4846]: E0202 12:10:16.178314 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs podName:fe216e68-8814-468d-a6f3-fe8721d94eba nodeName:}" failed. No retries permitted until 2026-02-02 12:10:24.178281233 +0000 UTC m=+55.406868136 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs") pod "network-metrics-daemon-j5m9k" (UID: "fe216e68-8814-468d-a6f3-fe8721d94eba") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.217816 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.217849 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.217858 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.217873 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.217890 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:16Z","lastTransitionTime":"2026-02-02T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.320359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.320406 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.320423 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.320447 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.320465 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:16Z","lastTransitionTime":"2026-02-02T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.398123 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 02:34:57.824759665 +0000 UTC Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.422588 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:16 crc kubenswrapper[4846]: E0202 12:10:16.422742 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.422881 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.422948 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.423049 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.423073 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.423122 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:16Z","lastTransitionTime":"2026-02-02T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.526151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.526199 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.526213 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.526234 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.526248 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:16Z","lastTransitionTime":"2026-02-02T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.628134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.628230 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.628244 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.628259 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.628269 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:16Z","lastTransitionTime":"2026-02-02T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.731478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.731550 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.731576 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.731606 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.731664 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:16Z","lastTransitionTime":"2026-02-02T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.834713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.834760 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.834777 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.834796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.834811 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:16Z","lastTransitionTime":"2026-02-02T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.937046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.937083 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.937098 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.937121 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:16 crc kubenswrapper[4846]: I0202 12:10:16.937135 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:16Z","lastTransitionTime":"2026-02-02T12:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.040669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.040744 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.040761 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.040789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.040804 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:17Z","lastTransitionTime":"2026-02-02T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.142501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.142536 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.142545 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.142560 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.142569 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:17Z","lastTransitionTime":"2026-02-02T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.245073 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.245106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.245114 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.245128 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.245145 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:17Z","lastTransitionTime":"2026-02-02T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.347954 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.348010 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.348028 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.348053 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.348070 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:17Z","lastTransitionTime":"2026-02-02T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.398980 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 08:58:32.388164737 +0000 UTC Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.423675 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.423696 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.423927 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:17 crc kubenswrapper[4846]: E0202 12:10:17.424392 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:17 crc kubenswrapper[4846]: E0202 12:10:17.424806 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:17 crc kubenswrapper[4846]: E0202 12:10:17.424981 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.451122 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.451196 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.451220 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.451254 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.451277 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:17Z","lastTransitionTime":"2026-02-02T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.555159 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.555218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.555232 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.555252 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.555267 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:17Z","lastTransitionTime":"2026-02-02T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.658504 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.658549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.658569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.658590 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.658603 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:17Z","lastTransitionTime":"2026-02-02T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.760932 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.760977 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.760985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.761013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.761023 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:17Z","lastTransitionTime":"2026-02-02T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.863784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.864135 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.864326 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.864558 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.864796 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:17Z","lastTransitionTime":"2026-02-02T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.968115 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.968161 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.968174 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.968192 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:17 crc kubenswrapper[4846]: I0202 12:10:17.968203 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:17Z","lastTransitionTime":"2026-02-02T12:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.071284 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.071332 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.071352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.071379 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.071398 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:18Z","lastTransitionTime":"2026-02-02T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.174537 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.174568 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.174576 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.174590 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.174597 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:18Z","lastTransitionTime":"2026-02-02T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.277466 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.277546 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.277557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.277574 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.277585 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:18Z","lastTransitionTime":"2026-02-02T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.380088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.380125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.380136 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.380154 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.380166 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:18Z","lastTransitionTime":"2026-02-02T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.399779 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 16:40:31.548652743 +0000 UTC Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.423402 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:18 crc kubenswrapper[4846]: E0202 12:10:18.423582 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.482473 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.482735 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.482806 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.482881 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.482943 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:18Z","lastTransitionTime":"2026-02-02T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.587069 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.587135 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.587151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.587172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.587191 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:18Z","lastTransitionTime":"2026-02-02T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.689428 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.689765 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.689877 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.689977 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.690157 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:18Z","lastTransitionTime":"2026-02-02T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.793075 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.793127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.793143 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.793168 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.793185 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:18Z","lastTransitionTime":"2026-02-02T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.896294 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.896375 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.896399 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.896431 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.896457 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:18Z","lastTransitionTime":"2026-02-02T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.999838 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.999915 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.999947 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:18 crc kubenswrapper[4846]: I0202 12:10:18.999978 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.000091 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:18Z","lastTransitionTime":"2026-02-02T12:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.102422 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.102469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.102481 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.102497 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.102506 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:19Z","lastTransitionTime":"2026-02-02T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.204433 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.204484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.204501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.204537 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.204547 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:19Z","lastTransitionTime":"2026-02-02T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.307307 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.307354 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.307369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.307390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.307406 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:19Z","lastTransitionTime":"2026-02-02T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.400852 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 21:29:20.544073714 +0000 UTC Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.410403 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.410492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.410570 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.410671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.410704 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:19Z","lastTransitionTime":"2026-02-02T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.422818 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.422900 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:19 crc kubenswrapper[4846]: E0202 12:10:19.423046 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.422824 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:19 crc kubenswrapper[4846]: E0202 12:10:19.423681 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:19 crc kubenswrapper[4846]: E0202 12:10:19.423927 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.424367 4846 scope.go:117] "RemoveContainer" containerID="67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.445897 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.457870 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.470445 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.481250 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.495656 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.508062 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.513646 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.513680 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.513693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.513711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.513722 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:19Z","lastTransitionTime":"2026-02-02T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.520428 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.533497 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.544657 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.570722 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.583757 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.595418 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.604755 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.615815 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.615839 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.615849 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.615864 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.615874 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:19Z","lastTransitionTime":"2026-02-02T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.616445 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.631265 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.642936 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.663710 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.717501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.717668 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.717692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.717714 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.717726 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:19Z","lastTransitionTime":"2026-02-02T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.747384 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/1.log" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.749098 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.749215 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.763694 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.779450 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.797065 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.813949 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.819800 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.819836 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.819844 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.819859 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.819869 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:19Z","lastTransitionTime":"2026-02-02T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.827680 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.839238 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.852076 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.860949 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.871843 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.880990 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.897784 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.909278 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.923000 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.923046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.923056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.923074 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.923115 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:19Z","lastTransitionTime":"2026-02-02T12:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.923845 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.946118 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.959101 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.974157 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:19 crc kubenswrapper[4846]: I0202 12:10:19.988023 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:19Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.025102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.025151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.025163 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.025178 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.025189 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:20Z","lastTransitionTime":"2026-02-02T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.127372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.127710 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.127760 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.127774 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.127782 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:20Z","lastTransitionTime":"2026-02-02T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.230411 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.230471 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.230486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.230512 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.230527 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:20Z","lastTransitionTime":"2026-02-02T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.332717 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.332751 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.332762 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.332778 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.332790 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:20Z","lastTransitionTime":"2026-02-02T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.402094 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 06:08:04.669344263 +0000 UTC Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.422597 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:20 crc kubenswrapper[4846]: E0202 12:10:20.422852 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.436029 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.436128 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.436146 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.436171 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.436227 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:20Z","lastTransitionTime":"2026-02-02T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.538935 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.538998 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.539011 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.539027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.539038 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:20Z","lastTransitionTime":"2026-02-02T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.640889 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.640954 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.640974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.641004 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.641039 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:20Z","lastTransitionTime":"2026-02-02T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.744321 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.744479 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.744503 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.744531 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.744549 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:20Z","lastTransitionTime":"2026-02-02T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.754928 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/2.log" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.756079 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/1.log" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.761185 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce" exitCode=1 Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.761248 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.761302 4846 scope.go:117] "RemoveContainer" containerID="67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.762497 4846 scope.go:117] "RemoveContainer" containerID="91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce" Feb 02 12:10:20 crc kubenswrapper[4846]: E0202 12:10:20.762811 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.781978 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.796931 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.832646 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.848941 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.848979 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.848995 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.849017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.849032 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:20Z","lastTransitionTime":"2026-02-02T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.850103 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.863448 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.879312 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.892039 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.906937 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.927286 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.941926 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.951982 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.952021 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.952033 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.952053 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.952068 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:20Z","lastTransitionTime":"2026-02-02T12:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.960925 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.974705 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:20 crc kubenswrapper[4846]: I0202 12:10:20.989831 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.002168 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:20Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.018321 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:21Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.034773 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:21Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.055296 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.055355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.055368 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.055390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.055404 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:21Z","lastTransitionTime":"2026-02-02T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.055434 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:21Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.158708 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.158757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.158770 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.158788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.158801 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:21Z","lastTransitionTime":"2026-02-02T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.231293 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.231408 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.231438 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.231470 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.231598 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.231658 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.231715 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:10:53.231662449 +0000 UTC m=+84.460249322 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.231665 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.231776 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:53.231764313 +0000 UTC m=+84.460351196 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.231786 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.231837 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:53.231820974 +0000 UTC m=+84.460407837 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.231764 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.232029 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:53.231980737 +0000 UTC m=+84.460567610 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.261348 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.261393 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.261404 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.261421 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.261433 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:21Z","lastTransitionTime":"2026-02-02T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.332163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.332284 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.332309 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.332324 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.332381 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 12:10:53.332369087 +0000 UTC m=+84.560955940 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.363959 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.364017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.364033 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.364056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.364072 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:21Z","lastTransitionTime":"2026-02-02T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.402760 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 17:51:20.078635005 +0000 UTC Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.423283 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.423321 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.423328 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.423506 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.423564 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:21 crc kubenswrapper[4846]: E0202 12:10:21.423662 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.466715 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.466758 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.466771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.466788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.466800 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:21Z","lastTransitionTime":"2026-02-02T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.569291 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.569325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.569337 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.569354 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.569365 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:21Z","lastTransitionTime":"2026-02-02T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.672019 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.672054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.672064 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.672081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.672090 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:21Z","lastTransitionTime":"2026-02-02T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.788198 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.788258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.788274 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.788295 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.788311 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:21Z","lastTransitionTime":"2026-02-02T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.791491 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/2.log" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.889122 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.890676 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.890707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.890717 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.890731 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.890739 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:21Z","lastTransitionTime":"2026-02-02T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.905136 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:21Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.931267 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:21Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.945064 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:21Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.958212 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:21Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.969951 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:21Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.985444 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:21Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.993119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.993315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.993407 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.993522 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:21 crc kubenswrapper[4846]: I0202 12:10:21.993663 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:21Z","lastTransitionTime":"2026-02-02T12:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.004582 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.018994 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.036512 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.045152 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.055392 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.059278 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.071653 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.085330 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.095478 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.095885 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.095919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.095927 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.095942 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.095951 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:22Z","lastTransitionTime":"2026-02-02T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.109047 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.123385 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.138260 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.148292 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.158820 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.168983 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.178863 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.198821 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.199124 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.199242 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.199333 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.199412 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:22Z","lastTransitionTime":"2026-02-02T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.205209 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.219420 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.240504 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.264422 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67609740fc407adc26e1ad357d522e9507c22133fb597506a9e8f8c300da5046\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:05Z\\\",\\\"message\\\":\\\" handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0202 12:10:05.550104 6271 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0202 12:10:05.550128 6271 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0202 12:10:05.550212 6271 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0202 12:10:05.550227 6271 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0202 12:10:05.550326 6271 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0202 12:10:05.550339 6271 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0202 12:10:05.550361 6271 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0202 12:10:05.550385 6271 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0202 12:10:05.550390 6271 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0202 12:10:05.550404 6271 factory.go:656] Stopping watch factory\\\\nI0202 12:10:05.550422 6271 ovnkube.go:599] Stopped ovnkube\\\\nI0202 12:10:05.550456 6271 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0202 12:10:05.550472 6271 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0202 12:10:05.550461 6271 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nF0202 12:10:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.277708 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.289551 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.301708 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.302138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.302190 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.302203 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.302225 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.302239 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:22Z","lastTransitionTime":"2026-02-02T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.314594 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.325067 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.338938 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.351517 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.361545 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.369238 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.378532 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.387512 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.402985 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 13:46:54.752829514 +0000 UTC Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.404652 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.404683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.404698 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.404722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.404733 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:22Z","lastTransitionTime":"2026-02-02T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.423313 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:22 crc kubenswrapper[4846]: E0202 12:10:22.423438 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.507232 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.507278 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.507289 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.507306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.507318 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:22Z","lastTransitionTime":"2026-02-02T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.609372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.609413 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.609424 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.609439 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.609451 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:22Z","lastTransitionTime":"2026-02-02T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.713872 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.713939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.713957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.713985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.714003 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:22Z","lastTransitionTime":"2026-02-02T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.817047 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.817088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.817099 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.817117 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.817129 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:22Z","lastTransitionTime":"2026-02-02T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.889191 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.890436 4846 scope.go:117] "RemoveContainer" containerID="91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce" Feb 02 12:10:22 crc kubenswrapper[4846]: E0202 12:10:22.890775 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.914380 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.919171 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.919396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.919474 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.919550 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.919636 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:22Z","lastTransitionTime":"2026-02-02T12:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.936451 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.953962 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.965550 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:22 crc kubenswrapper[4846]: I0202 12:10:22.986955 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:22Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.004585 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.018171 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.021838 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.021952 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.022023 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.022097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.022185 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:23Z","lastTransitionTime":"2026-02-02T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.031579 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.040614 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.055925 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.066748 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.077280 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.086472 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.095431 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.113415 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.124919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.125052 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.125069 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.125088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.125106 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:23Z","lastTransitionTime":"2026-02-02T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.128729 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.140713 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.153125 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:23Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.227984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.228032 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.228044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.228061 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.228074 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:23Z","lastTransitionTime":"2026-02-02T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.331606 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.331728 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.331746 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.331774 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.331790 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:23Z","lastTransitionTime":"2026-02-02T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.403229 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 07:55:23.237734398 +0000 UTC Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.422996 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.423114 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:23 crc kubenswrapper[4846]: E0202 12:10:23.423213 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.423028 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:23 crc kubenswrapper[4846]: E0202 12:10:23.423293 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:23 crc kubenswrapper[4846]: E0202 12:10:23.423438 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.434453 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.434494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.434508 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.434526 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.434541 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:23Z","lastTransitionTime":"2026-02-02T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.537002 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.537320 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.537511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.537695 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.537842 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:23Z","lastTransitionTime":"2026-02-02T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.641297 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.641362 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.641383 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.641411 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.641432 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:23Z","lastTransitionTime":"2026-02-02T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.743600 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.743637 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.743648 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.743661 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.743672 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:23Z","lastTransitionTime":"2026-02-02T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.846394 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.846436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.846445 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.846460 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.846470 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:23Z","lastTransitionTime":"2026-02-02T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.948419 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.948462 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.948476 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.948494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:23 crc kubenswrapper[4846]: I0202 12:10:23.948506 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:23Z","lastTransitionTime":"2026-02-02T12:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.051793 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.051859 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.051868 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.051902 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.051912 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:24Z","lastTransitionTime":"2026-02-02T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.154707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.154805 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.154824 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.154848 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.154867 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:24Z","lastTransitionTime":"2026-02-02T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.257193 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.257245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.257260 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.257279 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.257290 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:24Z","lastTransitionTime":"2026-02-02T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.262679 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:24 crc kubenswrapper[4846]: E0202 12:10:24.262844 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:24 crc kubenswrapper[4846]: E0202 12:10:24.262913 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs podName:fe216e68-8814-468d-a6f3-fe8721d94eba nodeName:}" failed. No retries permitted until 2026-02-02 12:10:40.262888584 +0000 UTC m=+71.491475487 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs") pod "network-metrics-daemon-j5m9k" (UID: "fe216e68-8814-468d-a6f3-fe8721d94eba") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.360957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.361017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.361037 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.361060 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.361077 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:24Z","lastTransitionTime":"2026-02-02T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.404101 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 22:11:49.939247388 +0000 UTC Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.422427 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:24 crc kubenswrapper[4846]: E0202 12:10:24.422606 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.464771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.465049 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.465113 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.465181 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.465263 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:24Z","lastTransitionTime":"2026-02-02T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.567924 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.568283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.568396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.568514 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.568698 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:24Z","lastTransitionTime":"2026-02-02T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.672101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.672142 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.672153 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.672169 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.672180 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:24Z","lastTransitionTime":"2026-02-02T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.774496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.774805 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.775055 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.775355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.775594 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:24Z","lastTransitionTime":"2026-02-02T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.878166 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.878236 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.878249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.878275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.878294 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:24Z","lastTransitionTime":"2026-02-02T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.981327 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.981384 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.981400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.981421 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:24 crc kubenswrapper[4846]: I0202 12:10:24.981439 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:24Z","lastTransitionTime":"2026-02-02T12:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.084249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.084290 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.084299 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.084315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.084328 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.186601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.186658 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.186669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.186686 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.186697 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.289211 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.289243 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.289251 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.289266 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.289274 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.391678 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.391716 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.391727 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.391749 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.391761 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.404568 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 20:23:38.75019362 +0000 UTC Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.422549 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:25 crc kubenswrapper[4846]: E0202 12:10:25.422799 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.422915 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.422928 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:25 crc kubenswrapper[4846]: E0202 12:10:25.423134 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:25 crc kubenswrapper[4846]: E0202 12:10:25.423346 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.494701 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.494758 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.494774 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.494797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.494814 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.597016 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.597085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.597109 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.597136 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.597158 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.700851 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.700916 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.700934 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.700961 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.700980 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.804265 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.804348 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.804365 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.804394 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.804410 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.872485 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.872547 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.872568 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.872594 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.872612 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: E0202 12:10:25.889080 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:25Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.894196 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.894259 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.894276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.894296 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.894310 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: E0202 12:10:25.908847 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:25Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.912854 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.912888 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.912899 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.912915 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.912927 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: E0202 12:10:25.926678 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:25Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.931362 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.931442 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.931460 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.931481 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.931496 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: E0202 12:10:25.945723 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:25Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.950051 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.950143 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.950166 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.950210 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.950250 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:25 crc kubenswrapper[4846]: E0202 12:10:25.971085 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:25Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:25 crc kubenswrapper[4846]: E0202 12:10:25.971321 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.973370 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.973414 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.973424 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.973443 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:25 crc kubenswrapper[4846]: I0202 12:10:25.973455 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:25Z","lastTransitionTime":"2026-02-02T12:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.076092 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.076167 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.076185 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.076211 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.076229 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:26Z","lastTransitionTime":"2026-02-02T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.178437 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.178472 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.178480 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.178494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.178503 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:26Z","lastTransitionTime":"2026-02-02T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.280939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.280975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.280984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.280997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.281011 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:26Z","lastTransitionTime":"2026-02-02T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.384245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.384301 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.384390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.384433 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.384450 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:26Z","lastTransitionTime":"2026-02-02T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.404768 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 08:17:53.386728863 +0000 UTC Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.423172 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:26 crc kubenswrapper[4846]: E0202 12:10:26.423371 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.487942 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.488005 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.488048 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.488080 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.488105 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:26Z","lastTransitionTime":"2026-02-02T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.590919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.590971 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.590988 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.591013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.591031 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:26Z","lastTransitionTime":"2026-02-02T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.693613 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.693672 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.693681 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.693694 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.693703 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:26Z","lastTransitionTime":"2026-02-02T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.795953 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.795988 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.795996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.796011 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.796023 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:26Z","lastTransitionTime":"2026-02-02T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.899400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.899690 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.899854 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.899989 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:26 crc kubenswrapper[4846]: I0202 12:10:26.900077 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:26Z","lastTransitionTime":"2026-02-02T12:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.002926 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.002997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.003017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.003039 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.003053 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:27Z","lastTransitionTime":"2026-02-02T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.106161 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.106409 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.106491 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.106580 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.106723 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:27Z","lastTransitionTime":"2026-02-02T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.208928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.209155 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.209249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.209353 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.209431 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:27Z","lastTransitionTime":"2026-02-02T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.312115 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.312473 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.312575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.312767 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.312899 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:27Z","lastTransitionTime":"2026-02-02T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.405344 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 03:04:30.044081557 +0000 UTC Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.415404 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.415439 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.415451 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.415467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.415480 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:27Z","lastTransitionTime":"2026-02-02T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.423396 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:27 crc kubenswrapper[4846]: E0202 12:10:27.423558 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.423742 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.423843 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:27 crc kubenswrapper[4846]: E0202 12:10:27.423985 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:27 crc kubenswrapper[4846]: E0202 12:10:27.424007 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.517789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.517819 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.517828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.517842 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.517851 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:27Z","lastTransitionTime":"2026-02-02T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.620247 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.620285 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.620298 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.620318 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.620329 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:27Z","lastTransitionTime":"2026-02-02T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.724170 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.724222 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.724242 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.724339 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.724417 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:27Z","lastTransitionTime":"2026-02-02T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.826540 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.826579 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.826590 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.826607 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.826646 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:27Z","lastTransitionTime":"2026-02-02T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.929496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.929550 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.929564 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.929586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:27 crc kubenswrapper[4846]: I0202 12:10:27.929603 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:27Z","lastTransitionTime":"2026-02-02T12:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.032046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.032100 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.032115 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.032137 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.032154 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:28Z","lastTransitionTime":"2026-02-02T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.134497 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.134537 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.134549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.134566 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.134578 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:28Z","lastTransitionTime":"2026-02-02T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.237304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.237362 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.237379 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.237403 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.237422 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:28Z","lastTransitionTime":"2026-02-02T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.340183 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.340220 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.340228 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.340242 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.340253 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:28Z","lastTransitionTime":"2026-02-02T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.406503 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 19:40:59.554427489 +0000 UTC Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.422945 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:28 crc kubenswrapper[4846]: E0202 12:10:28.423091 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.442486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.442557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.442573 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.442593 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.442607 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:28Z","lastTransitionTime":"2026-02-02T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.544961 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.544990 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.545004 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.545020 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.545029 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:28Z","lastTransitionTime":"2026-02-02T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.647560 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.647644 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.647661 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.647690 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.647707 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:28Z","lastTransitionTime":"2026-02-02T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.750306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.750382 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.750404 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.750433 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.750454 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:28Z","lastTransitionTime":"2026-02-02T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.853072 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.853108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.853116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.853131 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.853143 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:28Z","lastTransitionTime":"2026-02-02T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.956248 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.956313 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.956330 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.956353 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:28 crc kubenswrapper[4846]: I0202 12:10:28.956377 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:28Z","lastTransitionTime":"2026-02-02T12:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.058777 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.058813 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.058825 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.058839 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.058851 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:29Z","lastTransitionTime":"2026-02-02T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.160758 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.160801 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.160817 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.160835 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.160848 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:29Z","lastTransitionTime":"2026-02-02T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.267793 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.267890 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.267905 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.267925 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.267939 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:29Z","lastTransitionTime":"2026-02-02T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.370720 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.370967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.371027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.371088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.371143 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:29Z","lastTransitionTime":"2026-02-02T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.407542 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 09:54:02.389496927 +0000 UTC Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.422459 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.422535 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.422571 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:29 crc kubenswrapper[4846]: E0202 12:10:29.422678 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:29 crc kubenswrapper[4846]: E0202 12:10:29.422894 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:29 crc kubenswrapper[4846]: E0202 12:10:29.422957 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.435939 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.446422 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.456805 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.470231 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.473684 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.473730 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.473745 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.473768 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.473817 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:29Z","lastTransitionTime":"2026-02-02T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.483222 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.496954 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.510506 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.518842 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.530003 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.540529 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.551358 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.561657 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.579694 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.579740 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.579751 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.579768 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.579784 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:29Z","lastTransitionTime":"2026-02-02T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.582529 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.595478 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.606682 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.625315 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.637180 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.648215 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:29Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.682277 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.682330 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.682343 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.682362 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.682374 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:29Z","lastTransitionTime":"2026-02-02T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.784863 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.784925 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.784937 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.784955 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.785244 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:29Z","lastTransitionTime":"2026-02-02T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.888276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.888306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.888314 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.888328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.888336 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:29Z","lastTransitionTime":"2026-02-02T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.991246 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.991303 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.991321 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.991347 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:29 crc kubenswrapper[4846]: I0202 12:10:29.991365 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:29Z","lastTransitionTime":"2026-02-02T12:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.094389 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.094457 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.094478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.094504 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.094528 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:30Z","lastTransitionTime":"2026-02-02T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.198691 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.198761 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.198784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.198812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.198836 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:30Z","lastTransitionTime":"2026-02-02T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.301096 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.301158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.301180 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.301207 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.301228 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:30Z","lastTransitionTime":"2026-02-02T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.404755 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.405053 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.405066 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.405084 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.405096 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:30Z","lastTransitionTime":"2026-02-02T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.407970 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 07:34:29.758781676 +0000 UTC Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.423310 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:30 crc kubenswrapper[4846]: E0202 12:10:30.423460 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.507920 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.507967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.507976 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.508012 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.508026 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:30Z","lastTransitionTime":"2026-02-02T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.610750 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.610808 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.610819 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.610834 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.610844 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:30Z","lastTransitionTime":"2026-02-02T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.713540 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.713610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.713674 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.713706 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.713733 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:30Z","lastTransitionTime":"2026-02-02T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.816195 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.816245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.816256 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.816271 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.816281 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:30Z","lastTransitionTime":"2026-02-02T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.918906 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.918939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.918986 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.919002 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:30 crc kubenswrapper[4846]: I0202 12:10:30.919013 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:30Z","lastTransitionTime":"2026-02-02T12:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.022308 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.022372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.022415 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.022455 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.022478 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:31Z","lastTransitionTime":"2026-02-02T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.125175 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.125238 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.125255 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.125280 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.125304 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:31Z","lastTransitionTime":"2026-02-02T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.228686 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.228784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.228796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.228814 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.228825 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:31Z","lastTransitionTime":"2026-02-02T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.331276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.331331 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.331343 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.331360 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.331372 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:31Z","lastTransitionTime":"2026-02-02T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.408862 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 07:29:13.17177809 +0000 UTC Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.423264 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.423286 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.423357 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:31 crc kubenswrapper[4846]: E0202 12:10:31.423528 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:31 crc kubenswrapper[4846]: E0202 12:10:31.423759 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:31 crc kubenswrapper[4846]: E0202 12:10:31.423948 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.434233 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.434275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.434283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.434298 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.434307 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:31Z","lastTransitionTime":"2026-02-02T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.537180 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.537236 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.537249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.537268 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.537277 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:31Z","lastTransitionTime":"2026-02-02T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.640002 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.640046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.640054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.640067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.640076 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:31Z","lastTransitionTime":"2026-02-02T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.741931 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.742021 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.742040 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.742067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.742084 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:31Z","lastTransitionTime":"2026-02-02T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.844307 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.844363 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.844383 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.844405 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.844421 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:31Z","lastTransitionTime":"2026-02-02T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.946923 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.946985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.947003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.947028 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:31 crc kubenswrapper[4846]: I0202 12:10:31.947045 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:31Z","lastTransitionTime":"2026-02-02T12:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.049594 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.049710 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.049731 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.050129 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.050188 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:32Z","lastTransitionTime":"2026-02-02T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.153535 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.153576 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.153589 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.153608 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.153643 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:32Z","lastTransitionTime":"2026-02-02T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.256863 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.256911 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.256926 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.256950 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.256966 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:32Z","lastTransitionTime":"2026-02-02T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.360123 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.360155 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.360165 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.360182 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.360194 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:32Z","lastTransitionTime":"2026-02-02T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.409008 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 00:00:00.643657887 +0000 UTC Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.422407 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:32 crc kubenswrapper[4846]: E0202 12:10:32.422541 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.463920 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.463955 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.463964 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.463980 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.463990 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:32Z","lastTransitionTime":"2026-02-02T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.566464 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.566496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.566504 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.566517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.566525 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:32Z","lastTransitionTime":"2026-02-02T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.669655 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.669708 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.669722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.669746 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.669760 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:32Z","lastTransitionTime":"2026-02-02T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.771900 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.771945 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.771956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.771974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.771987 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:32Z","lastTransitionTime":"2026-02-02T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.874457 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.874485 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.874493 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.874505 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.874513 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:32Z","lastTransitionTime":"2026-02-02T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.976973 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.977027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.977036 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.977049 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:32 crc kubenswrapper[4846]: I0202 12:10:32.977057 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:32Z","lastTransitionTime":"2026-02-02T12:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.079964 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.080071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.080085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.080106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.080119 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:33Z","lastTransitionTime":"2026-02-02T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.182944 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.182980 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.182991 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.183007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.183020 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:33Z","lastTransitionTime":"2026-02-02T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.285644 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.285698 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.285707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.285722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.285731 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:33Z","lastTransitionTime":"2026-02-02T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.388596 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.388651 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.388666 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.388688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.388699 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:33Z","lastTransitionTime":"2026-02-02T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.409567 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 10:51:33.730272696 +0000 UTC Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.422892 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:33 crc kubenswrapper[4846]: E0202 12:10:33.423033 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.423061 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.422896 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:33 crc kubenswrapper[4846]: E0202 12:10:33.423179 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:33 crc kubenswrapper[4846]: E0202 12:10:33.423311 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.492224 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.492277 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.492287 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.492303 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.492313 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:33Z","lastTransitionTime":"2026-02-02T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.595345 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.595392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.595403 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.595421 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.595433 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:33Z","lastTransitionTime":"2026-02-02T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.697968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.698023 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.698035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.698436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.698467 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:33Z","lastTransitionTime":"2026-02-02T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.801062 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.801100 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.801112 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.801127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.801137 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:33Z","lastTransitionTime":"2026-02-02T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.903277 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.903315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.903326 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.903341 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:33 crc kubenswrapper[4846]: I0202 12:10:33.903352 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:33Z","lastTransitionTime":"2026-02-02T12:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.006280 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.006322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.006333 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.006353 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.006366 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:34Z","lastTransitionTime":"2026-02-02T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.108881 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.108915 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.108925 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.108940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.108951 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:34Z","lastTransitionTime":"2026-02-02T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.211101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.211127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.211138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.211153 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.211164 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:34Z","lastTransitionTime":"2026-02-02T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.313356 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.313382 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.313391 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.313405 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.313413 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:34Z","lastTransitionTime":"2026-02-02T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.410096 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 18:57:30.913526186 +0000 UTC Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.416204 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.416236 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.416248 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.416266 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.416278 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:34Z","lastTransitionTime":"2026-02-02T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.422492 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:34 crc kubenswrapper[4846]: E0202 12:10:34.422921 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.423193 4846 scope.go:117] "RemoveContainer" containerID="91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce" Feb 02 12:10:34 crc kubenswrapper[4846]: E0202 12:10:34.423449 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.518811 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.518842 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.518852 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.518866 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.518874 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:34Z","lastTransitionTime":"2026-02-02T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.621391 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.621445 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.621457 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.621477 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.621489 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:34Z","lastTransitionTime":"2026-02-02T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.724058 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.724094 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.724103 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.724116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.724126 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:34Z","lastTransitionTime":"2026-02-02T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.826770 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.826803 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.826810 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.826824 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.826834 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:34Z","lastTransitionTime":"2026-02-02T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.929798 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.929859 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.929872 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.929898 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:34 crc kubenswrapper[4846]: I0202 12:10:34.929912 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:34Z","lastTransitionTime":"2026-02-02T12:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.032066 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.032102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.032110 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.032123 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.032132 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:35Z","lastTransitionTime":"2026-02-02T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.134467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.134523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.134536 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.134552 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.134590 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:35Z","lastTransitionTime":"2026-02-02T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.237422 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.237459 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.237469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.237484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.237497 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:35Z","lastTransitionTime":"2026-02-02T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.340067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.340103 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.340113 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.340128 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.340139 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:35Z","lastTransitionTime":"2026-02-02T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.410823 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 13:43:57.235987658 +0000 UTC Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.423286 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.423331 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:35 crc kubenswrapper[4846]: E0202 12:10:35.423425 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.423667 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:35 crc kubenswrapper[4846]: E0202 12:10:35.423747 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:35 crc kubenswrapper[4846]: E0202 12:10:35.423868 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.441886 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.441922 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.441935 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.441949 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.441960 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:35Z","lastTransitionTime":"2026-02-02T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.544261 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.544569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.544705 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.544810 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.544891 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:35Z","lastTransitionTime":"2026-02-02T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.647408 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.647458 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.647476 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.647496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.647508 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:35Z","lastTransitionTime":"2026-02-02T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.749386 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.749452 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.749464 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.749483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.749495 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:35Z","lastTransitionTime":"2026-02-02T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.851546 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.851601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.851612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.851653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.851670 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:35Z","lastTransitionTime":"2026-02-02T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.953973 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.954019 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.954034 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.954051 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:35 crc kubenswrapper[4846]: I0202 12:10:35.954063 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:35Z","lastTransitionTime":"2026-02-02T12:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.055616 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.055683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.055695 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.055711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.055722 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.157869 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.157913 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.157924 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.157938 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.157948 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.207500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.207537 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.207547 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.207568 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.207579 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: E0202 12:10:36.219008 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:36Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.330124 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.330157 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.330166 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.330180 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.330189 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: E0202 12:10:36.343048 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:36Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.346396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.346437 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.346450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.346470 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.346482 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: E0202 12:10:36.358484 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:36Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.362265 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.362288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.362298 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.362318 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.362328 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: E0202 12:10:36.374533 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:36Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.377906 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.377947 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.377957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.377974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.377984 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: E0202 12:10:36.388951 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:36Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:36 crc kubenswrapper[4846]: E0202 12:10:36.389106 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.390492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.390526 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.390540 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.390561 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.390573 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.411847 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 09:55:15.085261513 +0000 UTC Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.423141 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:36 crc kubenswrapper[4846]: E0202 12:10:36.423277 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.493152 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.493209 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.493231 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.493257 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.493273 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.596539 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.596609 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.596642 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.596663 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.596675 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.699092 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.699134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.699145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.699161 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.699171 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.801668 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.801708 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.801719 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.801735 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.801746 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.903980 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.904019 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.904029 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.904044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:36 crc kubenswrapper[4846]: I0202 12:10:36.904054 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:36Z","lastTransitionTime":"2026-02-02T12:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.006983 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.007044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.007056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.007076 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.007089 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:37Z","lastTransitionTime":"2026-02-02T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.111583 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.111652 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.111672 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.111702 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.111725 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:37Z","lastTransitionTime":"2026-02-02T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.214350 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.214415 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.214442 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.214472 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.214495 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:37Z","lastTransitionTime":"2026-02-02T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.317799 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.317832 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.317842 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.317897 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.317911 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:37Z","lastTransitionTime":"2026-02-02T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.412068 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 10:59:30.406793981 +0000 UTC Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.419912 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.419960 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.419974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.419990 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.420000 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:37Z","lastTransitionTime":"2026-02-02T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.422410 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:37 crc kubenswrapper[4846]: E0202 12:10:37.422592 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.422902 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:37 crc kubenswrapper[4846]: E0202 12:10:37.423029 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.423391 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:37 crc kubenswrapper[4846]: E0202 12:10:37.423527 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.521883 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.521916 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.521925 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.521939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.521948 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:37Z","lastTransitionTime":"2026-02-02T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.624435 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.624485 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.624496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.624510 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.624523 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:37Z","lastTransitionTime":"2026-02-02T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.727411 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.727440 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.727449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.727480 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.727492 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:37Z","lastTransitionTime":"2026-02-02T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.830359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.830395 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.830405 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.830419 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.830429 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:37Z","lastTransitionTime":"2026-02-02T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.933342 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.933381 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.933392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.933409 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:37 crc kubenswrapper[4846]: I0202 12:10:37.933421 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:37Z","lastTransitionTime":"2026-02-02T12:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.036312 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.036354 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.036364 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.036381 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.036392 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:38Z","lastTransitionTime":"2026-02-02T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.138947 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.138979 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.138988 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.139004 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.139015 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:38Z","lastTransitionTime":"2026-02-02T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.242038 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.242082 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.242093 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.242108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.242120 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:38Z","lastTransitionTime":"2026-02-02T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.344093 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.344131 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.344141 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.344157 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.344169 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:38Z","lastTransitionTime":"2026-02-02T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.412185 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 13:42:37.361223195 +0000 UTC Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.422516 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:38 crc kubenswrapper[4846]: E0202 12:10:38.422684 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.446285 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.446323 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.446332 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.446348 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.446357 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:38Z","lastTransitionTime":"2026-02-02T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.549376 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.549425 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.549436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.549450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.549459 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:38Z","lastTransitionTime":"2026-02-02T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.651850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.651887 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.651895 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.651912 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.651921 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:38Z","lastTransitionTime":"2026-02-02T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.754058 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.754122 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.754144 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.754172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.754192 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:38Z","lastTransitionTime":"2026-02-02T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.856172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.856275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.856308 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.856326 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.856350 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:38Z","lastTransitionTime":"2026-02-02T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.958427 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.958475 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.958489 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.958507 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:38 crc kubenswrapper[4846]: I0202 12:10:38.958520 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:38Z","lastTransitionTime":"2026-02-02T12:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.060515 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.060558 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.060573 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.060594 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.060609 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:39Z","lastTransitionTime":"2026-02-02T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.162956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.162997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.163013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.163032 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.163044 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:39Z","lastTransitionTime":"2026-02-02T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.265067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.265102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.265111 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.265125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.265134 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:39Z","lastTransitionTime":"2026-02-02T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.366714 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.367042 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.367153 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.367228 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.367291 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:39Z","lastTransitionTime":"2026-02-02T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.413196 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 13:21:00.993238053 +0000 UTC Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.422484 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.422618 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:39 crc kubenswrapper[4846]: E0202 12:10:39.422640 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.422731 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:39 crc kubenswrapper[4846]: E0202 12:10:39.422896 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:39 crc kubenswrapper[4846]: E0202 12:10:39.423036 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.435812 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.445153 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.457204 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.469082 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.469108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.469120 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.469138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.469150 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:39Z","lastTransitionTime":"2026-02-02T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.469467 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.481673 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.493643 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.503495 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.513196 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.524380 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.532992 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.545563 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.555981 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.571395 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.571430 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.571440 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.571454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.571463 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:39Z","lastTransitionTime":"2026-02-02T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.582426 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.593884 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.606211 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.618974 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.643033 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.653159 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:39Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.673163 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.673224 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.673235 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.673250 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.673258 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:39Z","lastTransitionTime":"2026-02-02T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.775786 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.775827 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.775839 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.775856 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.775866 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:39Z","lastTransitionTime":"2026-02-02T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.878528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.878562 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.878571 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.878584 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.878593 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:39Z","lastTransitionTime":"2026-02-02T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.981605 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.981674 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.981683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.981698 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:39 crc kubenswrapper[4846]: I0202 12:10:39.981710 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:39Z","lastTransitionTime":"2026-02-02T12:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.083611 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.083673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.083685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.083702 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.083712 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:40Z","lastTransitionTime":"2026-02-02T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.185724 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.185762 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.185772 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.185788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.185798 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:40Z","lastTransitionTime":"2026-02-02T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.273425 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:40 crc kubenswrapper[4846]: E0202 12:10:40.273587 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:40 crc kubenswrapper[4846]: E0202 12:10:40.273744 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs podName:fe216e68-8814-468d-a6f3-fe8721d94eba nodeName:}" failed. No retries permitted until 2026-02-02 12:11:12.273719087 +0000 UTC m=+103.502305980 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs") pod "network-metrics-daemon-j5m9k" (UID: "fe216e68-8814-468d-a6f3-fe8721d94eba") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.288483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.288527 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.288544 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.288569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.288585 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:40Z","lastTransitionTime":"2026-02-02T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.391409 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.391483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.391535 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.391564 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.391588 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:40Z","lastTransitionTime":"2026-02-02T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.413905 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 09:22:43.330339596 +0000 UTC Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.423330 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:40 crc kubenswrapper[4846]: E0202 12:10:40.423531 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.496242 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.496288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.496304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.496325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.496343 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:40Z","lastTransitionTime":"2026-02-02T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.598156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.598214 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.598230 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.598253 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.598281 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:40Z","lastTransitionTime":"2026-02-02T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.700403 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.700439 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.700448 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.700500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.700509 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:40Z","lastTransitionTime":"2026-02-02T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.803127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.803164 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.803175 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.803191 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.803201 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:40Z","lastTransitionTime":"2026-02-02T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.904878 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.904922 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.904930 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.904943 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:40 crc kubenswrapper[4846]: I0202 12:10:40.904951 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:40Z","lastTransitionTime":"2026-02-02T12:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.006641 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.006681 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.006694 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.006709 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.006719 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:41Z","lastTransitionTime":"2026-02-02T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.108990 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.109028 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.109038 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.109084 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.109096 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:41Z","lastTransitionTime":"2026-02-02T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.211601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.211656 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.211667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.211683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.211727 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:41Z","lastTransitionTime":"2026-02-02T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.314703 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.314750 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.314761 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.314785 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.314799 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:41Z","lastTransitionTime":"2026-02-02T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.414425 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 06:35:20.970829618 +0000 UTC Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.417182 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.417222 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.417234 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.417252 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.417264 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:41Z","lastTransitionTime":"2026-02-02T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.422645 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.422698 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:41 crc kubenswrapper[4846]: E0202 12:10:41.422735 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.422779 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:41 crc kubenswrapper[4846]: E0202 12:10:41.422897 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:41 crc kubenswrapper[4846]: E0202 12:10:41.423053 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.519066 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.519107 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.519127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.519147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.519159 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:41Z","lastTransitionTime":"2026-02-02T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.621286 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.621322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.621333 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.621347 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.621356 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:41Z","lastTransitionTime":"2026-02-02T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.723188 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.723240 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.723252 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.723269 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.723601 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:41Z","lastTransitionTime":"2026-02-02T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.825598 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.825657 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.825671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.825689 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.825701 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:41Z","lastTransitionTime":"2026-02-02T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.928041 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.928077 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.928089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.928106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:41 crc kubenswrapper[4846]: I0202 12:10:41.928116 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:41Z","lastTransitionTime":"2026-02-02T12:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.030602 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.030693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.030711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.030737 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.030762 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:42Z","lastTransitionTime":"2026-02-02T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.133877 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.133917 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.133930 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.133950 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.133964 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:42Z","lastTransitionTime":"2026-02-02T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.236417 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.236494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.236506 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.236524 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.236537 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:42Z","lastTransitionTime":"2026-02-02T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.339038 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.339111 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.339143 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.339178 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.339200 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:42Z","lastTransitionTime":"2026-02-02T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.414609 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 05:05:07.430360318 +0000 UTC Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.422942 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:42 crc kubenswrapper[4846]: E0202 12:10:42.423071 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.441930 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.441987 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.441997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.442015 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.442027 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:42Z","lastTransitionTime":"2026-02-02T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.544434 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.544471 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.544482 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.544497 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.544507 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:42Z","lastTransitionTime":"2026-02-02T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.647059 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.647105 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.647114 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.647132 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.647143 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:42Z","lastTransitionTime":"2026-02-02T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.750494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.750536 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.750547 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.750566 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.750577 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:42Z","lastTransitionTime":"2026-02-02T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.852772 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.852837 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.852855 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.852871 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.852881 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:42Z","lastTransitionTime":"2026-02-02T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.859007 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7jshv_24fc18fa-9739-480c-8f43-ab30989164c2/kube-multus/0.log" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.859074 4846 generic.go:334] "Generic (PLEG): container finished" podID="24fc18fa-9739-480c-8f43-ab30989164c2" containerID="de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a" exitCode=1 Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.859111 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7jshv" event={"ID":"24fc18fa-9739-480c-8f43-ab30989164c2","Type":"ContainerDied","Data":"de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.859547 4846 scope.go:117] "RemoveContainer" containerID="de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.880876 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:42Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.892520 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:42Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.920268 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:42Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.936374 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:42Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.953310 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:42Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.955417 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.955450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.955484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.955502 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.955513 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:42Z","lastTransitionTime":"2026-02-02T12:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.972071 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:42Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.984797 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:42Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:42 crc kubenswrapper[4846]: I0202 12:10:42.994717 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:42Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.008911 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.021811 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.033230 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.046059 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.057905 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.057977 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.057991 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.058011 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.058081 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:43Z","lastTransitionTime":"2026-02-02T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.059937 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.073938 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.085178 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.098019 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:42Z\\\",\\\"message\\\":\\\"2026-02-02T12:09:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd\\\\n2026-02-02T12:09:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd to /host/opt/cni/bin/\\\\n2026-02-02T12:09:57Z [verbose] multus-daemon started\\\\n2026-02-02T12:09:57Z [verbose] Readiness Indicator file check\\\\n2026-02-02T12:10:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.109256 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.119364 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.160438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.160478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.160490 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.160512 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.160524 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:43Z","lastTransitionTime":"2026-02-02T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.263278 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.263490 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.263595 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.263715 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.263968 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:43Z","lastTransitionTime":"2026-02-02T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.366516 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.366586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.366607 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.366667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.366690 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:43Z","lastTransitionTime":"2026-02-02T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.415316 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 11:59:56.438085458 +0000 UTC Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.422681 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.422836 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.422980 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:43 crc kubenswrapper[4846]: E0202 12:10:43.422971 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:43 crc kubenswrapper[4846]: E0202 12:10:43.423149 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:43 crc kubenswrapper[4846]: E0202 12:10:43.423205 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.437537 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.468995 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.469077 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.469091 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.469138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.469151 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:43Z","lastTransitionTime":"2026-02-02T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.571798 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.571866 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.571883 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.571910 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.571928 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:43Z","lastTransitionTime":"2026-02-02T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.675124 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.675169 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.675179 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.675196 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.675209 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:43Z","lastTransitionTime":"2026-02-02T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.777775 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.777833 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.777851 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.777874 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.777893 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:43Z","lastTransitionTime":"2026-02-02T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.865286 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7jshv_24fc18fa-9739-480c-8f43-ab30989164c2/kube-multus/0.log" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.865418 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7jshv" event={"ID":"24fc18fa-9739-480c-8f43-ab30989164c2","Type":"ContainerStarted","Data":"abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.881475 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.881543 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.881565 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.881594 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.881648 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:43Z","lastTransitionTime":"2026-02-02T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.882087 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.895310 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.918122 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.933857 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.952688 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.967491 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.984392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.984436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.984451 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.984472 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.984548 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:43Z","lastTransitionTime":"2026-02-02T12:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:43 crc kubenswrapper[4846]: I0202 12:10:43.987717 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:42Z\\\",\\\"message\\\":\\\"2026-02-02T12:09:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd\\\\n2026-02-02T12:09:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd to /host/opt/cni/bin/\\\\n2026-02-02T12:09:57Z [verbose] multus-daemon started\\\\n2026-02-02T12:09:57Z [verbose] Readiness Indicator file check\\\\n2026-02-02T12:10:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.000572 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:43Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.018058 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.035262 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.050148 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.064108 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.077794 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.086422 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.086456 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.086467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.086483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.086493 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:44Z","lastTransitionTime":"2026-02-02T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.092370 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.105909 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.117132 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.127535 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.143851 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.152846 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31c6aa7-167b-4c1a-9830-65ce50827c03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b835d20acf3590c96063ca2b91a8f02cd5a374811972692fef36d29619697d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:44Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.188438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.188475 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.188483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.188499 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.188508 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:44Z","lastTransitionTime":"2026-02-02T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.291333 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.291387 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.291399 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.291419 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.291432 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:44Z","lastTransitionTime":"2026-02-02T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.395549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.395597 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.395610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.395640 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.395659 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:44Z","lastTransitionTime":"2026-02-02T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.416461 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 14:46:07.433374744 +0000 UTC Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.422902 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:44 crc kubenswrapper[4846]: E0202 12:10:44.423039 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.498608 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.498809 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.498831 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.498862 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.498880 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:44Z","lastTransitionTime":"2026-02-02T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.601911 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.601964 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.601978 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.602000 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.602014 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:44Z","lastTransitionTime":"2026-02-02T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.704671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.704715 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.704730 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.704747 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.704759 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:44Z","lastTransitionTime":"2026-02-02T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.807561 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.807610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.807667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.807710 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.807732 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:44Z","lastTransitionTime":"2026-02-02T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.910606 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.910717 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.910741 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.910772 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:44 crc kubenswrapper[4846]: I0202 12:10:44.910796 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:44Z","lastTransitionTime":"2026-02-02T12:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.014710 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.014771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.014790 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.014813 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.014831 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:45Z","lastTransitionTime":"2026-02-02T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.117575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.117689 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.117717 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.117743 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.117764 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:45Z","lastTransitionTime":"2026-02-02T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.220765 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.220818 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.220834 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.220857 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.220874 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:45Z","lastTransitionTime":"2026-02-02T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.323039 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.323095 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.323108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.323127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.323139 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:45Z","lastTransitionTime":"2026-02-02T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.416596 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 09:34:44.598409601 +0000 UTC Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.423146 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.423174 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:45 crc kubenswrapper[4846]: E0202 12:10:45.423369 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.423452 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:45 crc kubenswrapper[4846]: E0202 12:10:45.423685 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:45 crc kubenswrapper[4846]: E0202 12:10:45.423902 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.424999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.425226 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.425446 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.425735 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.425929 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:45Z","lastTransitionTime":"2026-02-02T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.528845 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.528888 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.528900 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.528919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.528930 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:45Z","lastTransitionTime":"2026-02-02T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.631539 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.631606 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.631670 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.631706 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.631732 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:45Z","lastTransitionTime":"2026-02-02T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.734597 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.734689 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.734707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.734730 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.734750 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:45Z","lastTransitionTime":"2026-02-02T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.837173 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.837233 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.837253 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.837291 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.837309 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:45Z","lastTransitionTime":"2026-02-02T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.940454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.940528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.940550 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.940579 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:45 crc kubenswrapper[4846]: I0202 12:10:45.940599 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:45Z","lastTransitionTime":"2026-02-02T12:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.042576 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.042688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.042716 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.042746 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.042771 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.144565 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.144608 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.144637 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.144653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.144662 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.246503 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.246554 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.246596 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.246615 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.246652 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.348982 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.349021 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.349036 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.349052 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.349063 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.417040 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 18:58:14.881265874 +0000 UTC Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.423380 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:46 crc kubenswrapper[4846]: E0202 12:10:46.423544 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.451789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.451829 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.451839 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.451856 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.451869 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.554407 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.554446 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.554456 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.554472 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.554481 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.656462 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.656523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.656533 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.656548 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.656557 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.667871 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.667949 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.667971 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.668003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.668025 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: E0202 12:10:46.681992 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:46Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.686977 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.687023 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.687034 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.687052 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.687064 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: E0202 12:10:46.699742 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:46Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.703536 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.703557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.703565 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.703580 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.703589 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: E0202 12:10:46.719697 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:46Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.723653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.723703 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.723713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.723730 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.723741 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: E0202 12:10:46.735754 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:46Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.738950 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.738983 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.738993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.739007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.739016 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: E0202 12:10:46.750939 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:46Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:46 crc kubenswrapper[4846]: E0202 12:10:46.751108 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.758372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.758405 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.758417 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.758433 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.758443 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.860919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.860971 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.860983 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.861002 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.861014 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.964014 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.964062 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.964078 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.964095 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:46 crc kubenswrapper[4846]: I0202 12:10:46.964106 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:46Z","lastTransitionTime":"2026-02-02T12:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.066532 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.066587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.066603 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.066668 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.066686 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:47Z","lastTransitionTime":"2026-02-02T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.169555 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.169602 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.169615 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.169661 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.169673 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:47Z","lastTransitionTime":"2026-02-02T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.273179 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.273253 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.273270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.273295 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.273315 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:47Z","lastTransitionTime":"2026-02-02T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.376967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.377199 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.377312 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.377426 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.377565 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:47Z","lastTransitionTime":"2026-02-02T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.417445 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 03:49:49.033813843 +0000 UTC Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.422918 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.422977 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:47 crc kubenswrapper[4846]: E0202 12:10:47.423035 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.423104 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:47 crc kubenswrapper[4846]: E0202 12:10:47.423323 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:47 crc kubenswrapper[4846]: E0202 12:10:47.423445 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.480643 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.480922 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.481237 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.481334 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.481427 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:47Z","lastTransitionTime":"2026-02-02T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.584784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.584835 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.584850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.584871 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.584886 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:47Z","lastTransitionTime":"2026-02-02T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.688033 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.688376 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.688532 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.688683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.688787 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:47Z","lastTransitionTime":"2026-02-02T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.791418 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.791477 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.791501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.791523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.791537 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:47Z","lastTransitionTime":"2026-02-02T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.894377 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.894422 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.894430 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.894446 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.894455 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:47Z","lastTransitionTime":"2026-02-02T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.997426 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.997492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.997508 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.997534 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:47 crc kubenswrapper[4846]: I0202 12:10:47.997551 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:47Z","lastTransitionTime":"2026-02-02T12:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.102947 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.103027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.103066 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.103101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.103122 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:48Z","lastTransitionTime":"2026-02-02T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.205408 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.205438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.205448 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.205464 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.205475 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:48Z","lastTransitionTime":"2026-02-02T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.308550 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.308661 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.308688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.308717 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.308739 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:48Z","lastTransitionTime":"2026-02-02T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.410659 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.410708 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.410717 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.410733 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.410743 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:48Z","lastTransitionTime":"2026-02-02T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.418063 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 01:34:31.603532096 +0000 UTC Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.423559 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:48 crc kubenswrapper[4846]: E0202 12:10:48.423871 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.513978 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.514051 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.514074 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.514106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.514131 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:48Z","lastTransitionTime":"2026-02-02T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.617218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.617285 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.617307 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.617335 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.617358 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:48Z","lastTransitionTime":"2026-02-02T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.720515 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.720588 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.720612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.720678 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.720703 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:48Z","lastTransitionTime":"2026-02-02T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.823291 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.823575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.823700 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.823784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.823879 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:48Z","lastTransitionTime":"2026-02-02T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.926348 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.926402 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.926420 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.926445 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:48 crc kubenswrapper[4846]: I0202 12:10:48.926464 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:48Z","lastTransitionTime":"2026-02-02T12:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.030022 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.030086 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.030104 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.030145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.030162 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:49Z","lastTransitionTime":"2026-02-02T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.132968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.133013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.133025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.133044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.133059 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:49Z","lastTransitionTime":"2026-02-02T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.235830 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.236171 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.236281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.236378 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.236477 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:49Z","lastTransitionTime":"2026-02-02T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.338993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.339229 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.339295 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.339373 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.339430 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:49Z","lastTransitionTime":"2026-02-02T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.418736 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 08:03:53.530455344 +0000 UTC Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.423150 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:49 crc kubenswrapper[4846]: E0202 12:10:49.423321 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.423712 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:49 crc kubenswrapper[4846]: E0202 12:10:49.424154 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.424212 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:49 crc kubenswrapper[4846]: E0202 12:10:49.424389 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.424592 4846 scope.go:117] "RemoveContainer" containerID="91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.442371 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.442901 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.442985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.443008 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.443035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.443053 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:49Z","lastTransitionTime":"2026-02-02T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.483045 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.498017 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31c6aa7-167b-4c1a-9830-65ce50827c03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b835d20acf3590c96063ca2b91a8f02cd5a374811972692fef36d29619697d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.520034 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.533187 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.545830 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.545865 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.545875 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.545892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.545904 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:49Z","lastTransitionTime":"2026-02-02T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.546036 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.561174 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.570816 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.584238 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.597946 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.614260 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.627601 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.642880 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.648164 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.648203 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.648212 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.648232 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.648244 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:49Z","lastTransitionTime":"2026-02-02T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.656762 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.668092 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.685299 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:42Z\\\",\\\"message\\\":\\\"2026-02-02T12:09:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd\\\\n2026-02-02T12:09:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd to /host/opt/cni/bin/\\\\n2026-02-02T12:09:57Z [verbose] multus-daemon started\\\\n2026-02-02T12:09:57Z [verbose] Readiness Indicator file check\\\\n2026-02-02T12:10:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.697153 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.710764 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.721224 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.751537 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.751578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.751588 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.751609 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.751636 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:49Z","lastTransitionTime":"2026-02-02T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.854273 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.854318 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.854326 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.854343 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.854354 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:49Z","lastTransitionTime":"2026-02-02T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.890087 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/2.log" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.892543 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.893129 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.910134 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.919749 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.932264 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:42Z\\\",\\\"message\\\":\\\"2026-02-02T12:09:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd\\\\n2026-02-02T12:09:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd to /host/opt/cni/bin/\\\\n2026-02-02T12:09:57Z [verbose] multus-daemon started\\\\n2026-02-02T12:09:57Z [verbose] Readiness Indicator file check\\\\n2026-02-02T12:10:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.945082 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.957084 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.957116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.957127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.957148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.957165 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:49Z","lastTransitionTime":"2026-02-02T12:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.962130 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.984084 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:49 crc kubenswrapper[4846]: I0202 12:10:49.995727 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:49Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.015677 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.034306 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31c6aa7-167b-4c1a-9830-65ce50827c03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b835d20acf3590c96063ca2b91a8f02cd5a374811972692fef36d29619697d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.053543 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.059345 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.059398 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.059410 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.059431 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.059443 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:50Z","lastTransitionTime":"2026-02-02T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.069380 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.082982 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.093581 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.111222 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.122473 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.138683 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.151864 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.161934 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.161996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.162046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.162075 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.162091 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:50Z","lastTransitionTime":"2026-02-02T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.213332 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.232322 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.264895 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.264938 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.264949 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.264965 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.264974 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:50Z","lastTransitionTime":"2026-02-02T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.367972 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.368011 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.368022 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.368038 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.368050 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:50Z","lastTransitionTime":"2026-02-02T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.419911 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 09:20:04.286100118 +0000 UTC Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.423221 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:50 crc kubenswrapper[4846]: E0202 12:10:50.423367 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.470929 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.470991 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.471003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.471027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.471042 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:50Z","lastTransitionTime":"2026-02-02T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.573752 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.573795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.573803 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.573819 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.573828 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:50Z","lastTransitionTime":"2026-02-02T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.677098 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.677160 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.677182 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.677206 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.677223 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:50Z","lastTransitionTime":"2026-02-02T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.779950 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.779995 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.780007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.780025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.780036 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:50Z","lastTransitionTime":"2026-02-02T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.882899 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.882939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.882948 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.882964 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.882977 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:50Z","lastTransitionTime":"2026-02-02T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.899094 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/3.log" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.899942 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/2.log" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.903109 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" exitCode=1 Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.903149 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.903192 4846 scope.go:117] "RemoveContainer" containerID="91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.903793 4846 scope.go:117] "RemoveContainer" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:10:50 crc kubenswrapper[4846]: E0202 12:10:50.903955 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.921989 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.932716 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.941649 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.962040 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.971958 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31c6aa7-167b-4c1a-9830-65ce50827c03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b835d20acf3590c96063ca2b91a8f02cd5a374811972692fef36d29619697d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.985726 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.985765 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.985773 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.985787 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.985798 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:50Z","lastTransitionTime":"2026-02-02T12:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:50 crc kubenswrapper[4846]: I0202 12:10:50.986093 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.000946 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.018881 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91aaa4d6115f44d7b54fb169af16641a074c7a5c9fcac59f0807f7e91f9ac1ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:20Z\\\",\\\"message\\\":\\\"achine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.183\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0202 12:10:20.212062 6505 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 12:10:20.2121\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:50Z\\\",\\\"message\\\":\\\" 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z]\\\\nI0202 12:10:50.514745 6930 services_controller.go:445] Built service openshift-marketplace/redhat-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nI0202 12:10:50.514732 6930 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-config-operator/metrics]} name:Service_openshift-config-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.161:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f32857b5-f652-4313-a0d7-455c3156dd99}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.030736 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.042197 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.055285 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.066751 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:42Z\\\",\\\"message\\\":\\\"2026-02-02T12:09:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd\\\\n2026-02-02T12:09:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd to /host/opt/cni/bin/\\\\n2026-02-02T12:09:57Z [verbose] multus-daemon started\\\\n2026-02-02T12:09:57Z [verbose] Readiness Indicator file check\\\\n2026-02-02T12:10:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.078412 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.088044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.088086 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.088098 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.088115 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.088126 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:51Z","lastTransitionTime":"2026-02-02T12:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.094584 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.108860 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.123076 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.133040 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.143423 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.155385 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.190967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.191010 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.191022 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.191041 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.191053 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:51Z","lastTransitionTime":"2026-02-02T12:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.295012 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.295075 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.295088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.295105 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.295122 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:51Z","lastTransitionTime":"2026-02-02T12:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.398401 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.398734 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.398953 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.399124 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.399241 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:51Z","lastTransitionTime":"2026-02-02T12:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.420832 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 14:01:05.5943113 +0000 UTC Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.423235 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.423272 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:51 crc kubenswrapper[4846]: E0202 12:10:51.423353 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.423453 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:51 crc kubenswrapper[4846]: E0202 12:10:51.423654 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:51 crc kubenswrapper[4846]: E0202 12:10:51.423695 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.501149 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.501193 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.501228 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.501246 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.501257 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:51Z","lastTransitionTime":"2026-02-02T12:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.604114 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.604160 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.604172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.604189 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.604201 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:51Z","lastTransitionTime":"2026-02-02T12:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.707089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.707146 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.707162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.707186 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.707204 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:51Z","lastTransitionTime":"2026-02-02T12:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.810514 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.810569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.810586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.810607 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.810664 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:51Z","lastTransitionTime":"2026-02-02T12:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.907554 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/3.log" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.910664 4846 scope.go:117] "RemoveContainer" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:10:51 crc kubenswrapper[4846]: E0202 12:10:51.910798 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.912254 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.912348 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.912358 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.912377 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.912388 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:51Z","lastTransitionTime":"2026-02-02T12:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.925442 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.937358 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.953890 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.970592 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:51 crc kubenswrapper[4846]: I0202 12:10:51.996979 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:50Z\\\",\\\"message\\\":\\\" 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z]\\\\nI0202 12:10:50.514745 6930 services_controller.go:445] Built service openshift-marketplace/redhat-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nI0202 12:10:50.514732 6930 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-config-operator/metrics]} name:Service_openshift-config-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.161:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f32857b5-f652-4313-a0d7-455c3156dd99}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:51Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.015409 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.015453 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.015465 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.015483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.015495 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:52Z","lastTransitionTime":"2026-02-02T12:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.016242 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.030035 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.042324 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.052723 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.065952 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:42Z\\\",\\\"message\\\":\\\"2026-02-02T12:09:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd\\\\n2026-02-02T12:09:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd to /host/opt/cni/bin/\\\\n2026-02-02T12:09:57Z [verbose] multus-daemon started\\\\n2026-02-02T12:09:57Z [verbose] Readiness Indicator file check\\\\n2026-02-02T12:10:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.077428 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.089018 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.098234 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.109713 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.118231 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.118311 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.118325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.118346 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.118359 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:52Z","lastTransitionTime":"2026-02-02T12:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.129252 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.139755 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31c6aa7-167b-4c1a-9830-65ce50827c03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b835d20acf3590c96063ca2b91a8f02cd5a374811972692fef36d29619697d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.152435 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.164918 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.177356 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:52Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.220135 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.220179 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.220189 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.220205 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.220214 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:52Z","lastTransitionTime":"2026-02-02T12:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.323063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.323114 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.323125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.323144 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.323156 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:52Z","lastTransitionTime":"2026-02-02T12:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.421490 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 16:56:34.837555272 +0000 UTC Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.422695 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:52 crc kubenswrapper[4846]: E0202 12:10:52.422814 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.425080 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.425103 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.425111 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.425121 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.425142 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:52Z","lastTransitionTime":"2026-02-02T12:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.527249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.527284 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.527293 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.527307 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.527315 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:52Z","lastTransitionTime":"2026-02-02T12:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.629513 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.629573 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.629590 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.629613 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.629659 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:52Z","lastTransitionTime":"2026-02-02T12:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.732905 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.732953 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.732964 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.732981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.732992 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:52Z","lastTransitionTime":"2026-02-02T12:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.836238 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.836307 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.836326 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.836352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.836369 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:52Z","lastTransitionTime":"2026-02-02T12:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.938475 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.938523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.938534 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.938552 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:52 crc kubenswrapper[4846]: I0202 12:10:52.938565 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:52Z","lastTransitionTime":"2026-02-02T12:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.041265 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.041327 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.041339 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.041359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.041372 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:53Z","lastTransitionTime":"2026-02-02T12:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.144653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.144710 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.144721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.144739 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.144750 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:53Z","lastTransitionTime":"2026-02-02T12:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.248035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.248088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.248099 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.248121 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.248142 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:53Z","lastTransitionTime":"2026-02-02T12:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.316672 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.316854 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:11:57.316822912 +0000 UTC m=+148.545409775 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.316891 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.316927 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.316984 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.317096 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.317147 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.317169 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.317190 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.317144 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.317171 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:11:57.317156901 +0000 UTC m=+148.545743764 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.317255 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 12:11:57.317244103 +0000 UTC m=+148.545830966 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.317346 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 12:11:57.317310335 +0000 UTC m=+148.545897198 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.351178 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.351245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.351262 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.351287 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.351307 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:53Z","lastTransitionTime":"2026-02-02T12:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.418458 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.418592 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.418641 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.418657 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.418717 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 12:11:57.418699611 +0000 UTC m=+148.647286474 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.422390 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 13:38:42.380359597 +0000 UTC Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.422489 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.422504 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.422584 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.422689 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.422759 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:53 crc kubenswrapper[4846]: E0202 12:10:53.422832 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.453789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.453849 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.453941 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.453970 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.453992 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:53Z","lastTransitionTime":"2026-02-02T12:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.556669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.556727 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.556746 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.556775 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.556793 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:53Z","lastTransitionTime":"2026-02-02T12:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.659489 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.659519 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.659527 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.659544 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.659562 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:53Z","lastTransitionTime":"2026-02-02T12:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.761254 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.761284 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.761292 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.761304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.761313 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:53Z","lastTransitionTime":"2026-02-02T12:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.864225 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.864250 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.864257 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.864271 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.864280 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:53Z","lastTransitionTime":"2026-02-02T12:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.967795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.967826 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.967834 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.967850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:53 crc kubenswrapper[4846]: I0202 12:10:53.967859 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:53Z","lastTransitionTime":"2026-02-02T12:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.071581 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.071780 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.071802 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.071827 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.071841 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:54Z","lastTransitionTime":"2026-02-02T12:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.175468 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.175554 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.175567 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.175588 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.175601 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:54Z","lastTransitionTime":"2026-02-02T12:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.278469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.278552 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.278570 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.278597 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.278613 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:54Z","lastTransitionTime":"2026-02-02T12:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.381781 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.381819 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.381831 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.381847 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.381859 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:54Z","lastTransitionTime":"2026-02-02T12:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.422660 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 20:32:27.188348283 +0000 UTC Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.422762 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:54 crc kubenswrapper[4846]: E0202 12:10:54.422914 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.484049 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.484118 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.484144 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.484176 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.484202 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:54Z","lastTransitionTime":"2026-02-02T12:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.585972 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.586011 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.586019 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.586036 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.586045 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:54Z","lastTransitionTime":"2026-02-02T12:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.688707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.688757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.688767 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.688786 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.688804 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:54Z","lastTransitionTime":"2026-02-02T12:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.792610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.792683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.792693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.792710 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.792725 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:54Z","lastTransitionTime":"2026-02-02T12:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.895479 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.895519 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.895531 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.895547 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:54 crc kubenswrapper[4846]: I0202 12:10:54.895560 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:54Z","lastTransitionTime":"2026-02-02T12:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.002452 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.002511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.002532 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.002609 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.002671 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:55Z","lastTransitionTime":"2026-02-02T12:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.105978 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.106386 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.106396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.106414 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.106424 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:55Z","lastTransitionTime":"2026-02-02T12:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.209650 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.209700 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.209709 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.209728 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.209740 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:55Z","lastTransitionTime":"2026-02-02T12:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.312398 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.312467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.312493 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.312524 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.312546 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:55Z","lastTransitionTime":"2026-02-02T12:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.416135 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.416182 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.416193 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.416212 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.416223 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:55Z","lastTransitionTime":"2026-02-02T12:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.422515 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.422609 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.422669 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:55 crc kubenswrapper[4846]: E0202 12:10:55.422799 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.422798 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 21:09:26.42904608 +0000 UTC Feb 02 12:10:55 crc kubenswrapper[4846]: E0202 12:10:55.422994 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:55 crc kubenswrapper[4846]: E0202 12:10:55.423186 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.519812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.519853 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.519870 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.519898 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.519917 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:55Z","lastTransitionTime":"2026-02-02T12:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.622724 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.622762 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.622771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.622786 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.622794 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:55Z","lastTransitionTime":"2026-02-02T12:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.724679 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.724738 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.724751 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.724770 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.724782 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:55Z","lastTransitionTime":"2026-02-02T12:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.827460 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.827502 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.827513 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.827528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.827539 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:55Z","lastTransitionTime":"2026-02-02T12:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.929574 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.929604 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.929612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.929652 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:55 crc kubenswrapper[4846]: I0202 12:10:55.929668 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:55Z","lastTransitionTime":"2026-02-02T12:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.032586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.032711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.032737 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.032770 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.032792 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.136219 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.136290 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.136314 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.136344 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.136367 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.240048 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.240102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.240114 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.240131 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.240143 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.343139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.343188 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.343205 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.343228 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.343245 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.423217 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 11:15:59.710353815 +0000 UTC Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.423363 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:56 crc kubenswrapper[4846]: E0202 12:10:56.423597 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.446739 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.446785 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.446796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.446812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.446825 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.549461 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.549506 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.549516 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.549532 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.549544 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.652411 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.652467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.652478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.652497 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.652512 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.756038 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.756093 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.756259 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.756358 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.756393 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.859549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.859594 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.859604 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.859640 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.859672 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.887291 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.887339 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.887353 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.887371 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.887381 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: E0202 12:10:56.902216 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.906213 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.906272 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.906325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.906351 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.906419 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: E0202 12:10:56.921251 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.925234 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.925267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.925278 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.925296 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.925307 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: E0202 12:10:56.942932 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.946788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.946856 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.946869 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.946913 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.946929 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: E0202 12:10:56.959837 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.964090 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.964142 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.964154 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.964173 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.964184 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:56 crc kubenswrapper[4846]: E0202 12:10:56.977336 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"90d0bd8e-8fff-4676-a040-40a8eb821a29\\\",\\\"systemUUID\\\":\\\"292d225e-9d96-4eab-9003-afa615d76032\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:56Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:56 crc kubenswrapper[4846]: E0202 12:10:56.977553 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.978773 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.978832 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.978855 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.978885 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:56 crc kubenswrapper[4846]: I0202 12:10:56.978908 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:56Z","lastTransitionTime":"2026-02-02T12:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.081329 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.081371 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.081385 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.081404 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.081413 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:57Z","lastTransitionTime":"2026-02-02T12:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.183315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.183349 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.183363 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.183378 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.183388 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:57Z","lastTransitionTime":"2026-02-02T12:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.285595 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.285655 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.285668 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.285685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.285696 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:57Z","lastTransitionTime":"2026-02-02T12:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.387917 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.387944 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.387953 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.387968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.387977 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:57Z","lastTransitionTime":"2026-02-02T12:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.423341 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.423335 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 16:26:52.237547909 +0000 UTC Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.423376 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:57 crc kubenswrapper[4846]: E0202 12:10:57.424079 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:57 crc kubenswrapper[4846]: E0202 12:10:57.424114 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.423414 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:57 crc kubenswrapper[4846]: E0202 12:10:57.424167 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.491479 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.491520 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.491531 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.491548 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.491559 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:57Z","lastTransitionTime":"2026-02-02T12:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.593834 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.593880 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.593889 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.593902 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.593911 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:57Z","lastTransitionTime":"2026-02-02T12:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.696780 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.696830 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.696842 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.696860 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.696874 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:57Z","lastTransitionTime":"2026-02-02T12:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.800184 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.800258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.800268 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.800284 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.800295 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:57Z","lastTransitionTime":"2026-02-02T12:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.902193 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.902238 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.902249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.902267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:57 crc kubenswrapper[4846]: I0202 12:10:57.902280 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:57Z","lastTransitionTime":"2026-02-02T12:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.004402 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.004553 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.004652 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.004752 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.004883 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:58Z","lastTransitionTime":"2026-02-02T12:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.107559 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.108103 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.108219 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.108340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.108457 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:58Z","lastTransitionTime":"2026-02-02T12:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.211522 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.211574 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.211594 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.211622 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.211655 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:58Z","lastTransitionTime":"2026-02-02T12:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.313571 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.313644 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.313656 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.313673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.313683 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:58Z","lastTransitionTime":"2026-02-02T12:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.415995 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.416030 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.416041 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.416058 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.416071 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:58Z","lastTransitionTime":"2026-02-02T12:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.422612 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:10:58 crc kubenswrapper[4846]: E0202 12:10:58.422776 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.424718 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 11:42:38.380741425 +0000 UTC Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.520064 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.520117 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.520167 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.520190 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.520205 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:58Z","lastTransitionTime":"2026-02-02T12:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.622249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.622283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.622293 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.622307 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.622316 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:58Z","lastTransitionTime":"2026-02-02T12:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.724573 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.724658 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.724671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.724691 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.724702 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:58Z","lastTransitionTime":"2026-02-02T12:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.827019 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.827048 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.827057 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.827070 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.827078 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:58Z","lastTransitionTime":"2026-02-02T12:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.930553 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.930657 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.930678 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.930701 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:58 crc kubenswrapper[4846]: I0202 12:10:58.930718 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:58Z","lastTransitionTime":"2026-02-02T12:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.033369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.033439 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.033454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.033476 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.033490 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:59Z","lastTransitionTime":"2026-02-02T12:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.135966 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.136023 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.136037 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.136057 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.136074 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:59Z","lastTransitionTime":"2026-02-02T12:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.239168 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.239232 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.239243 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.239261 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.239273 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:59Z","lastTransitionTime":"2026-02-02T12:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.342295 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.342352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.342379 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.342400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.342416 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:59Z","lastTransitionTime":"2026-02-02T12:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.422966 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.423057 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.422966 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:10:59 crc kubenswrapper[4846]: E0202 12:10:59.423143 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:10:59 crc kubenswrapper[4846]: E0202 12:10:59.423215 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:10:59 crc kubenswrapper[4846]: E0202 12:10:59.423303 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.424795 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 19:19:48.687958314 +0000 UTC Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.444896 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.444973 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.444985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.445012 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.445029 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:59Z","lastTransitionTime":"2026-02-02T12:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.474132 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.489330 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd4669820f87f286f3f673734d1b18f6606451f8ef83633a001192dee49e2a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.505800 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e681c0d1-0580-45d2-a381-540fded41685\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11fa69930a35c583fc670d8fbbb3a6d37f4a9215b5f91b0db330799e0b54ff89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3909b23ec010d84c2c6ea95551b944570ddb72c9e4290402c99beec6ca5151db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhr2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cscmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.518109 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe216e68-8814-468d-a6f3-fe8721d94eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zhqv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:10:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-j5m9k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.536240 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6395d48b-61a0-49b6-88d1-fac57fe4f3c1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://577adf49ec6f587180334b744acc8ff4bdfd09f9fd8fbb98df3217de65beac85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b2597011635ed77fae5245e08159cb79c5501df27e5ff4a9a03c2b38cc7c067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa0784794a3c6467302da083b8225b63cfd4f25d5f371101b3ebc4686cc66d86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeafc13c7a592b8d6d79ad724ce36d8f2430b88deecb00a9c1a4b91d145faa3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f88f17b3d209229121bcb4228de5affe7515d8ec428c3f7281ef75c6a1c6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eacebe26fe8b15a78c25998705b52b72f87bff16d99eec521939e2d6a3ca1752\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ba5f8e125e18a540d9cebfd05ac010eb04f2863612f67909c0af0898a092c58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://282b8b391dfba765acca3d487125f3b80cbdb2049c52f5341dfaaaa10f1defc2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.546737 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31c6aa7-167b-4c1a-9830-65ce50827c03\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b835d20acf3590c96063ca2b91a8f02cd5a374811972692fef36d29619697d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3680e252bf5f959c651f5e22dc90e801cab864052a31a8e40fb6c37127b7cbde\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.548619 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.548697 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.548709 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.548726 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.548745 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:59Z","lastTransitionTime":"2026-02-02T12:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.558763 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5062e12d3a217e234c5f7def30f0ae7ea8b5a26f70676c97653e78ef6a2b100d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.571923 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.589006 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:50Z\\\",\\\"message\\\":\\\" 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:50Z is after 2025-08-24T17:21:41Z]\\\\nI0202 12:10:50.514745 6930 services_controller.go:445] Built service openshift-marketplace/redhat-operators LB template configs for network=default: []services.lbConfig(nil)\\\\nI0202 12:10:50.514732 6930 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-config-operator/metrics]} name:Service_openshift-config-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.161:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f32857b5-f652-4313-a0d7-455c3156dd99}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:10:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-87zlr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ftvcw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.600322 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03fa522e-6321-4e78-a6ae-6ce53c63646b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cdafa904f6403344ffd922f686d4f8ad0eca2ddea0669bc877c31edfb73b271e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e78cb1e0c691e33cc436d31d03dae8b2a93d10e369293cbd70ea6617cf1408ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d70552ec34017a74407eb284f380acc0838e59e580836a79cb913b9af57b3544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.610180 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbab842b-f87a-4066-bb7f-194167196c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6354798ab237ac1ed731bffbbb9395dfb0533afe03cfaffaebcdd913e5e2a1a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6be81b2d78f489f367aa2a5579dd2e176dcf71a8baff5edd2decd52110385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12dfee247c22953d6ca8bbef626859b2c76f21ba657353d376678dbb2fc97a89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0fd7cf05e0fd897782f18a098023b6a8d3045fe18bb25906caf01666184c5d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.619917 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-44zpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40f86b56-873f-4c8a-81e2-abc5643f4869\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b948a674e830f3628553e5c9d4dd6533ef179627187b074dc2e33907beb74019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whqc6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-44zpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.630312 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7jshv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24fc18fa-9739-480c-8f43-ab30989164c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T12:10:42Z\\\",\\\"message\\\":\\\"2026-02-02T12:09:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd\\\\n2026-02-02T12:09:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ceecdeb4-af93-46a4-9d3c-53b9aaa95bbd to /host/opt/cni/bin/\\\\n2026-02-02T12:09:57Z [verbose] multus-daemon started\\\\n2026-02-02T12:09:57Z [verbose] Readiness Indicator file check\\\\n2026-02-02T12:10:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jns97\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7jshv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.639541 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05ed703a-a310-418e-8c5d-aeddb32977cf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a559173da5fb1872ffbd4fab2da4eed62df15ff608c76af92cf031d25465271b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dhzp7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.651332 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.651366 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.651374 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.651390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.651398 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:59Z","lastTransitionTime":"2026-02-02T12:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.653100 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebd8eb79-1502-48f4-98e1-45c365ac5638\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a10484f66894e95e83496ee958eae3d86f1e0e21dc113b40477c34cedbdb75c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cad3339365f29be8168f7c080f723c7283b4fc924bef8132de3246d83bae164c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c73690bd7cdc7e1ec7811eeb17982ab7b8a8528a7ff412c8f7eb47485f97fb3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a51c21c2bbd79d2445f607cb3df4c55ceb4e46e048e725ba1e6c2475cf228021\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7d3263b68920cfdbcf8340d3f626fe9bf5690e46d3cbf5731bc58a7f52021d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41adf976474680e9b6c6089d200f907c2231c9fbb2d98de76f98b8a1b0d66701\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa6f21f5bf9078ed24e9ac9de56b1f2b5823fd230953bccd8051664abcc7bd87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvl6z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6jrkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.664844 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61dd041-dcc8-4102-a956-9e7d1a4ce0b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:10:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T12:09:48Z\\\",\\\"message\\\":\\\"file observer\\\\nW0202 12:09:48.417930 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0202 12:09:48.418157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 12:09:48.419995 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-431033122/tls.crt::/tmp/serving-cert-431033122/tls.key\\\\\\\"\\\\nI0202 12:09:48.669501 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 12:09:48.673511 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 12:09:48.673584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 12:09:48.673650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 12:09:48.673685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 12:09:48.679613 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 12:09:48.679659 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679665 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 12:09:48.679669 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 12:09:48.679672 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 12:09:48.679675 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 12:09:48.679677 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 12:09:48.679745 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 12:09:48.681966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T12:09:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T12:09:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.677828 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45e02b40feb9c6acd510650ebe76f16eefa7493409cf29368a8d4b1c7220c430\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://915086fa3e1c5d2bff9cef7b595258826b813b9133a3d2b0ffa89447bd1c81a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.688996 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.700525 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-9vvp2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5aa2f93-ed54-4835-ad19-8b63707975d9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T12:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f423d71af5c09c3384711f8fa795fdbe2ba6cbc0e1d2ac5816bde55889024392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T12:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xt27x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T12:09:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-9vvp2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T12:10:59Z is after 2025-08-24T17:21:41Z" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.754145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.754188 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.754200 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.754217 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.754229 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:59Z","lastTransitionTime":"2026-02-02T12:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.859079 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.859136 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.859153 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.859177 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.859195 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:59Z","lastTransitionTime":"2026-02-02T12:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.962033 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.962071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.962081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.962097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:10:59 crc kubenswrapper[4846]: I0202 12:10:59.962109 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:10:59Z","lastTransitionTime":"2026-02-02T12:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.064248 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.064304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.064324 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.064347 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.064364 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:00Z","lastTransitionTime":"2026-02-02T12:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.167687 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.167740 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.167762 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.167795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.167817 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:00Z","lastTransitionTime":"2026-02-02T12:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.270886 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.270921 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.270932 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.270946 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.270957 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:00Z","lastTransitionTime":"2026-02-02T12:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.373593 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.373859 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.373992 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.374092 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.374184 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:00Z","lastTransitionTime":"2026-02-02T12:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.422877 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:00 crc kubenswrapper[4846]: E0202 12:11:00.423042 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.425912 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 01:08:22.649137563 +0000 UTC Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.477061 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.477106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.477119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.477137 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.477150 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:00Z","lastTransitionTime":"2026-02-02T12:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.579250 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.579538 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.579547 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.579561 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.579570 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:00Z","lastTransitionTime":"2026-02-02T12:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.682008 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.682059 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.682069 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.682086 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.682097 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:00Z","lastTransitionTime":"2026-02-02T12:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.784299 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.784376 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.784395 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.784423 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.784442 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:00Z","lastTransitionTime":"2026-02-02T12:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.886524 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.886575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.886587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.886605 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.886616 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:00Z","lastTransitionTime":"2026-02-02T12:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.989436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.989473 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.989484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.989500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:00 crc kubenswrapper[4846]: I0202 12:11:00.989510 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:00Z","lastTransitionTime":"2026-02-02T12:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.091906 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.091957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.091968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.091984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.091997 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:01Z","lastTransitionTime":"2026-02-02T12:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.195044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.195106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.195125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.195147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.195165 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:01Z","lastTransitionTime":"2026-02-02T12:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.297598 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.297662 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.297676 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.297694 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.297707 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:01Z","lastTransitionTime":"2026-02-02T12:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.400582 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.400685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.400705 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.400727 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.400745 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:01Z","lastTransitionTime":"2026-02-02T12:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.423188 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:01 crc kubenswrapper[4846]: E0202 12:11:01.423365 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.423431 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:01 crc kubenswrapper[4846]: E0202 12:11:01.423511 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.423210 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:01 crc kubenswrapper[4846]: E0202 12:11:01.423585 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.426044 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 10:56:16.513203405 +0000 UTC Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.503282 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.503340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.503365 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.503395 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.503418 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:01Z","lastTransitionTime":"2026-02-02T12:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.606499 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.606567 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.606586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.606609 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.606669 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:01Z","lastTransitionTime":"2026-02-02T12:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.708999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.709040 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.709051 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.709068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.709080 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:01Z","lastTransitionTime":"2026-02-02T12:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.811795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.811851 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.811866 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.811889 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.811904 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:01Z","lastTransitionTime":"2026-02-02T12:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.914803 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.914838 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.914849 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.914865 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:01 crc kubenswrapper[4846]: I0202 12:11:01.914876 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:01Z","lastTransitionTime":"2026-02-02T12:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.016475 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.016504 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.016511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.016523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.016532 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:02Z","lastTransitionTime":"2026-02-02T12:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.118614 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.118662 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.118683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.118701 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.118713 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:02Z","lastTransitionTime":"2026-02-02T12:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.220801 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.220858 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.220880 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.220921 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.220944 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:02Z","lastTransitionTime":"2026-02-02T12:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.323690 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.323734 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.323747 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.323764 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.323775 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:02Z","lastTransitionTime":"2026-02-02T12:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.422972 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:02 crc kubenswrapper[4846]: E0202 12:11:02.423161 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.425506 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.425556 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.425571 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.425587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.425599 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:02Z","lastTransitionTime":"2026-02-02T12:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.426868 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 04:52:34.17886265 +0000 UTC Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.528169 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.528202 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.528210 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.528223 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.528231 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:02Z","lastTransitionTime":"2026-02-02T12:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.630610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.630692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.630704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.630721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.630734 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:02Z","lastTransitionTime":"2026-02-02T12:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.733154 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.733184 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.733192 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.733205 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.733214 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:02Z","lastTransitionTime":"2026-02-02T12:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.835370 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.835419 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.835429 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.835446 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.835456 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:02Z","lastTransitionTime":"2026-02-02T12:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.938779 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.938816 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.938825 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.938840 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:02 crc kubenswrapper[4846]: I0202 12:11:02.938851 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:02Z","lastTransitionTime":"2026-02-02T12:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.041674 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.041729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.041741 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.041755 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.041766 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:03Z","lastTransitionTime":"2026-02-02T12:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.144072 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.144107 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.144115 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.144128 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.144138 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:03Z","lastTransitionTime":"2026-02-02T12:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.247172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.247268 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.247312 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.247334 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.247381 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:03Z","lastTransitionTime":"2026-02-02T12:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.351455 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.351571 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.351686 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.351719 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.351786 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:03Z","lastTransitionTime":"2026-02-02T12:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.423129 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.423227 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:03 crc kubenswrapper[4846]: E0202 12:11:03.423369 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.423431 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:03 crc kubenswrapper[4846]: E0202 12:11:03.423563 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:03 crc kubenswrapper[4846]: E0202 12:11:03.423666 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.427973 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 12:26:56.2810809 +0000 UTC Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.456260 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.456319 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.456340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.456365 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.456384 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:03Z","lastTransitionTime":"2026-02-02T12:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.559371 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.559441 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.559459 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.559484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.559500 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:03Z","lastTransitionTime":"2026-02-02T12:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.662997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.663031 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.663041 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.663056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.663068 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:03Z","lastTransitionTime":"2026-02-02T12:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.765813 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.766245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.766469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.766735 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.766990 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:03Z","lastTransitionTime":"2026-02-02T12:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.870248 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.870305 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.870321 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.870358 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.870396 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:03Z","lastTransitionTime":"2026-02-02T12:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.973098 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.973153 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.973170 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.973195 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:03 crc kubenswrapper[4846]: I0202 12:11:03.973212 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:03Z","lastTransitionTime":"2026-02-02T12:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.075497 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.075528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.075539 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.075553 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.075563 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:04Z","lastTransitionTime":"2026-02-02T12:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.178774 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.178819 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.178828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.178845 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.178854 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:04Z","lastTransitionTime":"2026-02-02T12:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.280742 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.280772 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.280780 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.280793 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.280802 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:04Z","lastTransitionTime":"2026-02-02T12:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.383620 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.383993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.384314 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.384580 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.384850 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:04Z","lastTransitionTime":"2026-02-02T12:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.423276 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:04 crc kubenswrapper[4846]: E0202 12:11:04.423414 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.423817 4846 scope.go:117] "RemoveContainer" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:11:04 crc kubenswrapper[4846]: E0202 12:11:04.424042 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.428825 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 22:17:25.628327274 +0000 UTC Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.487759 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.488518 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.488777 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.488948 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.489075 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:04Z","lastTransitionTime":"2026-02-02T12:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.590875 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.591139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.591266 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.591588 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.591877 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:04Z","lastTransitionTime":"2026-02-02T12:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.694725 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.694990 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.695162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.695261 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.695333 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:04Z","lastTransitionTime":"2026-02-02T12:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.797899 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.797935 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.797943 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.797956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.797965 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:04Z","lastTransitionTime":"2026-02-02T12:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.901042 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.901071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.901080 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.901095 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:04 crc kubenswrapper[4846]: I0202 12:11:04.901107 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:04Z","lastTransitionTime":"2026-02-02T12:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.004760 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.004828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.004850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.004878 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.004895 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:05Z","lastTransitionTime":"2026-02-02T12:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.106763 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.106795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.106805 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.106818 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.106827 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:05Z","lastTransitionTime":"2026-02-02T12:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.208883 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.209105 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.209187 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.209285 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.209373 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:05Z","lastTransitionTime":"2026-02-02T12:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.313889 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.313986 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.313999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.314029 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.314042 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:05Z","lastTransitionTime":"2026-02-02T12:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.416346 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.416414 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.416438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.416466 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.416488 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:05Z","lastTransitionTime":"2026-02-02T12:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.422908 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.423034 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:05 crc kubenswrapper[4846]: E0202 12:11:05.423208 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.423245 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:05 crc kubenswrapper[4846]: E0202 12:11:05.423300 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:05 crc kubenswrapper[4846]: E0202 12:11:05.423353 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.429649 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 02:19:32.11356992 +0000 UTC Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.519991 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.520043 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.520057 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.520076 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.520089 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:05Z","lastTransitionTime":"2026-02-02T12:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.622417 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.622490 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.622512 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.622575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.622598 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:05Z","lastTransitionTime":"2026-02-02T12:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.725091 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.725117 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.725124 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.725138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.725148 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:05Z","lastTransitionTime":"2026-02-02T12:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.827256 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.827328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.827352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.827381 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.827434 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:05Z","lastTransitionTime":"2026-02-02T12:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.930242 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.930281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.930294 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.930311 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:05 crc kubenswrapper[4846]: I0202 12:11:05.930322 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:05Z","lastTransitionTime":"2026-02-02T12:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.032952 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.033014 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.033031 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.033052 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.033068 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:06Z","lastTransitionTime":"2026-02-02T12:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.136092 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.136131 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.136147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.136201 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.136226 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:06Z","lastTransitionTime":"2026-02-02T12:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.239366 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.239412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.239423 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.239438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.239448 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:06Z","lastTransitionTime":"2026-02-02T12:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.342081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.342145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.342168 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.342204 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.342226 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:06Z","lastTransitionTime":"2026-02-02T12:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.422812 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:06 crc kubenswrapper[4846]: E0202 12:11:06.422988 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.430070 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 06:00:09.26834604 +0000 UTC Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.445533 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.445592 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.445660 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.445713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.445737 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:06Z","lastTransitionTime":"2026-02-02T12:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.549102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.549148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.549157 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.549173 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.549182 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:06Z","lastTransitionTime":"2026-02-02T12:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.652412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.652475 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.652492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.652516 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.652533 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:06Z","lastTransitionTime":"2026-02-02T12:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.754979 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.755260 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.755285 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.755315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.755339 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:06Z","lastTransitionTime":"2026-02-02T12:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.859043 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.859109 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.859134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.859164 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.859187 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:06Z","lastTransitionTime":"2026-02-02T12:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.960916 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.960963 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.960985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.961005 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:06 crc kubenswrapper[4846]: I0202 12:11:06.961024 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:06Z","lastTransitionTime":"2026-02-02T12:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.004845 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.004931 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.004958 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.004992 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.005032 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T12:11:07Z","lastTransitionTime":"2026-02-02T12:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.054799 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx"] Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.055298 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.057140 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.060038 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.060544 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.060542 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.095197 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.095181134 podStartE2EDuration="1m18.095181134s" podCreationTimestamp="2026-02-02 12:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.078318824 +0000 UTC m=+98.306905717" watchObservedRunningTime="2026-02-02 12:11:07.095181134 +0000 UTC m=+98.323767997" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.106167 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-44zpn" podStartSLOduration=74.106103856 podStartE2EDuration="1m14.106103856s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.104795711 +0000 UTC m=+98.333382574" watchObservedRunningTime="2026-02-02 12:11:07.106103856 +0000 UTC m=+98.334690729" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.135538 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-7jshv" podStartSLOduration=74.135514731 podStartE2EDuration="1m14.135514731s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.122198795 +0000 UTC m=+98.350785658" watchObservedRunningTime="2026-02-02 12:11:07.135514731 +0000 UTC m=+98.364101604" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.150582 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podStartSLOduration=74.150563642 podStartE2EDuration="1m14.150563642s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.13585789 +0000 UTC m=+98.364444763" watchObservedRunningTime="2026-02-02 12:11:07.150563642 +0000 UTC m=+98.379150515" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.150860 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-6jrkw" podStartSLOduration=74.150855261 podStartE2EDuration="1m14.150855261s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.150043509 +0000 UTC m=+98.378630412" watchObservedRunningTime="2026-02-02 12:11:07.150855261 +0000 UTC m=+98.379442134" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.161310 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.161364 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.161389 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.161427 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.161453 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.184137 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-9vvp2" podStartSLOduration=74.184118738 podStartE2EDuration="1m14.184118738s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.183758879 +0000 UTC m=+98.412345742" watchObservedRunningTime="2026-02-02 12:11:07.184118738 +0000 UTC m=+98.412705601" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.218182 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=76.218166017 podStartE2EDuration="1m16.218166017s" podCreationTimestamp="2026-02-02 12:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.208704695 +0000 UTC m=+98.437291558" watchObservedRunningTime="2026-02-02 12:11:07.218166017 +0000 UTC m=+98.446752880" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.229365 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=24.229348116 podStartE2EDuration="24.229348116s" podCreationTimestamp="2026-02-02 12:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.218712642 +0000 UTC m=+98.447299505" watchObservedRunningTime="2026-02-02 12:11:07.229348116 +0000 UTC m=+98.457934979" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.252601 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cscmz" podStartSLOduration=73.252584606 podStartE2EDuration="1m13.252584606s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.25235476 +0000 UTC m=+98.480941623" watchObservedRunningTime="2026-02-02 12:11:07.252584606 +0000 UTC m=+98.481171469" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.262311 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.262357 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.262393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.262421 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.262421 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.262440 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.262702 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.263387 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.269857 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.281345 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xlzwx\" (UID: \"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.295765 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.295744448 podStartE2EDuration="1m19.295744448s" podCreationTimestamp="2026-02-02 12:09:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.28348212 +0000 UTC m=+98.512068983" watchObservedRunningTime="2026-02-02 12:11:07.295744448 +0000 UTC m=+98.524331311" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.311265 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.311244461 podStartE2EDuration="45.311244461s" podCreationTimestamp="2026-02-02 12:10:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.296492148 +0000 UTC m=+98.525079011" watchObservedRunningTime="2026-02-02 12:11:07.311244461 +0000 UTC m=+98.539831334" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.372170 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.422641 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.422641 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.422709 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:07 crc kubenswrapper[4846]: E0202 12:11:07.422800 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:07 crc kubenswrapper[4846]: E0202 12:11:07.422852 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:07 crc kubenswrapper[4846]: E0202 12:11:07.423018 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.430820 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 00:14:20.691206595 +0000 UTC Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.431211 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.438323 4846 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.960265 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" event={"ID":"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d","Type":"ContainerStarted","Data":"6a1e993679cf932a8004f19e659adde0c44cb6654bac3375f354d495e705d27c"} Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.960566 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" event={"ID":"f84d87f2-f71d-4f74-9d7d-a5cdc8d1193d","Type":"ContainerStarted","Data":"17216ef425b123dfef1801fe3ef7c2428e444a77ecdcb702473132d5085b1767"} Feb 02 12:11:07 crc kubenswrapper[4846]: I0202 12:11:07.976255 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xlzwx" podStartSLOduration=73.976238552 podStartE2EDuration="1m13.976238552s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:07.975109742 +0000 UTC m=+99.203696615" watchObservedRunningTime="2026-02-02 12:11:07.976238552 +0000 UTC m=+99.204825435" Feb 02 12:11:08 crc kubenswrapper[4846]: I0202 12:11:08.423174 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:08 crc kubenswrapper[4846]: E0202 12:11:08.423336 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:09 crc kubenswrapper[4846]: I0202 12:11:09.422984 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:09 crc kubenswrapper[4846]: I0202 12:11:09.423066 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:09 crc kubenswrapper[4846]: E0202 12:11:09.424394 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:09 crc kubenswrapper[4846]: I0202 12:11:09.424425 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:09 crc kubenswrapper[4846]: E0202 12:11:09.424554 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:09 crc kubenswrapper[4846]: E0202 12:11:09.424693 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:10 crc kubenswrapper[4846]: I0202 12:11:10.423175 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:10 crc kubenswrapper[4846]: E0202 12:11:10.423294 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:11 crc kubenswrapper[4846]: I0202 12:11:11.423438 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:11 crc kubenswrapper[4846]: I0202 12:11:11.423532 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:11 crc kubenswrapper[4846]: E0202 12:11:11.424056 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:11 crc kubenswrapper[4846]: I0202 12:11:11.423669 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:11 crc kubenswrapper[4846]: E0202 12:11:11.424144 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:11 crc kubenswrapper[4846]: E0202 12:11:11.424356 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:12 crc kubenswrapper[4846]: I0202 12:11:12.324228 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:12 crc kubenswrapper[4846]: E0202 12:11:12.324676 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:11:12 crc kubenswrapper[4846]: E0202 12:11:12.324788 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs podName:fe216e68-8814-468d-a6f3-fe8721d94eba nodeName:}" failed. No retries permitted until 2026-02-02 12:12:16.324762047 +0000 UTC m=+167.553348950 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs") pod "network-metrics-daemon-j5m9k" (UID: "fe216e68-8814-468d-a6f3-fe8721d94eba") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 12:11:12 crc kubenswrapper[4846]: I0202 12:11:12.422904 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:12 crc kubenswrapper[4846]: E0202 12:11:12.423076 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:13 crc kubenswrapper[4846]: I0202 12:11:13.422654 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:13 crc kubenswrapper[4846]: I0202 12:11:13.422741 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:13 crc kubenswrapper[4846]: E0202 12:11:13.422824 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:13 crc kubenswrapper[4846]: I0202 12:11:13.422994 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:13 crc kubenswrapper[4846]: E0202 12:11:13.423028 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:13 crc kubenswrapper[4846]: E0202 12:11:13.423082 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:14 crc kubenswrapper[4846]: I0202 12:11:14.422598 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:14 crc kubenswrapper[4846]: E0202 12:11:14.422771 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:15 crc kubenswrapper[4846]: I0202 12:11:15.424565 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:15 crc kubenswrapper[4846]: I0202 12:11:15.424565 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:15 crc kubenswrapper[4846]: E0202 12:11:15.424735 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:15 crc kubenswrapper[4846]: E0202 12:11:15.424824 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:15 crc kubenswrapper[4846]: I0202 12:11:15.425170 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:15 crc kubenswrapper[4846]: E0202 12:11:15.425271 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:16 crc kubenswrapper[4846]: I0202 12:11:16.423200 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:16 crc kubenswrapper[4846]: E0202 12:11:16.423339 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:17 crc kubenswrapper[4846]: I0202 12:11:17.422690 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:17 crc kubenswrapper[4846]: I0202 12:11:17.422919 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:17 crc kubenswrapper[4846]: I0202 12:11:17.423077 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:17 crc kubenswrapper[4846]: E0202 12:11:17.424076 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:17 crc kubenswrapper[4846]: E0202 12:11:17.423969 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:17 crc kubenswrapper[4846]: E0202 12:11:17.424379 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:18 crc kubenswrapper[4846]: I0202 12:11:18.422505 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:18 crc kubenswrapper[4846]: E0202 12:11:18.422654 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:19 crc kubenswrapper[4846]: I0202 12:11:19.422577 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:19 crc kubenswrapper[4846]: I0202 12:11:19.422702 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:19 crc kubenswrapper[4846]: E0202 12:11:19.423781 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:19 crc kubenswrapper[4846]: I0202 12:11:19.423940 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:19 crc kubenswrapper[4846]: E0202 12:11:19.424361 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:19 crc kubenswrapper[4846]: E0202 12:11:19.424549 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:19 crc kubenswrapper[4846]: I0202 12:11:19.425068 4846 scope.go:117] "RemoveContainer" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:11:19 crc kubenswrapper[4846]: E0202 12:11:19.425346 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ftvcw_openshift-ovn-kubernetes(6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" Feb 02 12:11:20 crc kubenswrapper[4846]: I0202 12:11:20.423030 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:20 crc kubenswrapper[4846]: E0202 12:11:20.423173 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:21 crc kubenswrapper[4846]: I0202 12:11:21.422492 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:21 crc kubenswrapper[4846]: I0202 12:11:21.422528 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:21 crc kubenswrapper[4846]: E0202 12:11:21.422687 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:21 crc kubenswrapper[4846]: I0202 12:11:21.422752 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:21 crc kubenswrapper[4846]: E0202 12:11:21.422781 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:21 crc kubenswrapper[4846]: E0202 12:11:21.422938 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:22 crc kubenswrapper[4846]: I0202 12:11:22.423222 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:22 crc kubenswrapper[4846]: E0202 12:11:22.423364 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:23 crc kubenswrapper[4846]: I0202 12:11:23.423116 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:23 crc kubenswrapper[4846]: I0202 12:11:23.423170 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:23 crc kubenswrapper[4846]: E0202 12:11:23.423268 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:23 crc kubenswrapper[4846]: E0202 12:11:23.423362 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:23 crc kubenswrapper[4846]: I0202 12:11:23.423426 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:23 crc kubenswrapper[4846]: E0202 12:11:23.423558 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:24 crc kubenswrapper[4846]: I0202 12:11:24.423126 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:24 crc kubenswrapper[4846]: E0202 12:11:24.423278 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:25 crc kubenswrapper[4846]: I0202 12:11:25.423402 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:25 crc kubenswrapper[4846]: I0202 12:11:25.423502 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:25 crc kubenswrapper[4846]: E0202 12:11:25.423556 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:25 crc kubenswrapper[4846]: E0202 12:11:25.423753 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:25 crc kubenswrapper[4846]: I0202 12:11:25.423802 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:25 crc kubenswrapper[4846]: E0202 12:11:25.423854 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:26 crc kubenswrapper[4846]: I0202 12:11:26.423388 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:26 crc kubenswrapper[4846]: E0202 12:11:26.423776 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:27 crc kubenswrapper[4846]: I0202 12:11:27.422698 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:27 crc kubenswrapper[4846]: I0202 12:11:27.422800 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:27 crc kubenswrapper[4846]: E0202 12:11:27.422846 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:27 crc kubenswrapper[4846]: E0202 12:11:27.422948 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:27 crc kubenswrapper[4846]: I0202 12:11:27.422699 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:27 crc kubenswrapper[4846]: E0202 12:11:27.423339 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:28 crc kubenswrapper[4846]: I0202 12:11:28.422656 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:28 crc kubenswrapper[4846]: E0202 12:11:28.422797 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:29 crc kubenswrapper[4846]: I0202 12:11:29.036262 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7jshv_24fc18fa-9739-480c-8f43-ab30989164c2/kube-multus/1.log" Feb 02 12:11:29 crc kubenswrapper[4846]: I0202 12:11:29.037168 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7jshv_24fc18fa-9739-480c-8f43-ab30989164c2/kube-multus/0.log" Feb 02 12:11:29 crc kubenswrapper[4846]: I0202 12:11:29.037209 4846 generic.go:334] "Generic (PLEG): container finished" podID="24fc18fa-9739-480c-8f43-ab30989164c2" containerID="abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92" exitCode=1 Feb 02 12:11:29 crc kubenswrapper[4846]: I0202 12:11:29.037242 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7jshv" event={"ID":"24fc18fa-9739-480c-8f43-ab30989164c2","Type":"ContainerDied","Data":"abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92"} Feb 02 12:11:29 crc kubenswrapper[4846]: I0202 12:11:29.037277 4846 scope.go:117] "RemoveContainer" containerID="de690261aaabeb792a0f4900392c341b8e35b6529f6bc4194f2625ae2de6c63a" Feb 02 12:11:29 crc kubenswrapper[4846]: I0202 12:11:29.037700 4846 scope.go:117] "RemoveContainer" containerID="abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92" Feb 02 12:11:29 crc kubenswrapper[4846]: E0202 12:11:29.037846 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-7jshv_openshift-multus(24fc18fa-9739-480c-8f43-ab30989164c2)\"" pod="openshift-multus/multus-7jshv" podUID="24fc18fa-9739-480c-8f43-ab30989164c2" Feb 02 12:11:29 crc kubenswrapper[4846]: E0202 12:11:29.383458 4846 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 02 12:11:29 crc kubenswrapper[4846]: I0202 12:11:29.423127 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:29 crc kubenswrapper[4846]: I0202 12:11:29.423189 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:29 crc kubenswrapper[4846]: I0202 12:11:29.423269 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:29 crc kubenswrapper[4846]: E0202 12:11:29.424200 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:29 crc kubenswrapper[4846]: E0202 12:11:29.424370 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:29 crc kubenswrapper[4846]: E0202 12:11:29.424542 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:29 crc kubenswrapper[4846]: E0202 12:11:29.504921 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 12:11:30 crc kubenswrapper[4846]: I0202 12:11:30.042211 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7jshv_24fc18fa-9739-480c-8f43-ab30989164c2/kube-multus/1.log" Feb 02 12:11:30 crc kubenswrapper[4846]: I0202 12:11:30.423254 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:30 crc kubenswrapper[4846]: E0202 12:11:30.424946 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:31 crc kubenswrapper[4846]: I0202 12:11:31.423014 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:31 crc kubenswrapper[4846]: I0202 12:11:31.423105 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:31 crc kubenswrapper[4846]: E0202 12:11:31.423181 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:31 crc kubenswrapper[4846]: I0202 12:11:31.423236 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:31 crc kubenswrapper[4846]: E0202 12:11:31.423386 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:31 crc kubenswrapper[4846]: E0202 12:11:31.423490 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:32 crc kubenswrapper[4846]: I0202 12:11:32.423154 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:32 crc kubenswrapper[4846]: E0202 12:11:32.423655 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:33 crc kubenswrapper[4846]: I0202 12:11:33.423521 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:33 crc kubenswrapper[4846]: I0202 12:11:33.423573 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:33 crc kubenswrapper[4846]: I0202 12:11:33.423683 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:33 crc kubenswrapper[4846]: E0202 12:11:33.423794 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:33 crc kubenswrapper[4846]: E0202 12:11:33.424421 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:33 crc kubenswrapper[4846]: E0202 12:11:33.424470 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:33 crc kubenswrapper[4846]: I0202 12:11:33.425047 4846 scope.go:117] "RemoveContainer" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:11:34 crc kubenswrapper[4846]: I0202 12:11:34.058333 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/3.log" Feb 02 12:11:34 crc kubenswrapper[4846]: I0202 12:11:34.061814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerStarted","Data":"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa"} Feb 02 12:11:34 crc kubenswrapper[4846]: I0202 12:11:34.062997 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:11:34 crc kubenswrapper[4846]: I0202 12:11:34.095705 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podStartSLOduration=100.095675711 podStartE2EDuration="1m40.095675711s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:11:34.094877349 +0000 UTC m=+125.323464232" watchObservedRunningTime="2026-02-02 12:11:34.095675711 +0000 UTC m=+125.324262584" Feb 02 12:11:34 crc kubenswrapper[4846]: I0202 12:11:34.219067 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-j5m9k"] Feb 02 12:11:34 crc kubenswrapper[4846]: I0202 12:11:34.219239 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:34 crc kubenswrapper[4846]: E0202 12:11:34.219357 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:34 crc kubenswrapper[4846]: E0202 12:11:34.505670 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 12:11:35 crc kubenswrapper[4846]: I0202 12:11:35.423039 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:35 crc kubenswrapper[4846]: I0202 12:11:35.423112 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:35 crc kubenswrapper[4846]: I0202 12:11:35.423151 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:35 crc kubenswrapper[4846]: E0202 12:11:35.423760 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:35 crc kubenswrapper[4846]: E0202 12:11:35.423843 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:35 crc kubenswrapper[4846]: E0202 12:11:35.423937 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:36 crc kubenswrapper[4846]: I0202 12:11:36.423353 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:36 crc kubenswrapper[4846]: E0202 12:11:36.423506 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:37 crc kubenswrapper[4846]: I0202 12:11:37.422567 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:37 crc kubenswrapper[4846]: I0202 12:11:37.422652 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:37 crc kubenswrapper[4846]: E0202 12:11:37.422808 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:37 crc kubenswrapper[4846]: I0202 12:11:37.422858 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:37 crc kubenswrapper[4846]: E0202 12:11:37.423066 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:37 crc kubenswrapper[4846]: E0202 12:11:37.423161 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:38 crc kubenswrapper[4846]: I0202 12:11:38.422974 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:38 crc kubenswrapper[4846]: E0202 12:11:38.423229 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:39 crc kubenswrapper[4846]: I0202 12:11:39.423423 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:39 crc kubenswrapper[4846]: I0202 12:11:39.423524 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:39 crc kubenswrapper[4846]: E0202 12:11:39.424476 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:39 crc kubenswrapper[4846]: I0202 12:11:39.424504 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:39 crc kubenswrapper[4846]: E0202 12:11:39.424644 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:39 crc kubenswrapper[4846]: E0202 12:11:39.424855 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:39 crc kubenswrapper[4846]: E0202 12:11:39.507051 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 12:11:40 crc kubenswrapper[4846]: I0202 12:11:40.423419 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:40 crc kubenswrapper[4846]: E0202 12:11:40.423691 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:41 crc kubenswrapper[4846]: I0202 12:11:41.422571 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:41 crc kubenswrapper[4846]: I0202 12:11:41.422697 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:41 crc kubenswrapper[4846]: I0202 12:11:41.422586 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:41 crc kubenswrapper[4846]: E0202 12:11:41.422844 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:41 crc kubenswrapper[4846]: E0202 12:11:41.422982 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:41 crc kubenswrapper[4846]: E0202 12:11:41.423138 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:42 crc kubenswrapper[4846]: I0202 12:11:42.423124 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:42 crc kubenswrapper[4846]: E0202 12:11:42.423291 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:43 crc kubenswrapper[4846]: I0202 12:11:43.422772 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:43 crc kubenswrapper[4846]: I0202 12:11:43.422889 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:43 crc kubenswrapper[4846]: E0202 12:11:43.422952 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:43 crc kubenswrapper[4846]: E0202 12:11:43.423108 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:43 crc kubenswrapper[4846]: I0202 12:11:43.423124 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:43 crc kubenswrapper[4846]: E0202 12:11:43.423287 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:43 crc kubenswrapper[4846]: I0202 12:11:43.424288 4846 scope.go:117] "RemoveContainer" containerID="abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92" Feb 02 12:11:44 crc kubenswrapper[4846]: I0202 12:11:44.098960 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7jshv_24fc18fa-9739-480c-8f43-ab30989164c2/kube-multus/1.log" Feb 02 12:11:44 crc kubenswrapper[4846]: I0202 12:11:44.099317 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7jshv" event={"ID":"24fc18fa-9739-480c-8f43-ab30989164c2","Type":"ContainerStarted","Data":"4c7034f84bb93b8fc083ef27a05cac5d8f31339db2de4fe421ee29ac6ea3417e"} Feb 02 12:11:44 crc kubenswrapper[4846]: I0202 12:11:44.422438 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:44 crc kubenswrapper[4846]: E0202 12:11:44.422592 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:44 crc kubenswrapper[4846]: E0202 12:11:44.508838 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 12:11:45 crc kubenswrapper[4846]: I0202 12:11:45.423158 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:45 crc kubenswrapper[4846]: I0202 12:11:45.423217 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:45 crc kubenswrapper[4846]: I0202 12:11:45.423158 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:45 crc kubenswrapper[4846]: E0202 12:11:45.423324 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:45 crc kubenswrapper[4846]: E0202 12:11:45.423466 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:45 crc kubenswrapper[4846]: E0202 12:11:45.423594 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:46 crc kubenswrapper[4846]: I0202 12:11:46.423179 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:46 crc kubenswrapper[4846]: E0202 12:11:46.423405 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:47 crc kubenswrapper[4846]: I0202 12:11:47.422715 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:47 crc kubenswrapper[4846]: I0202 12:11:47.422739 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:47 crc kubenswrapper[4846]: I0202 12:11:47.422964 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:47 crc kubenswrapper[4846]: E0202 12:11:47.422914 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:47 crc kubenswrapper[4846]: E0202 12:11:47.423163 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:47 crc kubenswrapper[4846]: E0202 12:11:47.423246 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:48 crc kubenswrapper[4846]: I0202 12:11:48.422840 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:48 crc kubenswrapper[4846]: E0202 12:11:48.422999 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-j5m9k" podUID="fe216e68-8814-468d-a6f3-fe8721d94eba" Feb 02 12:11:49 crc kubenswrapper[4846]: I0202 12:11:49.423445 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:49 crc kubenswrapper[4846]: I0202 12:11:49.423530 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:49 crc kubenswrapper[4846]: I0202 12:11:49.423496 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:49 crc kubenswrapper[4846]: E0202 12:11:49.425300 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 12:11:49 crc kubenswrapper[4846]: E0202 12:11:49.425493 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 12:11:49 crc kubenswrapper[4846]: E0202 12:11:49.425680 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 12:11:50 crc kubenswrapper[4846]: I0202 12:11:50.422813 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:11:50 crc kubenswrapper[4846]: I0202 12:11:50.426296 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 02 12:11:50 crc kubenswrapper[4846]: I0202 12:11:50.426557 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 02 12:11:51 crc kubenswrapper[4846]: I0202 12:11:51.423900 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:51 crc kubenswrapper[4846]: I0202 12:11:51.423912 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:51 crc kubenswrapper[4846]: I0202 12:11:51.425188 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:51 crc kubenswrapper[4846]: I0202 12:11:51.426383 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 02 12:11:51 crc kubenswrapper[4846]: I0202 12:11:51.426401 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 02 12:11:51 crc kubenswrapper[4846]: I0202 12:11:51.426494 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 02 12:11:51 crc kubenswrapper[4846]: I0202 12:11:51.427129 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 02 12:11:52 crc kubenswrapper[4846]: I0202 12:11:52.906642 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.390485 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.390751 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:57 crc kubenswrapper[4846]: E0202 12:11:57.390771 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:13:59.390740515 +0000 UTC m=+270.619327368 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.390867 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.390923 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.392408 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.396671 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.397435 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.446932 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.456981 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.492346 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.496306 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:57 crc kubenswrapper[4846]: W0202 12:11:57.667527 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-59f58c439cf66f2f31f8b333040e26a448ecec112835ec76185457c164d41d7f WatchSource:0}: Error finding container 59f58c439cf66f2f31f8b333040e26a448ecec112835ec76185457c164d41d7f: Status 404 returned error can't find the container with id 59f58c439cf66f2f31f8b333040e26a448ecec112835ec76185457c164d41d7f Feb 02 12:11:57 crc kubenswrapper[4846]: I0202 12:11:57.766567 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:57 crc kubenswrapper[4846]: W0202 12:11:57.943361 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-a4a425acee1901f275a1af3ff7eb3ef0783135807b5b184af4d372d3adabee6b WatchSource:0}: Error finding container a4a425acee1901f275a1af3ff7eb3ef0783135807b5b184af4d372d3adabee6b: Status 404 returned error can't find the container with id a4a425acee1901f275a1af3ff7eb3ef0783135807b5b184af4d372d3adabee6b Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.145475 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3457a121f44451ee18d9fa08e769f77314a4fafc7608c994b05f9cb58e54e2de"} Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.145518 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a4a425acee1901f275a1af3ff7eb3ef0783135807b5b184af4d372d3adabee6b"} Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.146288 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.147599 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"090feffee3837c8e575c2c6f15cd6c156e0d7974dd9e02e35e3a7b17b82e471e"} Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.147659 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"38eab22352ad934aef1fd1cc01986c633943119d83424276b76592544132ace6"} Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.148875 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f9ab34d8c933a01a0bd33642326034d1587aac35312df01703355bc1e4dd659e"} Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.148902 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"59f58c439cf66f2f31f8b333040e26a448ecec112835ec76185457c164d41d7f"} Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.222913 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.258755 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bbmc8"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.259245 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.261174 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9h9tw"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.262058 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.263003 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.264765 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7qdsx"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.265335 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.267154 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.268066 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.271083 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.271410 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.271879 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.271916 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.272301 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.272828 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.273315 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.273502 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.273657 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.273712 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.273743 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.273807 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.273839 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.274004 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.274001 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.274113 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.273505 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.274727 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-79skc"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.274202 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.275060 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.275752 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.275841 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.275914 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.275988 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.276052 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.276173 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.276186 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.276218 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.276276 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.277066 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.277283 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.277901 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.278822 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.284688 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-bckqh"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.285277 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bckqh" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.286285 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.306975 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.307262 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.307094 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: W0202 12:11:58.307588 4846 reflector.go:561] object-"openshift-console"/"console-serving-cert": failed to list *v1.Secret: secrets "console-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.307536 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 02 12:11:58 crc kubenswrapper[4846]: E0202 12:11:58.307651 4846 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"console-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"console-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.307602 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.307712 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.307607 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.307933 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.307969 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.308042 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.308103 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.308502 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.308689 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.308928 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.321213 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.321704 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.321963 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.323138 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.323401 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-q57ps"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.323839 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.324833 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.325131 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6kzjp"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.325406 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.326981 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.328078 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.331559 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.337088 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.337438 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.337675 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.337951 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.338470 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.338574 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.341745 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5glmk"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.342229 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.347148 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94n5d"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.348028 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.350328 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zgkpl"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.350685 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9h9tw"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.350704 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.350992 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.351235 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.378251 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.381279 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.381817 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7qdsx"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.384236 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.384360 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.384428 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.384535 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.384606 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.384694 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.384762 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.384835 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.384904 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.384967 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.385026 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.386473 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.386982 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.387190 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.387349 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.387522 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.387706 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.387979 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.388166 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.388339 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.388613 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.388826 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.389003 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.389161 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.389337 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.392091 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.392213 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.392256 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.392314 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.392453 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.392477 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.397608 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.397930 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.398135 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.398838 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.405840 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.406069 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.406202 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.406431 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.406783 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.406919 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.407038 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.407155 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.407269 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.407484 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.407781 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.424397 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.429073 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e94b3ca8-f11a-4fee-8ef1-05a56d233593-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.429314 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddm9z\" (UniqueName: \"kubernetes.io/projected/e94b3ca8-f11a-4fee-8ef1-05a56d233593-kube-api-access-ddm9z\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.429420 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-config\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.429516 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a62db6be-7478-41ce-b1a7-fbf3f24967e2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5glmk\" (UID: \"a62db6be-7478-41ce-b1a7-fbf3f24967e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.429662 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.429755 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzzzd\" (UniqueName: \"kubernetes.io/projected/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-kube-api-access-mzzzd\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.429856 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12c5b3f2-6d12-4f68-8013-fd5a555b7cfc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2jvrx\" (UID: \"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.429949 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c142082-61b2-4ea2-9d3b-8f9bae616178-service-ca-bundle\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.430054 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c142082-61b2-4ea2-9d3b-8f9bae616178-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.430145 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e94b3ca8-f11a-4fee-8ef1-05a56d233593-audit-dir\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.430238 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-oauth-config\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.430345 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zftj\" (UniqueName: \"kubernetes.io/projected/614c989a-6be2-4824-a3e5-d810fd94e39b-kube-api-access-7zftj\") pod \"openshift-apiserver-operator-796bbdcf4f-ml2fm\" (UID: \"614c989a-6be2-4824-a3e5-d810fd94e39b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.430455 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h2p8\" (UniqueName: \"kubernetes.io/projected/a62db6be-7478-41ce-b1a7-fbf3f24967e2-kube-api-access-9h2p8\") pod \"openshift-config-operator-7777fb866f-5glmk\" (UID: \"a62db6be-7478-41ce-b1a7-fbf3f24967e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.430563 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8abc246a-7ca3-4550-ba36-491f830e5419-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-px4bj\" (UID: \"8abc246a-7ca3-4550-ba36-491f830e5419\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.430673 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e94b3ca8-f11a-4fee-8ef1-05a56d233593-serving-cert\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.430770 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-service-ca\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.430880 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-478m7\" (UniqueName: \"kubernetes.io/projected/af51b9e4-0ee3-4bf9-93e5-27b4039be121-kube-api-access-478m7\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.430985 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-config\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431079 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/614c989a-6be2-4824-a3e5-d810fd94e39b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ml2fm\" (UID: \"614c989a-6be2-4824-a3e5-d810fd94e39b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431180 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-oauth-serving-cert\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431333 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htdln\" (UniqueName: \"kubernetes.io/projected/2c142082-61b2-4ea2-9d3b-8f9bae616178-kube-api-access-htdln\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431392 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-client-ca\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431416 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e94b3ca8-f11a-4fee-8ef1-05a56d233593-encryption-config\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431462 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e94b3ca8-f11a-4fee-8ef1-05a56d233593-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431491 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkncs\" (UniqueName: \"kubernetes.io/projected/8abc246a-7ca3-4550-ba36-491f830e5419-kube-api-access-hkncs\") pod \"cluster-samples-operator-665b6dd947-px4bj\" (UID: \"8abc246a-7ca3-4550-ba36-491f830e5419\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431535 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c142082-61b2-4ea2-9d3b-8f9bae616178-config\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431558 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkmx6\" (UniqueName: \"kubernetes.io/projected/12c5b3f2-6d12-4f68-8013-fd5a555b7cfc-kube-api-access-gkmx6\") pod \"openshift-controller-manager-operator-756b6f6bc6-2jvrx\" (UID: \"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431581 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/614c989a-6be2-4824-a3e5-d810fd94e39b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ml2fm\" (UID: \"614c989a-6be2-4824-a3e5-d810fd94e39b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431646 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-serving-cert\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431665 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c142082-61b2-4ea2-9d3b-8f9bae616178-serving-cert\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431684 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a62db6be-7478-41ce-b1a7-fbf3f24967e2-serving-cert\") pod \"openshift-config-operator-7777fb866f-5glmk\" (UID: \"a62db6be-7478-41ce-b1a7-fbf3f24967e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431726 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-trusted-ca-bundle\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431750 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e94b3ca8-f11a-4fee-8ef1-05a56d233593-audit-policies\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431770 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c5b3f2-6d12-4f68-8013-fd5a555b7cfc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2jvrx\" (UID: \"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.431807 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e94b3ca8-f11a-4fee-8ef1-05a56d233593-etcd-client\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.432228 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bbmc8"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.432334 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.433169 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.433891 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.433889 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.450003 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.450524 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.450752 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.452438 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.452600 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.453134 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.453984 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.454161 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.460865 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tsz82"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.461297 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-h77rj"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.461534 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.462311 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.464673 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.464948 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zbpnp"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.464958 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.465151 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.465285 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.465296 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.466933 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.468835 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.469241 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.469485 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.470079 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.472401 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7xgq6"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.473118 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.473710 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.474301 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.475368 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2472r"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.475877 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2472r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.475913 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.476704 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7gk4g"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.477172 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.478201 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-79skc"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.479814 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.480747 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.481476 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.481849 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.482978 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.483535 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.484427 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.484917 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.485102 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.485718 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.486716 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6kzjp"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.487503 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.489192 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.491107 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.491592 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.491680 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.491844 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.492204 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.497825 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94n5d"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.499818 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.501712 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-q57ps"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.502803 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5glmk"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.504540 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-jn5pn"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.505290 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jn5pn" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.506820 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-h6lcv"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.510520 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.518051 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.521935 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zgkpl"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.522206 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.526587 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.527106 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.537169 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bckqh"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.537945 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/614c989a-6be2-4824-a3e5-d810fd94e39b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ml2fm\" (UID: \"614c989a-6be2-4824-a3e5-d810fd94e39b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.537982 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-serving-cert\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538010 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c142082-61b2-4ea2-9d3b-8f9bae616178-serving-cert\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538034 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a62db6be-7478-41ce-b1a7-fbf3f24967e2-serving-cert\") pod \"openshift-config-operator-7777fb866f-5glmk\" (UID: \"a62db6be-7478-41ce-b1a7-fbf3f24967e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538072 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-trusted-ca-bundle\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538096 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e94b3ca8-f11a-4fee-8ef1-05a56d233593-audit-policies\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538119 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c5b3f2-6d12-4f68-8013-fd5a555b7cfc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2jvrx\" (UID: \"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538142 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e94b3ca8-f11a-4fee-8ef1-05a56d233593-etcd-client\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538167 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e94b3ca8-f11a-4fee-8ef1-05a56d233593-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538189 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddm9z\" (UniqueName: \"kubernetes.io/projected/e94b3ca8-f11a-4fee-8ef1-05a56d233593-kube-api-access-ddm9z\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538209 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-config\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538231 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a62db6be-7478-41ce-b1a7-fbf3f24967e2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5glmk\" (UID: \"a62db6be-7478-41ce-b1a7-fbf3f24967e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538253 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538273 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzzzd\" (UniqueName: \"kubernetes.io/projected/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-kube-api-access-mzzzd\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538297 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12c5b3f2-6d12-4f68-8013-fd5a555b7cfc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2jvrx\" (UID: \"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538318 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c142082-61b2-4ea2-9d3b-8f9bae616178-service-ca-bundle\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538339 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c142082-61b2-4ea2-9d3b-8f9bae616178-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538360 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e94b3ca8-f11a-4fee-8ef1-05a56d233593-audit-dir\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538399 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zftj\" (UniqueName: \"kubernetes.io/projected/614c989a-6be2-4824-a3e5-d810fd94e39b-kube-api-access-7zftj\") pod \"openshift-apiserver-operator-796bbdcf4f-ml2fm\" (UID: \"614c989a-6be2-4824-a3e5-d810fd94e39b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538422 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-oauth-config\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538442 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h2p8\" (UniqueName: \"kubernetes.io/projected/a62db6be-7478-41ce-b1a7-fbf3f24967e2-kube-api-access-9h2p8\") pod \"openshift-config-operator-7777fb866f-5glmk\" (UID: \"a62db6be-7478-41ce-b1a7-fbf3f24967e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538464 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8abc246a-7ca3-4550-ba36-491f830e5419-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-px4bj\" (UID: \"8abc246a-7ca3-4550-ba36-491f830e5419\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538489 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e94b3ca8-f11a-4fee-8ef1-05a56d233593-serving-cert\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538508 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-service-ca\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538536 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-478m7\" (UniqueName: \"kubernetes.io/projected/af51b9e4-0ee3-4bf9-93e5-27b4039be121-kube-api-access-478m7\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538561 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-config\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538582 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/614c989a-6be2-4824-a3e5-d810fd94e39b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ml2fm\" (UID: \"614c989a-6be2-4824-a3e5-d810fd94e39b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538604 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-oauth-serving-cert\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538644 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htdln\" (UniqueName: \"kubernetes.io/projected/2c142082-61b2-4ea2-9d3b-8f9bae616178-kube-api-access-htdln\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538668 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-client-ca\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538687 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e94b3ca8-f11a-4fee-8ef1-05a56d233593-encryption-config\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538708 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e94b3ca8-f11a-4fee-8ef1-05a56d233593-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538729 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkncs\" (UniqueName: \"kubernetes.io/projected/8abc246a-7ca3-4550-ba36-491f830e5419-kube-api-access-hkncs\") pod \"cluster-samples-operator-665b6dd947-px4bj\" (UID: \"8abc246a-7ca3-4550-ba36-491f830e5419\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkmx6\" (UniqueName: \"kubernetes.io/projected/12c5b3f2-6d12-4f68-8013-fd5a555b7cfc-kube-api-access-gkmx6\") pod \"openshift-controller-manager-operator-756b6f6bc6-2jvrx\" (UID: \"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538776 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c142082-61b2-4ea2-9d3b-8f9bae616178-config\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.538870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e94b3ca8-f11a-4fee-8ef1-05a56d233593-audit-dir\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.539761 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c142082-61b2-4ea2-9d3b-8f9bae616178-config\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.540863 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a62db6be-7478-41ce-b1a7-fbf3f24967e2-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5glmk\" (UID: \"a62db6be-7478-41ce-b1a7-fbf3f24967e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.541318 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/614c989a-6be2-4824-a3e5-d810fd94e39b-config\") pod \"openshift-apiserver-operator-796bbdcf4f-ml2fm\" (UID: \"614c989a-6be2-4824-a3e5-d810fd94e39b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.541445 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e94b3ca8-f11a-4fee-8ef1-05a56d233593-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.541488 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12c5b3f2-6d12-4f68-8013-fd5a555b7cfc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2jvrx\" (UID: \"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.541867 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-config\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.542472 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.543144 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-service-ca\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.543434 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a62db6be-7478-41ce-b1a7-fbf3f24967e2-serving-cert\") pod \"openshift-config-operator-7777fb866f-5glmk\" (UID: \"a62db6be-7478-41ce-b1a7-fbf3f24967e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.543474 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-serving-cert\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.543533 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-trusted-ca-bundle\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.544079 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e94b3ca8-f11a-4fee-8ef1-05a56d233593-audit-policies\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.544260 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.544359 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-oauth-serving-cert\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.545555 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-config\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.545775 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-client-ca\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.546166 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e94b3ca8-f11a-4fee-8ef1-05a56d233593-serving-cert\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.546332 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c142082-61b2-4ea2-9d3b-8f9bae616178-service-ca-bundle\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.546425 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e94b3ca8-f11a-4fee-8ef1-05a56d233593-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.546765 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e94b3ca8-f11a-4fee-8ef1-05a56d233593-encryption-config\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.546938 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c142082-61b2-4ea2-9d3b-8f9bae616178-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.547163 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-oauth-config\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.547318 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.547567 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.549085 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e94b3ca8-f11a-4fee-8ef1-05a56d233593-etcd-client\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.551069 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.551081 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c142082-61b2-4ea2-9d3b-8f9bae616178-serving-cert\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.551683 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tsz82"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.552521 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8abc246a-7ca3-4550-ba36-491f830e5419-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-px4bj\" (UID: \"8abc246a-7ca3-4550-ba36-491f830e5419\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.553508 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-kbvlq"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.553933 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12c5b3f2-6d12-4f68-8013-fd5a555b7cfc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2jvrx\" (UID: \"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.554300 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kbvlq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.557665 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-kjwnm"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.558104 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.558175 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.558293 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.560549 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7xgq6"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.562464 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/614c989a-6be2-4824-a3e5-d810fd94e39b-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-ml2fm\" (UID: \"614c989a-6be2-4824-a3e5-d810fd94e39b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.562805 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.564601 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.565907 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.568064 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.569475 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7gk4g"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.570847 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.573782 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.575378 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zbpnp"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.577046 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.578273 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2472r"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.579608 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.581028 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jn5pn"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.582816 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-h6lcv"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.583545 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kbvlq"] Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.608187 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.627989 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.639530 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: E0202 12:11:58.639901 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.139886956 +0000 UTC m=+150.368473819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640173 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-policies\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-client-ca\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640261 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640316 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f149bcb-2741-4754-9b16-a87be939e504-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640334 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv6t2\" (UniqueName: \"kubernetes.io/projected/9f149bcb-2741-4754-9b16-a87be939e504-kube-api-access-gv6t2\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640350 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5256ddb9-8edb-42a8-a041-cb806d849e58-images\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640419 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/59df79f3-cb9f-412c-bc01-7f3376620c17-installation-pull-secrets\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640468 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5256ddb9-8edb-42a8-a041-cb806d849e58-config\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640546 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/459bf85f-e304-4481-a4cb-2c391953d6f5-etcd-client\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640592 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td668\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-kube-api-access-td668\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640638 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-config\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640672 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p47g5\" (UniqueName: \"kubernetes.io/projected/eba09ec1-c855-493b-8ab3-b40d4db28cc1-kube-api-access-p47g5\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640715 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b159dbf-2d80-429e-b473-1260d8f3d67b-config\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640739 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5256ddb9-8edb-42a8-a041-cb806d849e58-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640758 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/459bf85f-e304-4481-a4cb-2c391953d6f5-serving-cert\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640780 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640803 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640825 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/459bf85f-e304-4481-a4cb-2c391953d6f5-audit-dir\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640890 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640921 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.640965 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641000 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-etcd-serving-ca\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641049 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-tls\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641083 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641106 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/459bf85f-e304-4481-a4cb-2c391953d6f5-encryption-config\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641131 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/59df79f3-cb9f-412c-bc01-7f3376620c17-ca-trust-extracted\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641154 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/459bf85f-e304-4481-a4cb-2c391953d6f5-node-pullsecrets\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641220 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641257 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f149bcb-2741-4754-9b16-a87be939e504-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641306 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641363 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-bound-sa-token\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641385 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641402 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f149bcb-2741-4754-9b16-a87be939e504-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641423 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-dir\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641438 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjccf\" (UniqueName: \"kubernetes.io/projected/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-kube-api-access-tjccf\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641458 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggqvl\" (UniqueName: \"kubernetes.io/projected/459bf85f-e304-4481-a4cb-2c391953d6f5-kube-api-access-ggqvl\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641479 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba09ec1-c855-493b-8ab3-b40d4db28cc1-serving-cert\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641495 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-audit\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641511 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641525 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnvrf\" (UniqueName: \"kubernetes.io/projected/9b159dbf-2d80-429e-b473-1260d8f3d67b-kube-api-access-lnvrf\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641546 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9b159dbf-2d80-429e-b473-1260d8f3d67b-auth-proxy-config\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641566 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641581 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9b159dbf-2d80-429e-b473-1260d8f3d67b-machine-approver-tls\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641603 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bkzb\" (UniqueName: \"kubernetes.io/projected/5256ddb9-8edb-42a8-a041-cb806d849e58-kube-api-access-2bkzb\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641631 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dqvd\" (UniqueName: \"kubernetes.io/projected/4157225e-a9ff-4cbe-b56b-43d309030abd-kube-api-access-6dqvd\") pod \"downloads-7954f5f757-bckqh\" (UID: \"4157225e-a9ff-4cbe-b56b-43d309030abd\") " pod="openshift-console/downloads-7954f5f757-bckqh" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641655 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-certificates\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641669 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-trusted-ca\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641684 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641704 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-image-import-ca\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.641756 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-config\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.648144 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.668528 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.687569 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.707906 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.729855 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.742686 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:58 crc kubenswrapper[4846]: E0202 12:11:58.742864 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.242834761 +0000 UTC m=+150.471421634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.742974 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743013 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47a375de-20fc-4cd3-a224-b3fd6cd088c1-service-ca-bundle\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743052 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-serving-cert\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743071 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcx7t\" (UniqueName: \"kubernetes.io/projected/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-kube-api-access-qcx7t\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743090 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3766f763-0e01-466f-85dd-9cbb3a65326d-serving-cert\") pod \"service-ca-operator-777779d784-ff7cw\" (UID: \"3766f763-0e01-466f-85dd-9cbb3a65326d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743126 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pm6v\" (UniqueName: \"kubernetes.io/projected/7d559728-5463-4a87-b4e7-b767dfe50bd7-kube-api-access-6pm6v\") pod \"collect-profiles-29500560-wb4lk\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743145 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9b159dbf-2d80-429e-b473-1260d8f3d67b-machine-approver-tls\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743164 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjmdc\" (UniqueName: \"kubernetes.io/projected/52772605-2b37-4e17-a66a-5456e6e27d47-kube-api-access-tjmdc\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743207 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dqvd\" (UniqueName: \"kubernetes.io/projected/4157225e-a9ff-4cbe-b56b-43d309030abd-kube-api-access-6dqvd\") pod \"downloads-7954f5f757-bckqh\" (UID: \"4157225e-a9ff-4cbe-b56b-43d309030abd\") " pod="openshift-console/downloads-7954f5f757-bckqh" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743224 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-config-volume\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743240 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-certificates\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743255 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743292 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-plugins-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743325 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47a375de-20fc-4cd3-a224-b3fd6cd088c1-metrics-certs\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743362 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxnwd\" (UniqueName: \"kubernetes.io/projected/0d439a22-f822-4714-8135-70a093c340f6-kube-api-access-cxnwd\") pod \"olm-operator-6b444d44fb-s8wzs\" (UID: \"0d439a22-f822-4714-8135-70a093c340f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-policies\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743397 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3880113c-b614-4254-a166-ec2c967a86be-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sf5wg\" (UID: \"3880113c-b614-4254-a166-ec2c967a86be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743412 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zbpnp\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743916 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.743961 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5256ddb9-8edb-42a8-a041-cb806d849e58-images\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744039 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/59df79f3-cb9f-412c-bc01-7f3376620c17-installation-pull-secrets\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744089 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3880113c-b614-4254-a166-ec2c967a86be-proxy-tls\") pod \"machine-config-controller-84d6567774-sf5wg\" (UID: \"3880113c-b614-4254-a166-ec2c967a86be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744132 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-config\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744283 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-trusted-ca\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744308 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/739682cc-1b85-4417-95cf-09940455f8c1-etcd-ca\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744345 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/739682cc-1b85-4417-95cf-09940455f8c1-etcd-service-ca\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744363 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8a05f2a9-e378-4355-b28a-1d0ee567091f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7gk4g\" (UID: \"8a05f2a9-e378-4355-b28a-1d0ee567091f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744380 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6qrw\" (UniqueName: \"kubernetes.io/projected/802f21bc-5a4e-4fa2-9dc1-2b328f1cf924-kube-api-access-r6qrw\") pod \"migrator-59844c95c7-j2hmd\" (UID: \"802f21bc-5a4e-4fa2-9dc1-2b328f1cf924\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744393 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zbpnp\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744427 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-webhook-cert\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744445 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5256ddb9-8edb-42a8-a041-cb806d849e58-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744461 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6fe57969-bb65-43e3-a756-3bc0574727d8-profile-collector-cert\") pod \"catalog-operator-68c6474976-ph4nq\" (UID: \"6fe57969-bb65-43e3-a756-3bc0574727d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744562 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-registration-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744614 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01142a09-9483-44eb-93a3-746603cf21ed-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-h5lq9\" (UID: \"01142a09-9483-44eb-93a3-746603cf21ed\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744657 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7559fe1d-a033-4a6e-8227-ebd2f22dc496-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4zxqh\" (UID: \"7559fe1d-a033-4a6e-8227-ebd2f22dc496\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744674 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6fe57969-bb65-43e3-a756-3bc0574727d8-srv-cert\") pod \"catalog-operator-68c6474976-ph4nq\" (UID: \"6fe57969-bb65-43e3-a756-3bc0574727d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744732 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744753 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8nx7\" (UniqueName: \"kubernetes.io/projected/47a375de-20fc-4cd3-a224-b3fd6cd088c1-kube-api-access-j8nx7\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744804 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-csi-data-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744822 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88863c6c-303f-4f51-95f1-42bcbf8d7701-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hmldg\" (UID: \"88863c6c-303f-4f51-95f1-42bcbf8d7701\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744841 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744913 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-tls\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.744939 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/47a375de-20fc-4cd3-a224-b3fd6cd088c1-stats-auth\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745139 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chgzf\" (UniqueName: \"kubernetes.io/projected/3880113c-b614-4254-a166-ec2c967a86be-kube-api-access-chgzf\") pod \"machine-config-controller-84d6567774-sf5wg\" (UID: \"3880113c-b614-4254-a166-ec2c967a86be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745164 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/459bf85f-e304-4481-a4cb-2c391953d6f5-node-pullsecrets\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745181 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-apiservice-cert\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745218 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745261 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745303 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f1adc015-453c-4718-bb2c-25d6b829f12e-signing-cabundle\") pod \"service-ca-9c57cc56f-tsz82\" (UID: \"f1adc015-453c-4718-bb2c-25d6b829f12e\") " pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745344 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/52772605-2b37-4e17-a66a-5456e6e27d47-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745393 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c8cc0f5-61a4-482e-9413-675bc26d1781-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fdjlt\" (UID: \"8c8cc0f5-61a4-482e-9413-675bc26d1781\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745419 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-trusted-ca\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745462 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjccf\" (UniqueName: \"kubernetes.io/projected/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-kube-api-access-tjccf\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745483 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-metrics-tls\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745510 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-audit\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745557 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/20e43514-4056-412d-a96e-aa0d61e427f0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2ps68\" (UID: \"20e43514-4056-412d-a96e-aa0d61e427f0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745573 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/739682cc-1b85-4417-95cf-09940455f8c1-serving-cert\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745587 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88863c6c-303f-4f51-95f1-42bcbf8d7701-config\") pod \"kube-controller-manager-operator-78b949d7b-hmldg\" (UID: \"88863c6c-303f-4f51-95f1-42bcbf8d7701\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745639 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745669 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-policies\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745868 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnvrf\" (UniqueName: \"kubernetes.io/projected/9b159dbf-2d80-429e-b473-1260d8f3d67b-kube-api-access-lnvrf\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745889 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/47a375de-20fc-4cd3-a224-b3fd6cd088c1-default-certificate\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745959 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9b159dbf-2d80-429e-b473-1260d8f3d67b-auth-proxy-config\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.745982 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/739682cc-1b85-4417-95cf-09940455f8c1-etcd-client\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746030 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20e43514-4056-412d-a96e-aa0d61e427f0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2ps68\" (UID: \"20e43514-4056-412d-a96e-aa0d61e427f0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746050 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bkzb\" (UniqueName: \"kubernetes.io/projected/5256ddb9-8edb-42a8-a041-cb806d849e58-kube-api-access-2bkzb\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746068 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fklg\" (UniqueName: \"kubernetes.io/projected/f1adc015-453c-4718-bb2c-25d6b829f12e-kube-api-access-7fklg\") pod \"service-ca-9c57cc56f-tsz82\" (UID: \"f1adc015-453c-4718-bb2c-25d6b829f12e\") " pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746085 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5d0f31bd-b822-453c-8ed2-12d0e348946c-metrics-tls\") pod \"dns-operator-744455d44c-2472r\" (UID: \"5d0f31bd-b822-453c-8ed2-12d0e348946c\") " pod="openshift-dns-operator/dns-operator-744455d44c-2472r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746111 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-trusted-ca\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746129 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-image-import-ca\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746145 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-config\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746159 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f1adc015-453c-4718-bb2c-25d6b829f12e-signing-key\") pod \"service-ca-9c57cc56f-tsz82\" (UID: \"f1adc015-453c-4718-bb2c-25d6b829f12e\") " pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746174 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/52772605-2b37-4e17-a66a-5456e6e27d47-proxy-tls\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746202 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746220 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d559728-5463-4a87-b4e7-b767dfe50bd7-config-volume\") pod \"collect-profiles-29500560-wb4lk\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746237 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq9vv\" (UniqueName: \"kubernetes.io/projected/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-kube-api-access-vq9vv\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746255 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv6t2\" (UniqueName: \"kubernetes.io/projected/9f149bcb-2741-4754-9b16-a87be939e504-kube-api-access-gv6t2\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746273 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhqqf\" (UniqueName: \"kubernetes.io/projected/6fe57969-bb65-43e3-a756-3bc0574727d8-kube-api-access-qhqqf\") pod \"catalog-operator-68c6474976-ph4nq\" (UID: \"6fe57969-bb65-43e3-a756-3bc0574727d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-client-ca\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746333 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f149bcb-2741-4754-9b16-a87be939e504-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746339 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746350 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e43514-4056-412d-a96e-aa0d61e427f0-config\") pod \"kube-apiserver-operator-766d6c64bb-2ps68\" (UID: \"20e43514-4056-412d-a96e-aa0d61e427f0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746366 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/52772605-2b37-4e17-a66a-5456e6e27d47-images\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746404 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65vw8\" (UniqueName: \"kubernetes.io/projected/b3b18c12-56ac-4a04-bcda-f261a6446f24-kube-api-access-65vw8\") pod \"ingress-canary-jn5pn\" (UID: \"b3b18c12-56ac-4a04-bcda-f261a6446f24\") " pod="openshift-ingress-canary/ingress-canary-jn5pn" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746434 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5256ddb9-8edb-42a8-a041-cb806d849e58-config\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746471 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/459bf85f-e304-4481-a4cb-2c391953d6f5-etcd-client\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746490 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7559fe1d-a033-4a6e-8227-ebd2f22dc496-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4zxqh\" (UID: \"7559fe1d-a033-4a6e-8227-ebd2f22dc496\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746504 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3b18c12-56ac-4a04-bcda-f261a6446f24-cert\") pod \"ingress-canary-jn5pn\" (UID: \"b3b18c12-56ac-4a04-bcda-f261a6446f24\") " pod="openshift-ingress-canary/ingress-canary-jn5pn" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.746783 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: E0202 12:11:58.746925 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.246912329 +0000 UTC m=+150.475499192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.747104 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5256ddb9-8edb-42a8-a041-cb806d849e58-images\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.747109 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.747998 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-client-ca\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.747999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-config\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.748359 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9b159dbf-2d80-429e-b473-1260d8f3d67b-auth-proxy-config\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.749171 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5256ddb9-8edb-42a8-a041-cb806d849e58-config\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.750305 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-config\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.750317 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9b159dbf-2d80-429e-b473-1260d8f3d67b-machine-approver-tls\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.751370 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-image-import-ca\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.752606 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-certificates\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.752711 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.754117 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/459bf85f-e304-4481-a4cb-2c391953d6f5-etcd-client\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.754735 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755077 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-trusted-ca\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755331 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-certs\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755392 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td668\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-kube-api-access-td668\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755417 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p47g5\" (UniqueName: \"kubernetes.io/projected/eba09ec1-c855-493b-8ab3-b40d4db28cc1-kube-api-access-p47g5\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755462 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7559fe1d-a033-4a6e-8227-ebd2f22dc496-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4zxqh\" (UID: \"7559fe1d-a033-4a6e-8227-ebd2f22dc496\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755488 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/459bf85f-e304-4481-a4cb-2c391953d6f5-serving-cert\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755537 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c496r\" (UniqueName: \"kubernetes.io/projected/8a05f2a9-e378-4355-b28a-1d0ee567091f-kube-api-access-c496r\") pod \"multus-admission-controller-857f4d67dd-7gk4g\" (UID: \"8a05f2a9-e378-4355-b28a-1d0ee567091f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755573 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b159dbf-2d80-429e-b473-1260d8f3d67b-config\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755742 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-config\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755772 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755798 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/459bf85f-e304-4481-a4cb-2c391953d6f5-audit-dir\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755823 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6gnh\" (UniqueName: \"kubernetes.io/projected/8c8cc0f5-61a4-482e-9413-675bc26d1781-kube-api-access-z6gnh\") pod \"package-server-manager-789f6589d5-fdjlt\" (UID: \"8c8cc0f5-61a4-482e-9413-675bc26d1781\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755842 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755864 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-node-bootstrap-token\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755865 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-trusted-ca-bundle\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.755993 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3766f763-0e01-466f-85dd-9cbb3a65326d-config\") pod \"service-ca-operator-777779d784-ff7cw\" (UID: \"3766f763-0e01-466f-85dd-9cbb3a65326d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756019 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-etcd-serving-ca\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756044 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-socket-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756069 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d559728-5463-4a87-b4e7-b767dfe50bd7-secret-volume\") pod \"collect-profiles-29500560-wb4lk\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756089 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756114 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756133 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/459bf85f-e304-4481-a4cb-2c391953d6f5-encryption-config\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756153 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-metrics-tls\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756197 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0d439a22-f822-4714-8135-70a093c340f6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s8wzs\" (UID: \"0d439a22-f822-4714-8135-70a093c340f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756218 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01142a09-9483-44eb-93a3-746603cf21ed-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-h5lq9\" (UID: \"01142a09-9483-44eb-93a3-746603cf21ed\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756253 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bcnf\" (UniqueName: \"kubernetes.io/projected/739682cc-1b85-4417-95cf-09940455f8c1-kube-api-access-6bcnf\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756275 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/59df79f3-cb9f-412c-bc01-7f3376620c17-ca-trust-extracted\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756297 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88863c6c-303f-4f51-95f1-42bcbf8d7701-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hmldg\" (UID: \"88863c6c-303f-4f51-95f1-42bcbf8d7701\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756314 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0d439a22-f822-4714-8135-70a093c340f6-srv-cert\") pod \"olm-operator-6b444d44fb-s8wzs\" (UID: \"0d439a22-f822-4714-8135-70a093c340f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756366 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756377 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f149bcb-2741-4754-9b16-a87be939e504-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756399 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/739682cc-1b85-4417-95cf-09940455f8c1-config\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756419 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cthc\" (UniqueName: \"kubernetes.io/projected/3766f763-0e01-466f-85dd-9cbb3a65326d-kube-api-access-6cthc\") pod \"service-ca-operator-777779d784-ff7cw\" (UID: \"3766f763-0e01-466f-85dd-9cbb3a65326d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756444 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-tls\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756538 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmq7h\" (UniqueName: \"kubernetes.io/projected/9a4cbd38-0dde-440f-96a3-4b0bd3c2da81-kube-api-access-mmq7h\") pod \"control-plane-machine-set-operator-78cbb6b69f-6w8zm\" (UID: \"9a4cbd38-0dde-440f-96a3-4b0bd3c2da81\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756562 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxn6g\" (UniqueName: \"kubernetes.io/projected/b3e406b8-473a-40f4-b464-e6f9992ab0c5-kube-api-access-sxn6g\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756598 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756637 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9a4cbd38-0dde-440f-96a3-4b0bd3c2da81-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6w8zm\" (UID: \"9a4cbd38-0dde-440f-96a3-4b0bd3c2da81\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756663 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f149bcb-2741-4754-9b16-a87be939e504-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756687 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72nlv\" (UniqueName: \"kubernetes.io/projected/5d0f31bd-b822-453c-8ed2-12d0e348946c-kube-api-access-72nlv\") pod \"dns-operator-744455d44c-2472r\" (UID: \"5d0f31bd-b822-453c-8ed2-12d0e348946c\") " pod="openshift-dns-operator/dns-operator-744455d44c-2472r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756708 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-bound-sa-token\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756730 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-dir\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggqvl\" (UniqueName: \"kubernetes.io/projected/459bf85f-e304-4481-a4cb-2c391953d6f5-kube-api-access-ggqvl\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756781 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2nmb\" (UniqueName: \"kubernetes.io/projected/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-kube-api-access-x2nmb\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756809 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkxsp\" (UniqueName: \"kubernetes.io/projected/c0652d70-92dc-47c7-926c-d893e14aedfd-kube-api-access-qkxsp\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756837 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/07661a94-bbd8-4b80-842c-d9db04206967-tmpfs\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756849 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/59df79f3-cb9f-412c-bc01-7f3376620c17-ca-trust-extracted\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756857 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-mountpoint-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756919 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba09ec1-c855-493b-8ab3-b40d4db28cc1-serving-cert\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756952 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m795\" (UniqueName: \"kubernetes.io/projected/34dedc1d-10d1-4ca0-b059-99048ba37464-kube-api-access-6m795\") pod \"marketplace-operator-79b997595-zbpnp\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.756987 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghm4p\" (UniqueName: \"kubernetes.io/projected/07661a94-bbd8-4b80-842c-d9db04206967-kube-api-access-ghm4p\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.757021 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftw2p\" (UniqueName: \"kubernetes.io/projected/01142a09-9483-44eb-93a3-746603cf21ed-kube-api-access-ftw2p\") pod \"kube-storage-version-migrator-operator-b67b599dd-h5lq9\" (UID: \"01142a09-9483-44eb-93a3-746603cf21ed\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.757156 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b159dbf-2d80-429e-b473-1260d8f3d67b-config\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.757533 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-etcd-serving-ca\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.758334 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.758876 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/459bf85f-e304-4481-a4cb-2c391953d6f5-audit\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.759061 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5256ddb9-8edb-42a8-a041-cb806d849e58-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.759121 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/459bf85f-e304-4481-a4cb-2c391953d6f5-node-pullsecrets\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.760730 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f149bcb-2741-4754-9b16-a87be939e504-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.760874 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.761404 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.761566 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/59df79f3-cb9f-412c-bc01-7f3376620c17-installation-pull-secrets\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.761574 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba09ec1-c855-493b-8ab3-b40d4db28cc1-serving-cert\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.761804 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/459bf85f-e304-4481-a4cb-2c391953d6f5-audit-dir\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.761905 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-dir\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.762367 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/459bf85f-e304-4481-a4cb-2c391953d6f5-encryption-config\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.763222 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.764217 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.764938 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.765426 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f149bcb-2741-4754-9b16-a87be939e504-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.767497 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.768540 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/459bf85f-e304-4481-a4cb-2c391953d6f5-serving-cert\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.777686 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.788275 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.807964 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.828991 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.847497 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.858702 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:58 crc kubenswrapper[4846]: E0202 12:11:58.858882 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.358862065 +0000 UTC m=+150.587448928 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.858916 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-trusted-ca\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.858941 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/739682cc-1b85-4417-95cf-09940455f8c1-etcd-ca\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.858958 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/739682cc-1b85-4417-95cf-09940455f8c1-etcd-service-ca\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.858979 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8a05f2a9-e378-4355-b28a-1d0ee567091f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7gk4g\" (UID: \"8a05f2a9-e378-4355-b28a-1d0ee567091f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859009 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6qrw\" (UniqueName: \"kubernetes.io/projected/802f21bc-5a4e-4fa2-9dc1-2b328f1cf924-kube-api-access-r6qrw\") pod \"migrator-59844c95c7-j2hmd\" (UID: \"802f21bc-5a4e-4fa2-9dc1-2b328f1cf924\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859031 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zbpnp\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859055 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-webhook-cert\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859078 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6fe57969-bb65-43e3-a756-3bc0574727d8-profile-collector-cert\") pod \"catalog-operator-68c6474976-ph4nq\" (UID: \"6fe57969-bb65-43e3-a756-3bc0574727d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859099 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-registration-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859119 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01142a09-9483-44eb-93a3-746603cf21ed-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-h5lq9\" (UID: \"01142a09-9483-44eb-93a3-746603cf21ed\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859139 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7559fe1d-a033-4a6e-8227-ebd2f22dc496-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4zxqh\" (UID: \"7559fe1d-a033-4a6e-8227-ebd2f22dc496\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859161 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6fe57969-bb65-43e3-a756-3bc0574727d8-srv-cert\") pod \"catalog-operator-68c6474976-ph4nq\" (UID: \"6fe57969-bb65-43e3-a756-3bc0574727d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859187 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8nx7\" (UniqueName: \"kubernetes.io/projected/47a375de-20fc-4cd3-a224-b3fd6cd088c1-kube-api-access-j8nx7\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859209 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-csi-data-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859231 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88863c6c-303f-4f51-95f1-42bcbf8d7701-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hmldg\" (UID: \"88863c6c-303f-4f51-95f1-42bcbf8d7701\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859265 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/47a375de-20fc-4cd3-a224-b3fd6cd088c1-stats-auth\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859293 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chgzf\" (UniqueName: \"kubernetes.io/projected/3880113c-b614-4254-a166-ec2c967a86be-kube-api-access-chgzf\") pod \"machine-config-controller-84d6567774-sf5wg\" (UID: \"3880113c-b614-4254-a166-ec2c967a86be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859318 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-apiservice-cert\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859351 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f1adc015-453c-4718-bb2c-25d6b829f12e-signing-cabundle\") pod \"service-ca-9c57cc56f-tsz82\" (UID: \"f1adc015-453c-4718-bb2c-25d6b829f12e\") " pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859373 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/52772605-2b37-4e17-a66a-5456e6e27d47-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c8cc0f5-61a4-482e-9413-675bc26d1781-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fdjlt\" (UID: \"8c8cc0f5-61a4-482e-9413-675bc26d1781\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859408 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-trusted-ca\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859430 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-metrics-tls\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/20e43514-4056-412d-a96e-aa0d61e427f0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2ps68\" (UID: \"20e43514-4056-412d-a96e-aa0d61e427f0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859464 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/739682cc-1b85-4417-95cf-09940455f8c1-serving-cert\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859483 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88863c6c-303f-4f51-95f1-42bcbf8d7701-config\") pod \"kube-controller-manager-operator-78b949d7b-hmldg\" (UID: \"88863c6c-303f-4f51-95f1-42bcbf8d7701\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859517 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/47a375de-20fc-4cd3-a224-b3fd6cd088c1-default-certificate\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859549 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/739682cc-1b85-4417-95cf-09940455f8c1-etcd-client\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859604 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20e43514-4056-412d-a96e-aa0d61e427f0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2ps68\" (UID: \"20e43514-4056-412d-a96e-aa0d61e427f0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859657 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fklg\" (UniqueName: \"kubernetes.io/projected/f1adc015-453c-4718-bb2c-25d6b829f12e-kube-api-access-7fklg\") pod \"service-ca-9c57cc56f-tsz82\" (UID: \"f1adc015-453c-4718-bb2c-25d6b829f12e\") " pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859680 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5d0f31bd-b822-453c-8ed2-12d0e348946c-metrics-tls\") pod \"dns-operator-744455d44c-2472r\" (UID: \"5d0f31bd-b822-453c-8ed2-12d0e348946c\") " pod="openshift-dns-operator/dns-operator-744455d44c-2472r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859703 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f1adc015-453c-4718-bb2c-25d6b829f12e-signing-key\") pod \"service-ca-9c57cc56f-tsz82\" (UID: \"f1adc015-453c-4718-bb2c-25d6b829f12e\") " pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859728 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/52772605-2b37-4e17-a66a-5456e6e27d47-proxy-tls\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859773 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859795 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d559728-5463-4a87-b4e7-b767dfe50bd7-config-volume\") pod \"collect-profiles-29500560-wb4lk\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859813 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq9vv\" (UniqueName: \"kubernetes.io/projected/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-kube-api-access-vq9vv\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859836 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhqqf\" (UniqueName: \"kubernetes.io/projected/6fe57969-bb65-43e3-a756-3bc0574727d8-kube-api-access-qhqqf\") pod \"catalog-operator-68c6474976-ph4nq\" (UID: \"6fe57969-bb65-43e3-a756-3bc0574727d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859856 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e43514-4056-412d-a96e-aa0d61e427f0-config\") pod \"kube-apiserver-operator-766d6c64bb-2ps68\" (UID: \"20e43514-4056-412d-a96e-aa0d61e427f0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859874 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/52772605-2b37-4e17-a66a-5456e6e27d47-images\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859893 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65vw8\" (UniqueName: \"kubernetes.io/projected/b3b18c12-56ac-4a04-bcda-f261a6446f24-kube-api-access-65vw8\") pod \"ingress-canary-jn5pn\" (UID: \"b3b18c12-56ac-4a04-bcda-f261a6446f24\") " pod="openshift-ingress-canary/ingress-canary-jn5pn" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859915 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7559fe1d-a033-4a6e-8227-ebd2f22dc496-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4zxqh\" (UID: \"7559fe1d-a033-4a6e-8227-ebd2f22dc496\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859933 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3b18c12-56ac-4a04-bcda-f261a6446f24-cert\") pod \"ingress-canary-jn5pn\" (UID: \"b3b18c12-56ac-4a04-bcda-f261a6446f24\") " pod="openshift-ingress-canary/ingress-canary-jn5pn" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859953 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-certs\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859981 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7559fe1d-a033-4a6e-8227-ebd2f22dc496-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4zxqh\" (UID: \"7559fe1d-a033-4a6e-8227-ebd2f22dc496\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860000 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c496r\" (UniqueName: \"kubernetes.io/projected/8a05f2a9-e378-4355-b28a-1d0ee567091f-kube-api-access-c496r\") pod \"multus-admission-controller-857f4d67dd-7gk4g\" (UID: \"8a05f2a9-e378-4355-b28a-1d0ee567091f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860023 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-config\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860021 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-csi-data-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860040 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6gnh\" (UniqueName: \"kubernetes.io/projected/8c8cc0f5-61a4-482e-9413-675bc26d1781-kube-api-access-z6gnh\") pod \"package-server-manager-789f6589d5-fdjlt\" (UID: \"8c8cc0f5-61a4-482e-9413-675bc26d1781\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-node-bootstrap-token\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860200 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3766f763-0e01-466f-85dd-9cbb3a65326d-config\") pod \"service-ca-operator-777779d784-ff7cw\" (UID: \"3766f763-0e01-466f-85dd-9cbb3a65326d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860222 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-socket-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860274 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d559728-5463-4a87-b4e7-b767dfe50bd7-secret-volume\") pod \"collect-profiles-29500560-wb4lk\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860300 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860355 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-metrics-tls\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:11:58 crc kubenswrapper[4846]: E0202 12:11:58.860383 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.360369459 +0000 UTC m=+150.588956322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860416 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0d439a22-f822-4714-8135-70a093c340f6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s8wzs\" (UID: \"0d439a22-f822-4714-8135-70a093c340f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860441 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01142a09-9483-44eb-93a3-746603cf21ed-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-h5lq9\" (UID: \"01142a09-9483-44eb-93a3-746603cf21ed\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.860476 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-socket-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.859925 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-registration-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.861130 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/52772605-2b37-4e17-a66a-5456e6e27d47-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.861308 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-trusted-ca\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.861440 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88863c6c-303f-4f51-95f1-42bcbf8d7701-config\") pod \"kube-controller-manager-operator-78b949d7b-hmldg\" (UID: \"88863c6c-303f-4f51-95f1-42bcbf8d7701\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.861572 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-config\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.861644 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d559728-5463-4a87-b4e7-b767dfe50bd7-config-volume\") pod \"collect-profiles-29500560-wb4lk\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.861855 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e43514-4056-412d-a96e-aa0d61e427f0-config\") pod \"kube-apiserver-operator-766d6c64bb-2ps68\" (UID: \"20e43514-4056-412d-a96e-aa0d61e427f0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862161 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bcnf\" (UniqueName: \"kubernetes.io/projected/739682cc-1b85-4417-95cf-09940455f8c1-kube-api-access-6bcnf\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862207 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88863c6c-303f-4f51-95f1-42bcbf8d7701-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hmldg\" (UID: \"88863c6c-303f-4f51-95f1-42bcbf8d7701\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862232 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0d439a22-f822-4714-8135-70a093c340f6-srv-cert\") pod \"olm-operator-6b444d44fb-s8wzs\" (UID: \"0d439a22-f822-4714-8135-70a093c340f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/739682cc-1b85-4417-95cf-09940455f8c1-config\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862281 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cthc\" (UniqueName: \"kubernetes.io/projected/3766f763-0e01-466f-85dd-9cbb3a65326d-kube-api-access-6cthc\") pod \"service-ca-operator-777779d784-ff7cw\" (UID: \"3766f763-0e01-466f-85dd-9cbb3a65326d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862305 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmq7h\" (UniqueName: \"kubernetes.io/projected/9a4cbd38-0dde-440f-96a3-4b0bd3c2da81-kube-api-access-mmq7h\") pod \"control-plane-machine-set-operator-78cbb6b69f-6w8zm\" (UID: \"9a4cbd38-0dde-440f-96a3-4b0bd3c2da81\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862327 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxn6g\" (UniqueName: \"kubernetes.io/projected/b3e406b8-473a-40f4-b464-e6f9992ab0c5-kube-api-access-sxn6g\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862359 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9a4cbd38-0dde-440f-96a3-4b0bd3c2da81-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6w8zm\" (UID: \"9a4cbd38-0dde-440f-96a3-4b0bd3c2da81\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862387 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72nlv\" (UniqueName: \"kubernetes.io/projected/5d0f31bd-b822-453c-8ed2-12d0e348946c-kube-api-access-72nlv\") pod \"dns-operator-744455d44c-2472r\" (UID: \"5d0f31bd-b822-453c-8ed2-12d0e348946c\") " pod="openshift-dns-operator/dns-operator-744455d44c-2472r" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862429 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2nmb\" (UniqueName: \"kubernetes.io/projected/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-kube-api-access-x2nmb\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862690 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkxsp\" (UniqueName: \"kubernetes.io/projected/c0652d70-92dc-47c7-926c-d893e14aedfd-kube-api-access-qkxsp\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/07661a94-bbd8-4b80-842c-d9db04206967-tmpfs\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862759 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-mountpoint-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862785 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m795\" (UniqueName: \"kubernetes.io/projected/34dedc1d-10d1-4ca0-b059-99048ba37464-kube-api-access-6m795\") pod \"marketplace-operator-79b997595-zbpnp\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862809 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghm4p\" (UniqueName: \"kubernetes.io/projected/07661a94-bbd8-4b80-842c-d9db04206967-kube-api-access-ghm4p\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862833 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftw2p\" (UniqueName: \"kubernetes.io/projected/01142a09-9483-44eb-93a3-746603cf21ed-kube-api-access-ftw2p\") pod \"kube-storage-version-migrator-operator-b67b599dd-h5lq9\" (UID: \"01142a09-9483-44eb-93a3-746603cf21ed\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862880 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47a375de-20fc-4cd3-a224-b3fd6cd088c1-service-ca-bundle\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862904 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-serving-cert\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862932 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcx7t\" (UniqueName: \"kubernetes.io/projected/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-kube-api-access-qcx7t\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3766f763-0e01-466f-85dd-9cbb3a65326d-serving-cert\") pod \"service-ca-operator-777779d784-ff7cw\" (UID: \"3766f763-0e01-466f-85dd-9cbb3a65326d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.862985 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pm6v\" (UniqueName: \"kubernetes.io/projected/7d559728-5463-4a87-b4e7-b767dfe50bd7-kube-api-access-6pm6v\") pod \"collect-profiles-29500560-wb4lk\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863011 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjmdc\" (UniqueName: \"kubernetes.io/projected/52772605-2b37-4e17-a66a-5456e6e27d47-kube-api-access-tjmdc\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863045 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-config-volume\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863069 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-plugins-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863103 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47a375de-20fc-4cd3-a224-b3fd6cd088c1-metrics-certs\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863132 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxnwd\" (UniqueName: \"kubernetes.io/projected/0d439a22-f822-4714-8135-70a093c340f6-kube-api-access-cxnwd\") pod \"olm-operator-6b444d44fb-s8wzs\" (UID: \"0d439a22-f822-4714-8135-70a093c340f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863151 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/07661a94-bbd8-4b80-842c-d9db04206967-tmpfs\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863225 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20e43514-4056-412d-a96e-aa0d61e427f0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2ps68\" (UID: \"20e43514-4056-412d-a96e-aa0d61e427f0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3880113c-b614-4254-a166-ec2c967a86be-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sf5wg\" (UID: \"3880113c-b614-4254-a166-ec2c967a86be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863289 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zbpnp\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863322 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3880113c-b614-4254-a166-ec2c967a86be-proxy-tls\") pod \"machine-config-controller-84d6567774-sf5wg\" (UID: \"3880113c-b614-4254-a166-ec2c967a86be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863097 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-mountpoint-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863529 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b3e406b8-473a-40f4-b464-e6f9992ab0c5-plugins-dir\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.863949 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3880113c-b614-4254-a166-ec2c967a86be-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sf5wg\" (UID: \"3880113c-b614-4254-a166-ec2c967a86be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.866084 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6fe57969-bb65-43e3-a756-3bc0574727d8-profile-collector-cert\") pod \"catalog-operator-68c6474976-ph4nq\" (UID: \"6fe57969-bb65-43e3-a756-3bc0574727d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.866223 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d559728-5463-4a87-b4e7-b767dfe50bd7-secret-volume\") pod \"collect-profiles-29500560-wb4lk\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.872565 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.872812 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88863c6c-303f-4f51-95f1-42bcbf8d7701-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hmldg\" (UID: \"88863c6c-303f-4f51-95f1-42bcbf8d7701\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.873411 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-serving-cert\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.874650 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0d439a22-f822-4714-8135-70a093c340f6-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s8wzs\" (UID: \"0d439a22-f822-4714-8135-70a093c340f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.888555 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.893985 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f1adc015-453c-4718-bb2c-25d6b829f12e-signing-key\") pod \"service-ca-9c57cc56f-tsz82\" (UID: \"f1adc015-453c-4718-bb2c-25d6b829f12e\") " pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.908401 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.911555 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f1adc015-453c-4718-bb2c-25d6b829f12e-signing-cabundle\") pod \"service-ca-9c57cc56f-tsz82\" (UID: \"f1adc015-453c-4718-bb2c-25d6b829f12e\") " pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.927600 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.948007 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.964204 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:58 crc kubenswrapper[4846]: E0202 12:11:58.964417 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.464387385 +0000 UTC m=+150.692974248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.965112 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:58 crc kubenswrapper[4846]: E0202 12:11:58.965703 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.465692023 +0000 UTC m=+150.694278936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.967961 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.987444 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 02 12:11:58 crc kubenswrapper[4846]: I0202 12:11:58.997809 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47a375de-20fc-4cd3-a224-b3fd6cd088c1-metrics-certs\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.008367 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.015012 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47a375de-20fc-4cd3-a224-b3fd6cd088c1-service-ca-bundle\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.027669 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.034312 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/47a375de-20fc-4cd3-a224-b3fd6cd088c1-default-certificate\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.047944 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.052990 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/47a375de-20fc-4cd3-a224-b3fd6cd088c1-stats-auth\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.066171 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.066297 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.566273191 +0000 UTC m=+150.794860054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.067118 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.067760 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.067955 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.5676113 +0000 UTC m=+150.796198173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.088700 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.093953 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7559fe1d-a033-4a6e-8227-ebd2f22dc496-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4zxqh\" (UID: \"7559fe1d-a033-4a6e-8227-ebd2f22dc496\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.108016 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.128510 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.131929 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7559fe1d-a033-4a6e-8227-ebd2f22dc496-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4zxqh\" (UID: \"7559fe1d-a033-4a6e-8227-ebd2f22dc496\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.148848 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.167809 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.167885 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.667866788 +0000 UTC m=+150.896453651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.168142 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.168439 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.168801 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.668784974 +0000 UTC m=+150.897371847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.188831 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.207586 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.228232 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.256643 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.262694 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zbpnp\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.267882 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.269223 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.269426 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.769396243 +0000 UTC m=+150.997983156 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.270198 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.270784 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.770746952 +0000 UTC m=+150.999333855 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.276731 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3766f763-0e01-466f-85dd-9cbb3a65326d-serving-cert\") pod \"service-ca-operator-777779d784-ff7cw\" (UID: \"3766f763-0e01-466f-85dd-9cbb3a65326d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.288317 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.307709 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.311006 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3766f763-0e01-466f-85dd-9cbb3a65326d-config\") pod \"service-ca-operator-777779d784-ff7cw\" (UID: \"3766f763-0e01-466f-85dd-9cbb3a65326d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.327756 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.348586 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.357338 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zbpnp\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.368544 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.371067 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.371277 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.871249157 +0000 UTC m=+151.099836030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.371485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.371915 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.871900895 +0000 UTC m=+151.100487758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.376020 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01142a09-9483-44eb-93a3-746603cf21ed-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-h5lq9\" (UID: \"01142a09-9483-44eb-93a3-746603cf21ed\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.387442 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.391037 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01142a09-9483-44eb-93a3-746603cf21ed-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-h5lq9\" (UID: \"01142a09-9483-44eb-93a3-746603cf21ed\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.408283 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.428383 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.448093 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.467770 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.472656 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.472947 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.972904906 +0000 UTC m=+151.201491839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.473484 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.475111 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:11:59.975087309 +0000 UTC m=+151.203674212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.478150 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3880113c-b614-4254-a166-ec2c967a86be-proxy-tls\") pod \"machine-config-controller-84d6567774-sf5wg\" (UID: \"3880113c-b614-4254-a166-ec2c967a86be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.486385 4846 request.go:700] Waited for 1.015905656s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-controller-dockercfg-c2lfx&limit=500&resourceVersion=0 Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.487800 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.508190 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.513484 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/739682cc-1b85-4417-95cf-09940455f8c1-config\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.528748 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.541012 4846 secret.go:188] Couldn't get secret openshift-console/console-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.541086 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert podName:af51b9e4-0ee3-4bf9-93e5-27b4039be121 nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.041063716 +0000 UTC m=+151.269650589 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-serving-cert" (UniqueName: "kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert") pod "console-f9d7485db-79skc" (UID: "af51b9e4-0ee3-4bf9-93e5-27b4039be121") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.547826 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.553952 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/739682cc-1b85-4417-95cf-09940455f8c1-serving-cert\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.583095 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.583236 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.083188593 +0000 UTC m=+151.311775476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.583710 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.584174 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.084161112 +0000 UTC m=+151.312747985 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.584782 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.588034 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.589552 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/739682cc-1b85-4417-95cf-09940455f8c1-etcd-ca\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.593334 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/739682cc-1b85-4417-95cf-09940455f8c1-etcd-client\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.607570 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.610435 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/739682cc-1b85-4417-95cf-09940455f8c1-etcd-service-ca\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.633655 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.648262 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.667556 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.676630 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c8cc0f5-61a4-482e-9413-675bc26d1781-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fdjlt\" (UID: \"8c8cc0f5-61a4-482e-9413-675bc26d1781\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.685010 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.685253 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.185220753 +0000 UTC m=+151.413807666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.685319 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.685984 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.185952125 +0000 UTC m=+151.414539038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.687840 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.707972 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.727702 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.747770 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.768238 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.773047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5d0f31bd-b822-453c-8ed2-12d0e348946c-metrics-tls\") pod \"dns-operator-744455d44c-2472r\" (UID: \"5d0f31bd-b822-453c-8ed2-12d0e348946c\") " pod="openshift-dns-operator/dns-operator-744455d44c-2472r" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.787480 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.787561 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.287543131 +0000 UTC m=+151.516130004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.788402 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.788543 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.788965 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.288950721 +0000 UTC m=+151.517537584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.792387 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8a05f2a9-e378-4355-b28a-1d0ee567091f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7gk4g\" (UID: \"8a05f2a9-e378-4355-b28a-1d0ee567091f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.808457 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.828342 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.847484 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.853909 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-metrics-tls\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.859530 4846 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.859607 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-webhook-cert podName:07661a94-bbd8-4b80-842c-d9db04206967 nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.359583894 +0000 UTC m=+151.588170767 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-webhook-cert") pod "packageserver-d55dfcdfc-p8mks" (UID: "07661a94-bbd8-4b80-842c-d9db04206967") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.859805 4846 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.859899 4846 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.859915 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fe57969-bb65-43e3-a756-3bc0574727d8-srv-cert podName:6fe57969-bb65-43e3-a756-3bc0574727d8 nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.359894072 +0000 UTC m=+151.588480935 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/6fe57969-bb65-43e3-a756-3bc0574727d8-srv-cert") pod "catalog-operator-68c6474976-ph4nq" (UID: "6fe57969-bb65-43e3-a756-3bc0574727d8") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.859993 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-apiservice-cert podName:07661a94-bbd8-4b80-842c-d9db04206967 nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.359977315 +0000 UTC m=+151.588564268 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-apiservice-cert") pod "packageserver-d55dfcdfc-p8mks" (UID: "07661a94-bbd8-4b80-842c-d9db04206967") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860724 4846 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860783 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/52772605-2b37-4e17-a66a-5456e6e27d47-proxy-tls podName:52772605-2b37-4e17-a66a-5456e6e27d47 nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.360768808 +0000 UTC m=+151.589355681 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/52772605-2b37-4e17-a66a-5456e6e27d47-proxy-tls") pod "machine-config-operator-74547568cd-n7rfv" (UID: "52772605-2b37-4e17-a66a-5456e6e27d47") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860783 4846 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860811 4846 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860829 4846 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860842 4846 configmap.go:193] Couldn't get configMap openshift-ingress-operator/trusted-ca: failed to sync configmap cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860848 4846 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860801 4846 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860848 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-metrics-tls podName:618d005a-d18f-4ad9-92ba-cb1a97a9cebc nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.36083085 +0000 UTC m=+151.589417793 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-metrics-tls") pod "dns-default-kbvlq" (UID: "618d005a-d18f-4ad9-92ba-cb1a97a9cebc") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860961 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b3b18c12-56ac-4a04-bcda-f261a6446f24-cert podName:b3b18c12-56ac-4a04-bcda-f261a6446f24 nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.360939483 +0000 UTC m=+151.589526446 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b3b18c12-56ac-4a04-bcda-f261a6446f24-cert") pod "ingress-canary-jn5pn" (UID: "b3b18c12-56ac-4a04-bcda-f261a6446f24") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860982 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/52772605-2b37-4e17-a66a-5456e6e27d47-images podName:52772605-2b37-4e17-a66a-5456e6e27d47 nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.360972234 +0000 UTC m=+151.589559187 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/52772605-2b37-4e17-a66a-5456e6e27d47-images") pod "machine-config-operator-74547568cd-n7rfv" (UID: "52772605-2b37-4e17-a66a-5456e6e27d47") : failed to sync configmap cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.860997 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-trusted-ca podName:fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.360989654 +0000 UTC m=+151.589576617 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca" (UniqueName: "kubernetes.io/configmap/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-trusted-ca") pod "ingress-operator-5b745b69d9-wbtlv" (UID: "fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac") : failed to sync configmap cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.861201 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-node-bootstrap-token podName:c0652d70-92dc-47c7-926c-d893e14aedfd nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.36119115 +0000 UTC m=+151.589778103 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-node-bootstrap-token") pod "machine-config-server-kjwnm" (UID: "c0652d70-92dc-47c7-926c-d893e14aedfd") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.861223 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-certs podName:c0652d70-92dc-47c7-926c-d893e14aedfd nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.361216051 +0000 UTC m=+151.589803004 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-certs") pod "machine-config-server-kjwnm" (UID: "c0652d70-92dc-47c7-926c-d893e14aedfd") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.862896 4846 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.862928 4846 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.862974 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a4cbd38-0dde-440f-96a3-4b0bd3c2da81-control-plane-machine-set-operator-tls podName:9a4cbd38-0dde-440f-96a3-4b0bd3c2da81 nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.362958111 +0000 UTC m=+151.591545034 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/9a4cbd38-0dde-440f-96a3-4b0bd3c2da81-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-6w8zm" (UID: "9a4cbd38-0dde-440f-96a3-4b0bd3c2da81") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.863005 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d439a22-f822-4714-8135-70a093c340f6-srv-cert podName:0d439a22-f822-4714-8135-70a093c340f6 nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.362993272 +0000 UTC m=+151.591580255 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/0d439a22-f822-4714-8135-70a093c340f6-srv-cert") pod "olm-operator-6b444d44fb-s8wzs" (UID: "0d439a22-f822-4714-8135-70a093c340f6") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.864059 4846 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.864117 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-config-volume podName:618d005a-d18f-4ad9-92ba-cb1a97a9cebc nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.364097464 +0000 UTC m=+151.592684407 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-config-volume") pod "dns-default-kbvlq" (UID: "618d005a-d18f-4ad9-92ba-cb1a97a9cebc") : failed to sync configmap cache: timed out waiting for the condition Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.868559 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.889212 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.889365 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.389339164 +0000 UTC m=+151.617926047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.889515 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.889965 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.389948081 +0000 UTC m=+151.618534944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.894976 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.907473 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.927684 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.948155 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.967808 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.987452 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.991440 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.991740 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.491717183 +0000 UTC m=+151.720304046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:11:59 crc kubenswrapper[4846]: I0202 12:11:59.991898 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:11:59 crc kubenswrapper[4846]: E0202 12:11:59.992594 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.492572148 +0000 UTC m=+151.721159041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.009482 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.028478 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.048347 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.068011 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.088891 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.093858 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.094023 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.593998929 +0000 UTC m=+151.822585802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.094251 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.094585 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.095055 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.59503723 +0000 UTC m=+151.823624173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.108399 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.128088 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.148384 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.168984 4846 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.187857 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.196030 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.196269 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.696241015 +0000 UTC m=+151.924827888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.196892 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.197255 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.697239984 +0000 UTC m=+151.925826857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.208341 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.244395 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zftj\" (UniqueName: \"kubernetes.io/projected/614c989a-6be2-4824-a3e5-d810fd94e39b-kube-api-access-7zftj\") pod \"openshift-apiserver-operator-796bbdcf4f-ml2fm\" (UID: \"614c989a-6be2-4824-a3e5-d810fd94e39b\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.270047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzzzd\" (UniqueName: \"kubernetes.io/projected/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-kube-api-access-mzzzd\") pod \"route-controller-manager-6576b87f9c-cfv6g\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.284833 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.289467 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h2p8\" (UniqueName: \"kubernetes.io/projected/a62db6be-7478-41ce-b1a7-fbf3f24967e2-kube-api-access-9h2p8\") pod \"openshift-config-operator-7777fb866f-5glmk\" (UID: \"a62db6be-7478-41ce-b1a7-fbf3f24967e2\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.298304 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.298479 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.79845273 +0000 UTC m=+152.027039593 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.298751 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.299027 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.799019857 +0000 UTC m=+152.027606720 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.311725 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddm9z\" (UniqueName: \"kubernetes.io/projected/e94b3ca8-f11a-4fee-8ef1-05a56d233593-kube-api-access-ddm9z\") pod \"apiserver-7bbb656c7d-2984r\" (UID: \"e94b3ca8-f11a-4fee-8ef1-05a56d233593\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.322093 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-478m7\" (UniqueName: \"kubernetes.io/projected/af51b9e4-0ee3-4bf9-93e5-27b4039be121-kube-api-access-478m7\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.322403 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.343174 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htdln\" (UniqueName: \"kubernetes.io/projected/2c142082-61b2-4ea2-9d3b-8f9bae616178-kube-api-access-htdln\") pod \"authentication-operator-69f744f599-6kzjp\" (UID: \"2c142082-61b2-4ea2-9d3b-8f9bae616178\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.366608 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkncs\" (UniqueName: \"kubernetes.io/projected/8abc246a-7ca3-4550-ba36-491f830e5419-kube-api-access-hkncs\") pod \"cluster-samples-operator-665b6dd947-px4bj\" (UID: \"8abc246a-7ca3-4550-ba36-491f830e5419\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.387060 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkmx6\" (UniqueName: \"kubernetes.io/projected/12c5b3f2-6d12-4f68-8013-fd5a555b7cfc-kube-api-access-gkmx6\") pod \"openshift-controller-manager-operator-756b6f6bc6-2jvrx\" (UID: \"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.387897 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.400958 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401178 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-node-bootstrap-token\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401208 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-metrics-tls\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401241 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0d439a22-f822-4714-8135-70a093c340f6-srv-cert\") pod \"olm-operator-6b444d44fb-s8wzs\" (UID: \"0d439a22-f822-4714-8135-70a093c340f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401280 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9a4cbd38-0dde-440f-96a3-4b0bd3c2da81-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6w8zm\" (UID: \"9a4cbd38-0dde-440f-96a3-4b0bd3c2da81\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401570 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-config-volume\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401622 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-webhook-cert\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401660 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6fe57969-bb65-43e3-a756-3bc0574727d8-srv-cert\") pod \"catalog-operator-68c6474976-ph4nq\" (UID: \"6fe57969-bb65-43e3-a756-3bc0574727d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401695 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-apiservice-cert\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401712 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-trusted-ca\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/52772605-2b37-4e17-a66a-5456e6e27d47-proxy-tls\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401801 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/52772605-2b37-4e17-a66a-5456e6e27d47-images\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.403588 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:00.903558698 +0000 UTC m=+152.132145581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.401822 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3b18c12-56ac-4a04-bcda-f261a6446f24-cert\") pod \"ingress-canary-jn5pn\" (UID: \"b3b18c12-56ac-4a04-bcda-f261a6446f24\") " pod="openshift-ingress-canary/ingress-canary-jn5pn" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.403978 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-certs\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.404271 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/52772605-2b37-4e17-a66a-5456e6e27d47-images\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.404771 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-config-volume\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.406841 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6fe57969-bb65-43e3-a756-3bc0574727d8-srv-cert\") pod \"catalog-operator-68c6474976-ph4nq\" (UID: \"6fe57969-bb65-43e3-a756-3bc0574727d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.407223 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0d439a22-f822-4714-8135-70a093c340f6-srv-cert\") pod \"olm-operator-6b444d44fb-s8wzs\" (UID: \"0d439a22-f822-4714-8135-70a093c340f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.408184 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9a4cbd38-0dde-440f-96a3-4b0bd3c2da81-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6w8zm\" (UID: \"9a4cbd38-0dde-440f-96a3-4b0bd3c2da81\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.408575 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.408925 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/52772605-2b37-4e17-a66a-5456e6e27d47-proxy-tls\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.410029 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b3b18c12-56ac-4a04-bcda-f261a6446f24-cert\") pod \"ingress-canary-jn5pn\" (UID: \"b3b18c12-56ac-4a04-bcda-f261a6446f24\") " pod="openshift-ingress-canary/ingress-canary-jn5pn" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.410184 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-trusted-ca\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.411277 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-apiservice-cert\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.412071 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/07661a94-bbd8-4b80-842c-d9db04206967-webhook-cert\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.430045 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.435987 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.436904 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-metrics-tls\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.455604 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.459344 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.461639 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm"] Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.470937 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.472346 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-certs\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.479155 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.479212 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.489587 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.495958 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c0652d70-92dc-47c7-926c-d893e14aedfd-node-bootstrap-token\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.498749 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5glmk"] Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.505891 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.506147 4846 request.go:700] Waited for 1.938291872s due to client-side throttling, not priority and fairness, request: PATCH:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/pods/machine-approver-56656f9798-ghkkq/status Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.506366 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.006332949 +0000 UTC m=+152.234919822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: W0202 12:12:00.512465 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda62db6be_7478_41ce_b1a7_fbf3f24967e2.slice/crio-e8a15946e3577b574b8f2a28bfce030ee9c066a09e3499c271d8c6631842ad12 WatchSource:0}: Error finding container e8a15946e3577b574b8f2a28bfce030ee9c066a09e3499c271d8c6631842ad12: Status 404 returned error can't find the container with id e8a15946e3577b574b8f2a28bfce030ee9c066a09e3499c271d8c6631842ad12 Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.545148 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnvrf\" (UniqueName: \"kubernetes.io/projected/9b159dbf-2d80-429e-b473-1260d8f3d67b-kube-api-access-lnvrf\") pod \"machine-approver-56656f9798-ghkkq\" (UID: \"9b159dbf-2d80-429e-b473-1260d8f3d67b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.554317 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.567213 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv6t2\" (UniqueName: \"kubernetes.io/projected/9f149bcb-2741-4754-9b16-a87be939e504-kube-api-access-gv6t2\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.575086 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.582224 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f149bcb-2741-4754-9b16-a87be939e504-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-bmwts\" (UID: \"9f149bcb-2741-4754-9b16-a87be939e504\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.592175 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.606695 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.607118 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.107096912 +0000 UTC m=+152.335683775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.607219 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.607614 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.107602707 +0000 UTC m=+152.336189570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.613801 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.614794 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bkzb\" (UniqueName: \"kubernetes.io/projected/5256ddb9-8edb-42a8-a041-cb806d849e58-kube-api-access-2bkzb\") pod \"machine-api-operator-5694c8668f-7qdsx\" (UID: \"5256ddb9-8edb-42a8-a041-cb806d849e58\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.626314 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dqvd\" (UniqueName: \"kubernetes.io/projected/4157225e-a9ff-4cbe-b56b-43d309030abd-kube-api-access-6dqvd\") pod \"downloads-7954f5f757-bckqh\" (UID: \"4157225e-a9ff-4cbe-b56b-43d309030abd\") " pod="openshift-console/downloads-7954f5f757-bckqh" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.649374 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjccf\" (UniqueName: \"kubernetes.io/projected/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-kube-api-access-tjccf\") pod \"oauth-openshift-558db77b4-q57ps\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.672454 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td668\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-kube-api-access-td668\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.673264 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g"] Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.685013 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.694997 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggqvl\" (UniqueName: \"kubernetes.io/projected/459bf85f-e304-4481-a4cb-2c391953d6f5-kube-api-access-ggqvl\") pod \"apiserver-76f77b778f-9h9tw\" (UID: \"459bf85f-e304-4481-a4cb-2c391953d6f5\") " pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.699834 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj"] Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.707146 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.707827 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.708069 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.20804034 +0000 UTC m=+152.436627203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.708141 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-bound-sa-token\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.708152 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.708408 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.20839755 +0000 UTC m=+152.436984503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.723096 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p47g5\" (UniqueName: \"kubernetes.io/projected/eba09ec1-c855-493b-8ab3-b40d4db28cc1-kube-api-access-p47g5\") pod \"controller-manager-879f6c89f-bbmc8\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.746457 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6qrw\" (UniqueName: \"kubernetes.io/projected/802f21bc-5a4e-4fa2-9dc1-2b328f1cf924-kube-api-access-r6qrw\") pod \"migrator-59844c95c7-j2hmd\" (UID: \"802f21bc-5a4e-4fa2-9dc1-2b328f1cf924\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.761818 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.763723 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6kzjp"] Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.769969 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7559fe1d-a033-4a6e-8227-ebd2f22dc496-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4zxqh\" (UID: \"7559fe1d-a033-4a6e-8227-ebd2f22dc496\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.781392 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bckqh" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.783772 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chgzf\" (UniqueName: \"kubernetes.io/projected/3880113c-b614-4254-a166-ec2c967a86be-kube-api-access-chgzf\") pod \"machine-config-controller-84d6567774-sf5wg\" (UID: \"3880113c-b614-4254-a166-ec2c967a86be\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.791560 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:12:00 crc kubenswrapper[4846]: W0202 12:12:00.792999 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b159dbf_2d80_429e_b473_1260d8f3d67b.slice/crio-deea11af12863176e6a51391d2739178726ade39e2631ca4a4e5046968341ad8 WatchSource:0}: Error finding container deea11af12863176e6a51391d2739178726ade39e2631ca4a4e5046968341ad8: Status 404 returned error can't find the container with id deea11af12863176e6a51391d2739178726ade39e2631ca4a4e5046968341ad8 Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.806413 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r"] Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.806473 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8nx7\" (UniqueName: \"kubernetes.io/projected/47a375de-20fc-4cd3-a224-b3fd6cd088c1-kube-api-access-j8nx7\") pod \"router-default-5444994796-h77rj\" (UID: \"47a375de-20fc-4cd3-a224-b3fd6cd088c1\") " pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.810811 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.811253 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.311238873 +0000 UTC m=+152.539825736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.824850 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq9vv\" (UniqueName: \"kubernetes.io/projected/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-kube-api-access-vq9vv\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.847442 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6gnh\" (UniqueName: \"kubernetes.io/projected/8c8cc0f5-61a4-482e-9413-675bc26d1781-kube-api-access-z6gnh\") pod \"package-server-manager-789f6589d5-fdjlt\" (UID: \"8c8cc0f5-61a4-482e-9413-675bc26d1781\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.862304 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx"] Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.868438 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhqqf\" (UniqueName: \"kubernetes.io/projected/6fe57969-bb65-43e3-a756-3bc0574727d8-kube-api-access-qhqqf\") pod \"catalog-operator-68c6474976-ph4nq\" (UID: \"6fe57969-bb65-43e3-a756-3bc0574727d8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.892334 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fklg\" (UniqueName: \"kubernetes.io/projected/f1adc015-453c-4718-bb2c-25d6b829f12e-kube-api-access-7fklg\") pod \"service-ca-9c57cc56f-tsz82\" (UID: \"f1adc015-453c-4718-bb2c-25d6b829f12e\") " pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.905820 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65vw8\" (UniqueName: \"kubernetes.io/projected/b3b18c12-56ac-4a04-bcda-f261a6446f24-kube-api-access-65vw8\") pod \"ingress-canary-jn5pn\" (UID: \"b3b18c12-56ac-4a04-bcda-f261a6446f24\") " pod="openshift-ingress-canary/ingress-canary-jn5pn" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.912469 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:00 crc kubenswrapper[4846]: E0202 12:12:00.912864 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.4128458 +0000 UTC m=+152.641432663 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.929030 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wbtlv\" (UID: \"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.946460 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/20e43514-4056-412d-a96e-aa0d61e427f0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2ps68\" (UID: \"20e43514-4056-412d-a96e-aa0d61e427f0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.962611 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c496r\" (UniqueName: \"kubernetes.io/projected/8a05f2a9-e378-4355-b28a-1d0ee567091f-kube-api-access-c496r\") pod \"multus-admission-controller-857f4d67dd-7gk4g\" (UID: \"8a05f2a9-e378-4355-b28a-1d0ee567091f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.973356 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:12:00 crc kubenswrapper[4846]: I0202 12:12:00.982894 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bcnf\" (UniqueName: \"kubernetes.io/projected/739682cc-1b85-4417-95cf-09940455f8c1-kube-api-access-6bcnf\") pod \"etcd-operator-b45778765-7xgq6\" (UID: \"739682cc-1b85-4417-95cf-09940455f8c1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.002828 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/88863c6c-303f-4f51-95f1-42bcbf8d7701-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hmldg\" (UID: \"88863c6c-303f-4f51-95f1-42bcbf8d7701\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.006545 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.013446 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.014021 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.014157 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.514137588 +0000 UTC m=+152.742724451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.021772 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.023318 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-9h9tw"] Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.027160 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.031199 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmq7h\" (UniqueName: \"kubernetes.io/projected/9a4cbd38-0dde-440f-96a3-4b0bd3c2da81-kube-api-access-mmq7h\") pod \"control-plane-machine-set-operator-78cbb6b69f-6w8zm\" (UID: \"9a4cbd38-0dde-440f-96a3-4b0bd3c2da81\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.052225 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72nlv\" (UniqueName: \"kubernetes.io/projected/5d0f31bd-b822-453c-8ed2-12d0e348946c-kube-api-access-72nlv\") pod \"dns-operator-744455d44c-2472r\" (UID: \"5d0f31bd-b822-453c-8ed2-12d0e348946c\") " pod="openshift-dns-operator/dns-operator-744455d44c-2472r" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.060803 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.064636 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-q57ps"] Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.067529 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.067610 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cthc\" (UniqueName: \"kubernetes.io/projected/3766f763-0e01-466f-85dd-9cbb3a65326d-kube-api-access-6cthc\") pod \"service-ca-operator-777779d784-ff7cw\" (UID: \"3766f763-0e01-466f-85dd-9cbb3a65326d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.075699 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.085121 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2472r" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.092678 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" Feb 02 12:12:01 crc kubenswrapper[4846]: W0202 12:12:01.093807 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod459bf85f_e304_4481_a4cb_2c391953d6f5.slice/crio-9ded45d39fece387ff657868dd236f7903ee45e14e3863bc465484ae870619a7 WatchSource:0}: Error finding container 9ded45d39fece387ff657868dd236f7903ee45e14e3863bc465484ae870619a7: Status 404 returned error can't find the container with id 9ded45d39fece387ff657868dd236f7903ee45e14e3863bc465484ae870619a7 Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.105641 4846 secret.go:188] Couldn't get secret openshift-console/console-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.105760 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert podName:af51b9e4-0ee3-4bf9-93e5-27b4039be121 nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.105740257 +0000 UTC m=+153.334327120 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "console-serving-cert" (UniqueName: "kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert") pod "console-f9d7485db-79skc" (UID: "af51b9e4-0ee3-4bf9-93e5-27b4039be121") : failed to sync secret cache: timed out waiting for the condition Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.106051 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.107481 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxn6g\" (UniqueName: \"kubernetes.io/projected/b3e406b8-473a-40f4-b464-e6f9992ab0c5-kube-api-access-sxn6g\") pod \"csi-hostpathplugin-h6lcv\" (UID: \"b3e406b8-473a-40f4-b464-e6f9992ab0c5\") " pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.110981 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.112236 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2nmb\" (UniqueName: \"kubernetes.io/projected/618d005a-d18f-4ad9-92ba-cb1a97a9cebc-kube-api-access-x2nmb\") pod \"dns-default-kbvlq\" (UID: \"618d005a-d18f-4ad9-92ba-cb1a97a9cebc\") " pod="openshift-dns/dns-default-kbvlq" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.112238 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bckqh"] Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.114498 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.115000 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.614986033 +0000 UTC m=+152.843572896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.135562 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkxsp\" (UniqueName: \"kubernetes.io/projected/c0652d70-92dc-47c7-926c-d893e14aedfd-kube-api-access-qkxsp\") pod \"machine-config-server-kjwnm\" (UID: \"c0652d70-92dc-47c7-926c-d893e14aedfd\") " pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.136353 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.155720 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts"] Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.156354 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftw2p\" (UniqueName: \"kubernetes.io/projected/01142a09-9483-44eb-93a3-746603cf21ed-kube-api-access-ftw2p\") pod \"kube-storage-version-migrator-operator-b67b599dd-h5lq9\" (UID: \"01142a09-9483-44eb-93a3-746603cf21ed\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:12:01 crc kubenswrapper[4846]: W0202 12:12:01.158377 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4157225e_a9ff_4cbe_b56b_43d309030abd.slice/crio-886e63ca053a1fcb2f50dcad325fce0ca03815edbef7043e056b787c0fd3b4d2 WatchSource:0}: Error finding container 886e63ca053a1fcb2f50dcad325fce0ca03815edbef7043e056b787c0fd3b4d2: Status 404 returned error can't find the container with id 886e63ca053a1fcb2f50dcad325fce0ca03815edbef7043e056b787c0fd3b4d2 Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.170972 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jn5pn" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.171949 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m795\" (UniqueName: \"kubernetes.io/projected/34dedc1d-10d1-4ca0-b059-99048ba37464-kube-api-access-6m795\") pod \"marketplace-operator-79b997595-zbpnp\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.173116 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7qdsx"] Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.183709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" event={"ID":"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc","Type":"ContainerStarted","Data":"6de0055f7f499114f6736bd7d4383cd6d513681c48f1a57c19f4ede587cab427"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.183769 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" event={"ID":"12c5b3f2-6d12-4f68-8013-fd5a555b7cfc","Type":"ContainerStarted","Data":"56a2fe6796c35c9ed230873106ff33009f94adedfc1e666b73f2fd8d68fa96d2"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.189200 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghm4p\" (UniqueName: \"kubernetes.io/projected/07661a94-bbd8-4b80-842c-d9db04206967-kube-api-access-ghm4p\") pod \"packageserver-d55dfcdfc-p8mks\" (UID: \"07661a94-bbd8-4b80-842c-d9db04206967\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.191554 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" event={"ID":"459bf85f-e304-4481-a4cb-2c391953d6f5","Type":"ContainerStarted","Data":"9ded45d39fece387ff657868dd236f7903ee45e14e3863bc465484ae870619a7"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.196704 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.199079 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" event={"ID":"8abc246a-7ca3-4550-ba36-491f830e5419","Type":"ContainerStarted","Data":"933e0a8863d3a600eca001ad3c12750041ece16fe4d65fee335eb17f0fdb58ac"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.201932 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" event={"ID":"8abc246a-7ca3-4550-ba36-491f830e5419","Type":"ContainerStarted","Data":"4ce419f910ad22c06056690bb4019072dea65fccc2b4786753fa903cd5d25d77"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.203900 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kbvlq" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.206612 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kjwnm" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.211516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjmdc\" (UniqueName: \"kubernetes.io/projected/52772605-2b37-4e17-a66a-5456e6e27d47-kube-api-access-tjmdc\") pod \"machine-config-operator-74547568cd-n7rfv\" (UID: \"52772605-2b37-4e17-a66a-5456e6e27d47\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.216360 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.216684 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.716666933 +0000 UTC m=+152.945253796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.218333 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" event={"ID":"9b159dbf-2d80-429e-b473-1260d8f3d67b","Type":"ContainerStarted","Data":"deea11af12863176e6a51391d2739178726ade39e2631ca4a4e5046968341ad8"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.222819 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bbmc8"] Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.223345 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcx7t\" (UniqueName: \"kubernetes.io/projected/b30d0b23-8b4d-4863-93e9-5afd111cb2c9-kube-api-access-qcx7t\") pod \"console-operator-58897d9998-zgkpl\" (UID: \"b30d0b23-8b4d-4863-93e9-5afd111cb2c9\") " pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.224536 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" event={"ID":"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0","Type":"ContainerStarted","Data":"cefc6f9d60a83ed6f04d9a8e16bd284f7da588841f6173e060f0226e8594a0fd"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.236148 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" event={"ID":"2c142082-61b2-4ea2-9d3b-8f9bae616178","Type":"ContainerStarted","Data":"b5c4fd5aa1f0e87a41be29d66067b484444908933c9cc59f98e1239a106e9f26"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.236200 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" event={"ID":"2c142082-61b2-4ea2-9d3b-8f9bae616178","Type":"ContainerStarted","Data":"cd7b02588af9094c0f34570b6e15cb9582750b9a0c4374131cc3fe50448b8dd8"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.244803 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.247313 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" event={"ID":"ec6797b4-3f50-4496-97a9-eb84dd09d5c9","Type":"ContainerStarted","Data":"8f468181088598a609c761679bf1da6cc164283ecc0ec927cdb207c614833cec"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.247366 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" event={"ID":"ec6797b4-3f50-4496-97a9-eb84dd09d5c9","Type":"ContainerStarted","Data":"f4a1b4c888e3cce8ee145c1f1fadcb05a16b3283d275e890e22600768817df26"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.248592 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.249587 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pm6v\" (UniqueName: \"kubernetes.io/projected/7d559728-5463-4a87-b4e7-b767dfe50bd7-kube-api-access-6pm6v\") pod \"collect-profiles-29500560-wb4lk\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.253389 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" event={"ID":"614c989a-6be2-4824-a3e5-d810fd94e39b","Type":"ContainerStarted","Data":"3156ca4377047f3b473eb0e2df550cab31925afe0172ce8200e30d29a5ea3131"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.253433 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" event={"ID":"614c989a-6be2-4824-a3e5-d810fd94e39b","Type":"ContainerStarted","Data":"34e1645e2b0b7a38596008ed62d93a94e4e8022b8df19128cff6ce117cdcdfa0"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.262268 4846 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-cfv6g container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.262327 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" podUID="ec6797b4-3f50-4496-97a9-eb84dd09d5c9" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.265682 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" event={"ID":"e94b3ca8-f11a-4fee-8ef1-05a56d233593","Type":"ContainerStarted","Data":"c3bf5631b20aca58f0e0167d1d93f4d9aa87968c2c27979201eaf3d38c0d4300"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.267714 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxnwd\" (UniqueName: \"kubernetes.io/projected/0d439a22-f822-4714-8135-70a093c340f6-kube-api-access-cxnwd\") pod \"olm-operator-6b444d44fb-s8wzs\" (UID: \"0d439a22-f822-4714-8135-70a093c340f6\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.269413 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.269998 4846 generic.go:334] "Generic (PLEG): container finished" podID="a62db6be-7478-41ce-b1a7-fbf3f24967e2" containerID="91681d50297cffb336186debfe1d4544e947c38b37695ec68d91e7bab70b68b0" exitCode=0 Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.270045 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" event={"ID":"a62db6be-7478-41ce-b1a7-fbf3f24967e2","Type":"ContainerDied","Data":"91681d50297cffb336186debfe1d4544e947c38b37695ec68d91e7bab70b68b0"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.270081 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" event={"ID":"a62db6be-7478-41ce-b1a7-fbf3f24967e2","Type":"ContainerStarted","Data":"e8a15946e3577b574b8f2a28bfce030ee9c066a09e3499c271d8c6631842ad12"} Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.280978 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.287165 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:12:01 crc kubenswrapper[4846]: W0202 12:12:01.292175 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47a375de_20fc_4cd3_a224_b3fd6cd088c1.slice/crio-c80ced1cf2f1b89c2e85b7ff0f7d6bfc8519bb34fe19b41d1ddd076a302433fc WatchSource:0}: Error finding container c80ced1cf2f1b89c2e85b7ff0f7d6bfc8519bb34fe19b41d1ddd076a302433fc: Status 404 returned error can't find the container with id c80ced1cf2f1b89c2e85b7ff0f7d6bfc8519bb34fe19b41d1ddd076a302433fc Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.295879 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.319449 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.320671 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.820639448 +0000 UTC m=+153.049226411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.337269 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.345053 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.347005 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd"] Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.351136 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.417896 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.420466 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.421249 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:01.921223886 +0000 UTC m=+153.149810749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.442453 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.450320 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.522324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.522853 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.022830194 +0000 UTC m=+153.251417067 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.623386 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.623841 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.123818473 +0000 UTC m=+153.352405336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.726015 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.726743 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.226726537 +0000 UTC m=+153.455313400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.830717 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2472r"] Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.830829 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.830978 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.330955371 +0000 UTC m=+153.559542244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.831075 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.831441 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.331428304 +0000 UTC m=+153.560015167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.832759 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tsz82"] Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.835348 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg"] Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.932266 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.932510 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.432479775 +0000 UTC m=+153.661066648 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:01 crc kubenswrapper[4846]: I0202 12:12:01.932880 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:01 crc kubenswrapper[4846]: E0202 12:12:01.933223 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.433214737 +0000 UTC m=+153.661801600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.039143 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.039627 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.539603322 +0000 UTC m=+153.768190185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.099556 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" podStartSLOduration=128.099535245 podStartE2EDuration="2m8.099535245s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:02.085769217 +0000 UTC m=+153.314356100" watchObservedRunningTime="2026-02-02 12:12:02.099535245 +0000 UTC m=+153.328122108" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.142769 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.142878 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.143190 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.643175576 +0000 UTC m=+153.871762449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.152302 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert\") pod \"console-f9d7485db-79skc\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.222024 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-ml2fm" podStartSLOduration=129.221995924 podStartE2EDuration="2m9.221995924s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:02.212350446 +0000 UTC m=+153.440937309" watchObservedRunningTime="2026-02-02 12:12:02.221995924 +0000 UTC m=+153.450582797" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.245279 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.245605 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.745587667 +0000 UTC m=+153.974174530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.271669 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.300894 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh"] Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.302162 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" event={"ID":"5256ddb9-8edb-42a8-a041-cb806d849e58","Type":"ContainerStarted","Data":"61ab34e2d9ca1658fe01befd3a86861e81a2e80e80113198d2ec477075d44c58"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.302216 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" event={"ID":"5256ddb9-8edb-42a8-a041-cb806d849e58","Type":"ContainerStarted","Data":"830e01aff7e03519bb66d364b0b837f59f451f0bf09f9d9cd551bd5e5dcc6cdc"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.307093 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" event={"ID":"9f149bcb-2741-4754-9b16-a87be939e504","Type":"ContainerStarted","Data":"145047ec24ffb09575d0dc4cea6c0547f424639cff5b44f7e1f6f7114c2591b9"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.307261 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" event={"ID":"9f149bcb-2741-4754-9b16-a87be939e504","Type":"ContainerStarted","Data":"6773128281bc22c5f44ec67d5843dc66434e5ca3c3caf9c031c74a699f548536"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.310280 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7gk4g"] Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.317391 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" event={"ID":"3880113c-b614-4254-a166-ec2c967a86be","Type":"ContainerStarted","Data":"4c0d8f2e20188030e2e0cb0aa43287c70ee1f0d5d4e749c539511d6e51fde179"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.322447 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd" event={"ID":"802f21bc-5a4e-4fa2-9dc1-2b328f1cf924","Type":"ContainerStarted","Data":"024db0bdc30476373112821ad2ef3c74e61959616c80c97fc0923813b47474cc"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.322498 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd" event={"ID":"802f21bc-5a4e-4fa2-9dc1-2b328f1cf924","Type":"ContainerStarted","Data":"7ade22d96a268a8463f54e3c932178a9c756aa1278ca8239d1b76afbe50c2249"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.328740 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-6kzjp" podStartSLOduration=129.32872318 podStartE2EDuration="2m9.32872318s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:02.327241377 +0000 UTC m=+153.555828250" watchObservedRunningTime="2026-02-02 12:12:02.32872318 +0000 UTC m=+153.557310043" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.330979 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kjwnm" event={"ID":"c0652d70-92dc-47c7-926c-d893e14aedfd","Type":"ContainerStarted","Data":"e3d6846173dc554c3684fc7a42673354a352e08ef071d6f8dec04ac6fdbc86bc"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.335137 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7xgq6"] Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.340235 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" event={"ID":"eba09ec1-c855-493b-8ab3-b40d4db28cc1","Type":"ContainerStarted","Data":"35b11d93c537c1b9170c391cf4becfe22201a1cb6cdbabdf505aa014394503eb"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.340302 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" event={"ID":"eba09ec1-c855-493b-8ab3-b40d4db28cc1","Type":"ContainerStarted","Data":"b15410055d3c6790086cbd87db8cd26d65a19c5677f1e0efdfca918b8560580d"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.341320 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.347351 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" event={"ID":"f1adc015-453c-4718-bb2c-25d6b829f12e","Type":"ContainerStarted","Data":"4d097aa31b9f3e3926296e87c7bddb37bd582729ff5886cf142d2fafc4106497"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.347902 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.348268 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.848249824 +0000 UTC m=+154.076836687 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.366240 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2jvrx" podStartSLOduration=128.366219394 podStartE2EDuration="2m8.366219394s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:02.364306339 +0000 UTC m=+153.592893222" watchObservedRunningTime="2026-02-02 12:12:02.366219394 +0000 UTC m=+153.594806267" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.370740 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" event={"ID":"9b159dbf-2d80-429e-b473-1260d8f3d67b","Type":"ContainerStarted","Data":"1a679e75baf5a5b30695a7ecb5ed9d434cb535eb2bf8a1d474e6e548e75b7dfa"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.379904 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" event={"ID":"a62db6be-7478-41ce-b1a7-fbf3f24967e2","Type":"ContainerStarted","Data":"ee3022c39e64d298daa5d4917e3cb521e035650d11f9984cf4faf1fe51b27562"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.380264 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.384724 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2472r" event={"ID":"5d0f31bd-b822-453c-8ed2-12d0e348946c","Type":"ContainerStarted","Data":"27221fc5013b9f12b177e59c735bd1756bc95b0c6b977681b4e1403cb70f9053"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.388149 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.390358 4846 generic.go:334] "Generic (PLEG): container finished" podID="459bf85f-e304-4481-a4cb-2c391953d6f5" containerID="83220c3fa18be4a652a5979b0360078dfc804508bdfcb5a02710ee1fe5e21ddc" exitCode=0 Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.390691 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" event={"ID":"459bf85f-e304-4481-a4cb-2c391953d6f5","Type":"ContainerDied","Data":"83220c3fa18be4a652a5979b0360078dfc804508bdfcb5a02710ee1fe5e21ddc"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.401305 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" event={"ID":"8abc246a-7ca3-4550-ba36-491f830e5419","Type":"ContainerStarted","Data":"41b5d538b8e0baadff5fa35b8f9c9e2f1245804e1ca191b6cf2cea941cb834a8"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.413932 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bckqh" event={"ID":"4157225e-a9ff-4cbe-b56b-43d309030abd","Type":"ContainerStarted","Data":"62c9c597f40fbbbb3f67c384a11f99f505ceb1817ca799d05f545a20819c7f0d"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.413978 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bckqh" event={"ID":"4157225e-a9ff-4cbe-b56b-43d309030abd","Type":"ContainerStarted","Data":"886e63ca053a1fcb2f50dcad325fce0ca03815edbef7043e056b787c0fd3b4d2"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.414734 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-bckqh" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.420514 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-h77rj" event={"ID":"47a375de-20fc-4cd3-a224-b3fd6cd088c1","Type":"ContainerStarted","Data":"309bc22e5e28580f8031a91131d575b444bce946a82701ce3c59b9ad3983d642"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.420574 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-h77rj" event={"ID":"47a375de-20fc-4cd3-a224-b3fd6cd088c1","Type":"ContainerStarted","Data":"c80ced1cf2f1b89c2e85b7ff0f7d6bfc8519bb34fe19b41d1ddd076a302433fc"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.428511 4846 generic.go:334] "Generic (PLEG): container finished" podID="e94b3ca8-f11a-4fee-8ef1-05a56d233593" containerID="a5525ae382ab746461ee1da2e306aa5dec2c5aae0e31e848e1ce6929296e4a7c" exitCode=0 Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.429833 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" event={"ID":"e94b3ca8-f11a-4fee-8ef1-05a56d233593","Type":"ContainerDied","Data":"a5525ae382ab746461ee1da2e306aa5dec2c5aae0e31e848e1ce6929296e4a7c"} Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.449796 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.449979 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.949957274 +0000 UTC m=+154.178544137 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.450806 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.453947 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:02.95393543 +0000 UTC m=+154.182522293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.552289 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.553085 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.053065494 +0000 UTC m=+154.281652357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.628676 4846 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-q57ps container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.628739 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" podUID="ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.628800 4846 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-bbmc8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.628848 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" podUID="eba09ec1-c855-493b-8ab3-b40d4db28cc1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.631106 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.641848 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-bckqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.641919 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bckqh" podUID="4157225e-a9ff-4cbe-b56b-43d309030abd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.654793 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.655084 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.155073093 +0000 UTC m=+154.383659956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: W0202 12:12:02.665903 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7559fe1d_a033_4a6e_8227_ebd2f22dc496.slice/crio-15367aea404a0c394d6a45e440dda0b5ed0700ad28f5a5a8fcda56f4b103364b WatchSource:0}: Error finding container 15367aea404a0c394d6a45e440dda0b5ed0700ad28f5a5a8fcda56f4b103364b: Status 404 returned error can't find the container with id 15367aea404a0c394d6a45e440dda0b5ed0700ad28f5a5a8fcda56f4b103364b Feb 02 12:12:02 crc kubenswrapper[4846]: W0202 12:12:02.685915 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod739682cc_1b85_4417_95cf_09940455f8c1.slice/crio-af792d57975b45c9aed7a731dbd0bd6cc19e840ac178eb250f5297c59d6aa109 WatchSource:0}: Error finding container af792d57975b45c9aed7a731dbd0bd6cc19e840ac178eb250f5297c59d6aa109: Status 404 returned error can't find the container with id af792d57975b45c9aed7a731dbd0bd6cc19e840ac178eb250f5297c59d6aa109 Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.759514 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.759923 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.259901374 +0000 UTC m=+154.488488237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.860716 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.861509 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.361494371 +0000 UTC m=+154.590081244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.956942 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt"] Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.965455 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:02 crc kubenswrapper[4846]: E0202 12:12:02.965914 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.465893908 +0000 UTC m=+154.694480771 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.973924 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm"] Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.977284 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv"] Feb 02 12:12:02 crc kubenswrapper[4846]: I0202 12:12:02.991552 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.025977 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.035676 4846 patch_prober.go:28] interesting pod/router-default-5444994796-h77rj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 12:12:03 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Feb 02 12:12:03 crc kubenswrapper[4846]: [+]process-running ok Feb 02 12:12:03 crc kubenswrapper[4846]: healthz check failed Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.035724 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h77rj" podUID="47a375de-20fc-4cd3-a224-b3fd6cd088c1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.069695 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.070391 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.570371279 +0000 UTC m=+154.798958152 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.172572 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.172843 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.67281729 +0000 UTC m=+154.901404163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.172905 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.173281 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.673271493 +0000 UTC m=+154.901858356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.243327 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.268260 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jn5pn"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.275272 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.275743 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.775723215 +0000 UTC m=+155.004310078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.290433 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" podStartSLOduration=130.29041123 podStartE2EDuration="2m10.29041123s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.270838374 +0000 UTC m=+154.499425237" watchObservedRunningTime="2026-02-02 12:12:03.29041123 +0000 UTC m=+154.518998093" Feb 02 12:12:03 crc kubenswrapper[4846]: W0202 12:12:03.326833 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3b18c12_56ac_4a04_bcda_f261a6446f24.slice/crio-8339c23d3916877df00ce3a04aa7a2b644d8d01a42fa5757d707d8614e1b24f5 WatchSource:0}: Error finding container 8339c23d3916877df00ce3a04aa7a2b644d8d01a42fa5757d707d8614e1b24f5: Status 404 returned error can't find the container with id 8339c23d3916877df00ce3a04aa7a2b644d8d01a42fa5757d707d8614e1b24f5 Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.380458 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" podStartSLOduration=129.380438172 podStartE2EDuration="2m9.380438172s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.353352059 +0000 UTC m=+154.581938922" watchObservedRunningTime="2026-02-02 12:12:03.380438172 +0000 UTC m=+154.609025035" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.381017 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-bckqh" podStartSLOduration=129.381010909 podStartE2EDuration="2m9.381010909s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.380805633 +0000 UTC m=+154.609392496" watchObservedRunningTime="2026-02-02 12:12:03.381010909 +0000 UTC m=+154.609597772" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.381394 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.381715 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.881704748 +0000 UTC m=+155.110291611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.415179 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.431212 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" podStartSLOduration=130.431185689 podStartE2EDuration="2m10.431185689s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.417585186 +0000 UTC m=+154.646172049" watchObservedRunningTime="2026-02-02 12:12:03.431185689 +0000 UTC m=+154.659772552" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.463845 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" podStartSLOduration=130.463828163 podStartE2EDuration="2m10.463828163s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.456745378 +0000 UTC m=+154.685332241" watchObservedRunningTime="2026-02-02 12:12:03.463828163 +0000 UTC m=+154.692415026" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.482153 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.482664 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:03.982625216 +0000 UTC m=+155.211212129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.526363 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-bmwts" podStartSLOduration=129.52634699 podStartE2EDuration="2m9.52634699s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.494080078 +0000 UTC m=+154.722666961" watchObservedRunningTime="2026-02-02 12:12:03.52634699 +0000 UTC m=+154.754933853" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.546764 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.546793 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" event={"ID":"6fe57969-bb65-43e3-a756-3bc0574727d8","Type":"ContainerStarted","Data":"491dfe4cb4768d8469c570754df5e0b47b1dfd1848129fe5a910c9a9004e5c14"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.546817 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.546826 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" event={"ID":"6fe57969-bb65-43e3-a756-3bc0574727d8","Type":"ContainerStarted","Data":"f07ca4c97651e8eed902b4b04419c04dce9452f2b6ade0f16d8c89b644973301"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.546837 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zgkpl"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.546850 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.547614 4846 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-ph4nq container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.547662 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" podUID="6fe57969-bb65-43e3-a756-3bc0574727d8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.549950 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-h6lcv"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.560700 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.562587 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" event={"ID":"9a4cbd38-0dde-440f-96a3-4b0bd3c2da81","Type":"ContainerStarted","Data":"6da2e6d40f26fe9d5d52289029ab77e3d793e4356733a5a0ea939c21a5ba7667"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.562625 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" event={"ID":"9a4cbd38-0dde-440f-96a3-4b0bd3c2da81","Type":"ContainerStarted","Data":"df1be8b2dba058d2e3fcef9c157ebd47a7e93a3a671e35f11ee3e24fb154b789"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.574394 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" event={"ID":"e94b3ca8-f11a-4fee-8ef1-05a56d233593","Type":"ContainerStarted","Data":"bd247df8bf65cdf47a0eb5b6053b7592267a0ac69194bc418290244bfb719287"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.574432 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.586231 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.586957 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.086944602 +0000 UTC m=+155.315531465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.588893 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kbvlq"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.598099 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zbpnp"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.602290 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.604026 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" event={"ID":"f1adc015-453c-4718-bb2c-25d6b829f12e","Type":"ContainerStarted","Data":"86f61d57d2ed2fd1e97b10b9d621ab821751995be7f7c94a7f7d31105aa0b0ea"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.606219 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.607222 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" podStartSLOduration=129.607208737 podStartE2EDuration="2m9.607208737s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.605998322 +0000 UTC m=+154.834585185" watchObservedRunningTime="2026-02-02 12:12:03.607208737 +0000 UTC m=+154.835795590" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.615938 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" event={"ID":"7559fe1d-a033-4a6e-8227-ebd2f22dc496","Type":"ContainerStarted","Data":"8fda9b7d3745b3357e20fb1c9f54c5a24b2157fa02669b1259a4e40adebfa319"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.615972 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" event={"ID":"7559fe1d-a033-4a6e-8227-ebd2f22dc496","Type":"ContainerStarted","Data":"15367aea404a0c394d6a45e440dda0b5ed0700ad28f5a5a8fcda56f4b103364b"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.640299 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-79skc"] Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.647411 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" event={"ID":"8c8cc0f5-61a4-482e-9413-675bc26d1781","Type":"ContainerStarted","Data":"dd0d1bbaffee17deaf527f7532135b0a1ad527001adb992b7e81b9627cec5102"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.647446 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" event={"ID":"8c8cc0f5-61a4-482e-9413-675bc26d1781","Type":"ContainerStarted","Data":"c11397ce39b74c3918c2b59cb605e9818346f8366b37e8959ca78d990040b17e"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.650040 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" event={"ID":"8a05f2a9-e378-4355-b28a-1d0ee567091f","Type":"ContainerStarted","Data":"cc45c45052cd22023fa0b29154bc30412ec01bc54a956024ea2001838305df1c"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.650067 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" event={"ID":"8a05f2a9-e378-4355-b28a-1d0ee567091f","Type":"ContainerStarted","Data":"3da8fec09328f7ce22c2f0bd40464567607e221b4691c908387eb169be65c17f"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.660903 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" event={"ID":"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0","Type":"ContainerStarted","Data":"9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.662345 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" event={"ID":"88863c6c-303f-4f51-95f1-42bcbf8d7701","Type":"ContainerStarted","Data":"e5539f434b6306364008fab8cad839cdae35c3e531c5d217e56c5b37c7c3fe20"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.683293 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-kjwnm" podStartSLOduration=5.683278227 podStartE2EDuration="5.683278227s" podCreationTimestamp="2026-02-02 12:11:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.68304248 +0000 UTC m=+154.911629343" watchObservedRunningTime="2026-02-02 12:12:03.683278227 +0000 UTC m=+154.911865090" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.687729 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.687862 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.187839408 +0000 UTC m=+155.416426271 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.688160 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.691944 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.191928917 +0000 UTC m=+155.420515780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.704670 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" event={"ID":"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac","Type":"ContainerStarted","Data":"6d133fea37151ddccebc4745c0483355355ec0a4a65ba945ae04f19a6303d500"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.704744 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" event={"ID":"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac","Type":"ContainerStarted","Data":"e61fd7bff7c18e4b3c90e1ac10aa9d5e995845a403692e3106180c8c11531048"} Feb 02 12:12:03 crc kubenswrapper[4846]: W0202 12:12:03.731766 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07661a94_bbd8_4b80_842c_d9db04206967.slice/crio-0256aa244d3564e96a41080e55be35109f9ea63727ea70deda3bf136ed5504b2 WatchSource:0}: Error finding container 0256aa244d3564e96a41080e55be35109f9ea63727ea70deda3bf136ed5504b2: Status 404 returned error can't find the container with id 0256aa244d3564e96a41080e55be35109f9ea63727ea70deda3bf136ed5504b2 Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.755514 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-h77rj" podStartSLOduration=129.755495304 podStartE2EDuration="2m9.755495304s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.732403337 +0000 UTC m=+154.960990200" watchObservedRunningTime="2026-02-02 12:12:03.755495304 +0000 UTC m=+154.984082167" Feb 02 12:12:03 crc kubenswrapper[4846]: W0202 12:12:03.775848 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3766f763_0e01_466f_85dd_9cbb3a65326d.slice/crio-1718494a6045b142b3359f549a169e272e5ed2987fd0434ea9010081a72fa07e WatchSource:0}: Error finding container 1718494a6045b142b3359f549a169e272e5ed2987fd0434ea9010081a72fa07e: Status 404 returned error can't find the container with id 1718494a6045b142b3359f549a169e272e5ed2987fd0434ea9010081a72fa07e Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.776113 4846 csr.go:261] certificate signing request csr-7pl28 is approved, waiting to be issued Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.777879 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" event={"ID":"459bf85f-e304-4481-a4cb-2c391953d6f5","Type":"ContainerStarted","Data":"be26a5fd85f719f61df06c2ac26dc80a6d3110518610849de422ee4b31f99fce"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.790021 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.790861 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.290843146 +0000 UTC m=+155.519430009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.791652 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.791955 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.291933117 +0000 UTC m=+155.520519980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.792176 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4zxqh" podStartSLOduration=129.792163183 podStartE2EDuration="2m9.792163183s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.77785519 +0000 UTC m=+155.006442053" watchObservedRunningTime="2026-02-02 12:12:03.792163183 +0000 UTC m=+155.020750046" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.807378 4846 csr.go:257] certificate signing request csr-7pl28 is issued Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.827283 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ghkkq" event={"ID":"9b159dbf-2d80-429e-b473-1260d8f3d67b","Type":"ContainerStarted","Data":"b588d9834d625dd18e133834180d88d44fb0a41ab1ac3966401960cdad84ccf9"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.829002 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" podStartSLOduration=129.828979238 podStartE2EDuration="2m9.828979238s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.827016621 +0000 UTC m=+155.055603484" watchObservedRunningTime="2026-02-02 12:12:03.828979238 +0000 UTC m=+155.057566101" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.834184 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jn5pn" event={"ID":"b3b18c12-56ac-4a04-bcda-f261a6446f24","Type":"ContainerStarted","Data":"8339c23d3916877df00ce3a04aa7a2b644d8d01a42fa5757d707d8614e1b24f5"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.841009 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kjwnm" event={"ID":"c0652d70-92dc-47c7-926c-d893e14aedfd","Type":"ContainerStarted","Data":"a9fcf51748a9f04b75a68344fafc04429726d65499021ec463e68df8d8c65903"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.866216 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" podStartSLOduration=129.866201594 podStartE2EDuration="2m9.866201594s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.86396316 +0000 UTC m=+155.092550013" watchObservedRunningTime="2026-02-02 12:12:03.866201594 +0000 UTC m=+155.094788457" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.902584 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.902899 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.402884805 +0000 UTC m=+155.631471668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.903783 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:03 crc kubenswrapper[4846]: E0202 12:12:03.906380 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.406365195 +0000 UTC m=+155.634952058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.911889 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6w8zm" podStartSLOduration=129.911869434 podStartE2EDuration="2m9.911869434s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.901740382 +0000 UTC m=+155.130327245" watchObservedRunningTime="2026-02-02 12:12:03.911869434 +0000 UTC m=+155.140456297" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.920071 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2472r" event={"ID":"5d0f31bd-b822-453c-8ed2-12d0e348946c","Type":"ContainerStarted","Data":"779975ee7e7846e9c5cb88d7a9674a75841e35127f2e218f9f86024753e8c2ee"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.928775 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" event={"ID":"3880113c-b614-4254-a166-ec2c967a86be","Type":"ContainerStarted","Data":"eca064f19bdfbde252f6b9cbac9489e24da0f4038f302fdd0f3f06517021d5e7"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.928840 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" event={"ID":"3880113c-b614-4254-a166-ec2c967a86be","Type":"ContainerStarted","Data":"c76be29491e8833bd151b8372ef970556d7ff830d493765a837d2168f7c72b4a"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.948129 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-tsz82" podStartSLOduration=129.948082741 podStartE2EDuration="2m9.948082741s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.946334341 +0000 UTC m=+155.174921204" watchObservedRunningTime="2026-02-02 12:12:03.948082741 +0000 UTC m=+155.176669604" Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.975978 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd" event={"ID":"802f21bc-5a4e-4fa2-9dc1-2b328f1cf924","Type":"ContainerStarted","Data":"dd8803d93fc6297ecab02c65ccb601f9a84e2aaaa4d79acb179c7172da1e3a36"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.987286 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" event={"ID":"739682cc-1b85-4417-95cf-09940455f8c1","Type":"ContainerStarted","Data":"972a06045af57f7c29a7e67bc70fe1b6eb91d27d1f16999164ea0e89c608be53"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.987327 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" event={"ID":"739682cc-1b85-4417-95cf-09940455f8c1","Type":"ContainerStarted","Data":"af792d57975b45c9aed7a731dbd0bd6cc19e840ac178eb250f5297c59d6aa109"} Feb 02 12:12:03 crc kubenswrapper[4846]: I0202 12:12:03.988825 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sf5wg" podStartSLOduration=129.988799888 podStartE2EDuration="2m9.988799888s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:03.975910026 +0000 UTC m=+155.204496879" watchObservedRunningTime="2026-02-02 12:12:03.988799888 +0000 UTC m=+155.217386751" Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.004542 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.004790 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.504759539 +0000 UTC m=+155.733346402 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.010502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" event={"ID":"5256ddb9-8edb-42a8-a041-cb806d849e58","Type":"ContainerStarted","Data":"4e764ef31f21428bd159a186359de4f84d447c16a5782aff3c773d3cd72242e9"} Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.019291 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-bckqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.019379 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bckqh" podUID="4157225e-a9ff-4cbe-b56b-43d309030abd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.020371 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2hmd" podStartSLOduration=130.02034767 podStartE2EDuration="2m10.02034767s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:04.020285488 +0000 UTC m=+155.248872351" watchObservedRunningTime="2026-02-02 12:12:04.02034767 +0000 UTC m=+155.248934533" Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.026137 4846 patch_prober.go:28] interesting pod/router-default-5444994796-h77rj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 12:12:04 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Feb 02 12:12:04 crc kubenswrapper[4846]: [+]process-running ok Feb 02 12:12:04 crc kubenswrapper[4846]: healthz check failed Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.026194 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h77rj" podUID="47a375de-20fc-4cd3-a224-b3fd6cd088c1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.026671 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.106415 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7qdsx" podStartSLOduration=130.106398128 podStartE2EDuration="2m10.106398128s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:04.077613295 +0000 UTC m=+155.306200178" watchObservedRunningTime="2026-02-02 12:12:04.106398128 +0000 UTC m=+155.334984991" Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.107197 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7xgq6" podStartSLOduration=130.10718989 podStartE2EDuration="2m10.10718989s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:04.103737181 +0000 UTC m=+155.332324074" watchObservedRunningTime="2026-02-02 12:12:04.10718989 +0000 UTC m=+155.335776753" Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.110453 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.112739 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.61272672 +0000 UTC m=+155.841313583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.211403 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.211615 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.711601059 +0000 UTC m=+155.940187922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.313192 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.313844 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.813833324 +0000 UTC m=+156.042420187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.415731 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.416027 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:04.916012708 +0000 UTC m=+156.144599571 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.519441 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.519799 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.019783347 +0000 UTC m=+156.248370210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.621788 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.622210 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.122180627 +0000 UTC m=+156.350767490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.622499 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.622805 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.122793846 +0000 UTC m=+156.351380709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.652101 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.727493 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.727664 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.227612136 +0000 UTC m=+156.456198999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.727692 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.728012 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.228003477 +0000 UTC m=+156.456590390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.809048 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-02 12:07:03 +0000 UTC, rotation deadline is 2026-11-28 15:07:47.094913364 +0000 UTC Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.809158 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7178h55m42.285758501s for next certificate rotation Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.828381 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.828791 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.32876944 +0000 UTC m=+156.557356303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:04 crc kubenswrapper[4846]: I0202 12:12:04.929418 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:04 crc kubenswrapper[4846]: E0202 12:12:04.929927 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.429911283 +0000 UTC m=+156.658498146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.030302 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:05 crc kubenswrapper[4846]: E0202 12:12:05.030714 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.530686706 +0000 UTC m=+156.759273579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.041770 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zgkpl" event={"ID":"b30d0b23-8b4d-4863-93e9-5afd111cb2c9","Type":"ContainerStarted","Data":"cfa1781f283caf1402f88c191357c19efc47d10137163795ddbc4b2500a8f9c9"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.041809 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zgkpl" event={"ID":"b30d0b23-8b4d-4863-93e9-5afd111cb2c9","Type":"ContainerStarted","Data":"d7c524b32c4f35f8124f3e003d465ffe6cbf310b635efbd74b4eca77f196bf50"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.042920 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.043055 4846 patch_prober.go:28] interesting pod/router-default-5444994796-h77rj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 12:12:05 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Feb 02 12:12:05 crc kubenswrapper[4846]: [+]process-running ok Feb 02 12:12:05 crc kubenswrapper[4846]: healthz check failed Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.043117 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h77rj" podUID="47a375de-20fc-4cd3-a224-b3fd6cd088c1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.045408 4846 patch_prober.go:28] interesting pod/console-operator-58897d9998-zgkpl container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.045441 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-zgkpl" podUID="b30d0b23-8b4d-4863-93e9-5afd111cb2c9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.064894 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" event={"ID":"01142a09-9483-44eb-93a3-746603cf21ed","Type":"ContainerStarted","Data":"5ecbad5f57724bc3cae2f23da13323dbf3820fc98e405246e730e370602c345c"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.064950 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" event={"ID":"01142a09-9483-44eb-93a3-746603cf21ed","Type":"ContainerStarted","Data":"911478585b9a7683fe5981f4b2825982e1c35a6d565c578087c4369326e2dcb1"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.068792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" event={"ID":"8a05f2a9-e378-4355-b28a-1d0ee567091f","Type":"ContainerStarted","Data":"e1a3c84517ff4491d7e0360125cdce258ed0e7766c3c5cdc1b18d9a593d57016"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.116936 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-zgkpl" podStartSLOduration=131.116915969 podStartE2EDuration="2m11.116915969s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.080123755 +0000 UTC m=+156.308710618" watchObservedRunningTime="2026-02-02 12:12:05.116915969 +0000 UTC m=+156.345502832" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.119309 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-h5lq9" podStartSLOduration=131.119297478 podStartE2EDuration="2m11.119297478s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.118745842 +0000 UTC m=+156.347332705" watchObservedRunningTime="2026-02-02 12:12:05.119297478 +0000 UTC m=+156.347884341" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.123608 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" event={"ID":"20e43514-4056-412d-a96e-aa0d61e427f0","Type":"ContainerStarted","Data":"fe09cc07447af0973ec236b87d1609c49459b720db6d5c2c85cbc685752dfe03"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.123887 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" event={"ID":"20e43514-4056-412d-a96e-aa0d61e427f0","Type":"ContainerStarted","Data":"2290970a02d9b1a6ffe3bedb398bd223000253ba431bce1610f648592b9bc435"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.133317 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:05 crc kubenswrapper[4846]: E0202 12:12:05.133657 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.633621842 +0000 UTC m=+156.862208705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.139940 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" event={"ID":"b3e406b8-473a-40f4-b464-e6f9992ab0c5","Type":"ContainerStarted","Data":"8c5605440bd31a3d23cfd8d3541076e2719a794acb6beed3d71d4163c9190988"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.159223 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7gk4g" podStartSLOduration=131.159200352 podStartE2EDuration="2m11.159200352s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.141027446 +0000 UTC m=+156.369614309" watchObservedRunningTime="2026-02-02 12:12:05.159200352 +0000 UTC m=+156.387787215" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.164349 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" event={"ID":"34dedc1d-10d1-4ca0-b059-99048ba37464","Type":"ContainerStarted","Data":"ceeb99133435ec0fb619d2010f5a3b2136022c9011c8ee1f62129edd65db609a"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.164398 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" event={"ID":"34dedc1d-10d1-4ca0-b059-99048ba37464","Type":"ContainerStarted","Data":"0bf34f2ea3c6d0ba73912a7216abc2920b974c47d3be95e3e3b3bc61a2b205fc"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.165202 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.169802 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2ps68" podStartSLOduration=131.169783877 podStartE2EDuration="2m11.169783877s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.168097068 +0000 UTC m=+156.396683931" watchObservedRunningTime="2026-02-02 12:12:05.169783877 +0000 UTC m=+156.398370740" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.172474 4846 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zbpnp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.172535 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" podUID="34dedc1d-10d1-4ca0-b059-99048ba37464" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.193907 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-79skc" event={"ID":"af51b9e4-0ee3-4bf9-93e5-27b4039be121","Type":"ContainerStarted","Data":"15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.194448 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-79skc" event={"ID":"af51b9e4-0ee3-4bf9-93e5-27b4039be121","Type":"ContainerStarted","Data":"3793dbe275be8eb1c21e9d5d00438ea825623b5a249f9e167b1c04f4a30a45cb"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.206749 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" podStartSLOduration=131.206735076 podStartE2EDuration="2m11.206735076s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.206162539 +0000 UTC m=+156.434749402" watchObservedRunningTime="2026-02-02 12:12:05.206735076 +0000 UTC m=+156.435321939" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.229534 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" event={"ID":"0d439a22-f822-4714-8135-70a093c340f6","Type":"ContainerStarted","Data":"8797604533c2d7a75315f4658988c9dd76996ec0a83b2d327422f6c731ccb343"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.229592 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" event={"ID":"0d439a22-f822-4714-8135-70a093c340f6","Type":"ContainerStarted","Data":"4d493409772598c9b04c5d8faf5c3a871689b875c82ec58178248922149d5bc8"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.230751 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.231761 4846 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-s8wzs container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.231809 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" podUID="0d439a22-f822-4714-8135-70a093c340f6" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.234133 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:05 crc kubenswrapper[4846]: E0202 12:12:05.235737 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.735715843 +0000 UTC m=+156.964302706 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.252301 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" event={"ID":"fe6e43d2-c8e5-4393-a8aa-cd2de4d331ac","Type":"ContainerStarted","Data":"70be6dbc990994743c4548d076f5e816ee8f7010ac717cdd912c5bd24c004d84"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.261857 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-79skc" podStartSLOduration=131.261837588 podStartE2EDuration="2m11.261837588s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.26119259 +0000 UTC m=+156.489779473" watchObservedRunningTime="2026-02-02 12:12:05.261837588 +0000 UTC m=+156.490424451" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.283616 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" event={"ID":"3766f763-0e01-466f-85dd-9cbb3a65326d","Type":"ContainerStarted","Data":"48e3f9b57d44d6770cefd10e9241f21be4be1ff5a236602467c4eb90cb72eff3"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.283686 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" event={"ID":"3766f763-0e01-466f-85dd-9cbb3a65326d","Type":"ContainerStarted","Data":"1718494a6045b142b3359f549a169e272e5ed2987fd0434ea9010081a72fa07e"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.288930 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" event={"ID":"07661a94-bbd8-4b80-842c-d9db04206967","Type":"ContainerStarted","Data":"4bef6d9cff15f4d90194f33e8b80370e80c8d934b35f940135ef7e9fdf0eb97f"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.288962 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" event={"ID":"07661a94-bbd8-4b80-842c-d9db04206967","Type":"ContainerStarted","Data":"0256aa244d3564e96a41080e55be35109f9ea63727ea70deda3bf136ed5504b2"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.289901 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.304132 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" event={"ID":"88863c6c-303f-4f51-95f1-42bcbf8d7701","Type":"ContainerStarted","Data":"3763a53b46085122b89567f85fa3030a7dcbb622b10bfa36585b318eda41c6f6"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.309019 4846 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-p8mks container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.309075 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" podUID="07661a94-bbd8-4b80-842c-d9db04206967" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.314215 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" podStartSLOduration=131.314195772 podStartE2EDuration="2m11.314195772s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.313351548 +0000 UTC m=+156.541938411" watchObservedRunningTime="2026-02-02 12:12:05.314195772 +0000 UTC m=+156.542782645" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.325775 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" event={"ID":"7d559728-5463-4a87-b4e7-b767dfe50bd7","Type":"ContainerStarted","Data":"b77000692f2fa0089b2dc53b65e836c076ed0cf6d08245baa4599490ad60038d"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.326368 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" event={"ID":"7d559728-5463-4a87-b4e7-b767dfe50bd7","Type":"ContainerStarted","Data":"942bdc4a37858f977897c029f6418ee44e764beda839c691fc5762adfc207fba"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.335208 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:05 crc kubenswrapper[4846]: E0202 12:12:05.336871 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.836857767 +0000 UTC m=+157.065444630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.355353 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wbtlv" podStartSLOduration=131.355334251 podStartE2EDuration="2m11.355334251s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.354013063 +0000 UTC m=+156.582599936" watchObservedRunningTime="2026-02-02 12:12:05.355334251 +0000 UTC m=+156.583921124" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.370651 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" event={"ID":"459bf85f-e304-4481-a4cb-2c391953d6f5","Type":"ContainerStarted","Data":"7c345794ae546ab7368a7d87fdf2efa8a8b65574ec893f3544b15587ebb5c406"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.392895 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" event={"ID":"8c8cc0f5-61a4-482e-9413-675bc26d1781","Type":"ContainerStarted","Data":"34581ff387a7e9a8ded439f099687e77f40f0e0f8f8b7fced45e399816b73393"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.394668 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.409850 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jn5pn" event={"ID":"b3b18c12-56ac-4a04-bcda-f261a6446f24","Type":"ContainerStarted","Data":"e8f4fd0a3cc8e93eaf982970c178d771ad2d7eac0006a6251d58afd58330608e"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.409226 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ff7cw" podStartSLOduration=131.409201478 podStartE2EDuration="2m11.409201478s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.40512717 +0000 UTC m=+156.633714043" watchObservedRunningTime="2026-02-02 12:12:05.409201478 +0000 UTC m=+156.637788341" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.436882 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:05 crc kubenswrapper[4846]: E0202 12:12:05.438438 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:05.938414923 +0000 UTC m=+157.167001796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.474098 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2472r" event={"ID":"5d0f31bd-b822-453c-8ed2-12d0e348946c","Type":"ContainerStarted","Data":"8038e126c3c1aae03c6b6f008d9c798993bd44a32d3a5abbe9bbeae4450ea0fd"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.488384 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" event={"ID":"52772605-2b37-4e17-a66a-5456e6e27d47","Type":"ContainerStarted","Data":"11c81fa1db83883eed0d495891e8c4c77d49780849ef1a7c4b93e5ff2b8f190d"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.488843 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" event={"ID":"52772605-2b37-4e17-a66a-5456e6e27d47","Type":"ContainerStarted","Data":"7861c9777f941f3891091aa87f3dcb0fbd9497645995c15532590c0f0b1059b2"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.488862 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" event={"ID":"52772605-2b37-4e17-a66a-5456e6e27d47","Type":"ContainerStarted","Data":"79c09a402cbf3b0e3e6fe15cfb1b5a63798ecffd8faa80f8f06c78c6473e5ed0"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.490587 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hmldg" podStartSLOduration=131.49056719 podStartE2EDuration="2m11.49056719s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.485801543 +0000 UTC m=+156.714388416" watchObservedRunningTime="2026-02-02 12:12:05.49056719 +0000 UTC m=+156.719154053" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.490941 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" podStartSLOduration=132.490936311 podStartE2EDuration="2m12.490936311s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.455446295 +0000 UTC m=+156.684033168" watchObservedRunningTime="2026-02-02 12:12:05.490936311 +0000 UTC m=+156.719523174" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.497255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kbvlq" event={"ID":"618d005a-d18f-4ad9-92ba-cb1a97a9cebc","Type":"ContainerStarted","Data":"ee0bb9206a4db3528434ae0fda64c6951ad485c38497fca8ae29ad5942899907"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.497297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kbvlq" event={"ID":"618d005a-d18f-4ad9-92ba-cb1a97a9cebc","Type":"ContainerStarted","Data":"1b2c50fd1f0c2cd535d0d0d944684a4e7e1ff5fa951a55ed71a041f2b5a54448"} Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.541732 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:05 crc kubenswrapper[4846]: E0202 12:12:05.544466 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.044452118 +0000 UTC m=+157.273038981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.546171 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" podStartSLOduration=131.546151727 podStartE2EDuration="2m11.546151727s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.540520254 +0000 UTC m=+156.769107117" watchObservedRunningTime="2026-02-02 12:12:05.546151727 +0000 UTC m=+156.774738600" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.556996 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.557144 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.590844 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" podStartSLOduration=132.590830329 podStartE2EDuration="2m12.590830329s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.589080908 +0000 UTC m=+156.817667781" watchObservedRunningTime="2026-02-02 12:12:05.590830329 +0000 UTC m=+156.819417182" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.593408 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ph4nq" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.608850 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.643966 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:05 crc kubenswrapper[4846]: E0202 12:12:05.645822 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.145802858 +0000 UTC m=+157.374389731 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.668629 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" podStartSLOduration=131.668602757 podStartE2EDuration="2m11.668602757s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.641314308 +0000 UTC m=+156.869901171" watchObservedRunningTime="2026-02-02 12:12:05.668602757 +0000 UTC m=+156.897189620" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.684997 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n7rfv" podStartSLOduration=131.68497712 podStartE2EDuration="2m11.68497712s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.684936709 +0000 UTC m=+156.913523572" watchObservedRunningTime="2026-02-02 12:12:05.68497712 +0000 UTC m=+156.913563973" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.710604 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.711082 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.746520 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:05 crc kubenswrapper[4846]: E0202 12:12:05.746836 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.246824648 +0000 UTC m=+157.475411501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.788694 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-2472r" podStartSLOduration=131.788677098 podStartE2EDuration="2m11.788677098s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.720993321 +0000 UTC m=+156.949580184" watchObservedRunningTime="2026-02-02 12:12:05.788677098 +0000 UTC m=+157.017263951" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.788799 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-jn5pn" podStartSLOduration=7.788794721 podStartE2EDuration="7.788794721s" podCreationTimestamp="2026-02-02 12:11:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:05.788007109 +0000 UTC m=+157.016593972" watchObservedRunningTime="2026-02-02 12:12:05.788794721 +0000 UTC m=+157.017381594" Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.849332 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:05 crc kubenswrapper[4846]: E0202 12:12:05.849697 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.349678831 +0000 UTC m=+157.578265694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:05 crc kubenswrapper[4846]: I0202 12:12:05.950424 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:05 crc kubenswrapper[4846]: E0202 12:12:05.950828 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.450813275 +0000 UTC m=+157.679400138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.031932 4846 patch_prober.go:28] interesting pod/router-default-5444994796-h77rj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 12:12:06 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Feb 02 12:12:06 crc kubenswrapper[4846]: [+]process-running ok Feb 02 12:12:06 crc kubenswrapper[4846]: healthz check failed Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.032008 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h77rj" podUID="47a375de-20fc-4cd3-a224-b3fd6cd088c1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.051752 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.051911 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.551887916 +0000 UTC m=+157.780474779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.052003 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.052325 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.552313759 +0000 UTC m=+157.780900612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.153002 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.153212 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.653173475 +0000 UTC m=+157.881760338 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.153361 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.153723 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.6537115 +0000 UTC m=+157.882298363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.254990 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.255185 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.755153083 +0000 UTC m=+157.983739956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.255287 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.255693 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.755678648 +0000 UTC m=+157.984265511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.335515 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5glmk" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.356089 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.356310 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.856285236 +0000 UTC m=+158.084872099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.356510 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.356851 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.856839012 +0000 UTC m=+158.085425875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.462596 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.462817 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.962785015 +0000 UTC m=+158.191371878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.462992 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.464166 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:06.964158545 +0000 UTC m=+158.192745408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.509497 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kbvlq" event={"ID":"618d005a-d18f-4ad9-92ba-cb1a97a9cebc","Type":"ContainerStarted","Data":"f0e8cd26f8587a793b64516448a58ccdfd3d5ec7398638548f1b6c26697485d6"} Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.509662 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-kbvlq" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.512753 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" event={"ID":"b3e406b8-473a-40f4-b464-e6f9992ab0c5","Type":"ContainerStarted","Data":"5a1b34ff70439df079aea3843e32c117ee34b4980980381232127c0e09cc6861"} Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.512803 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" event={"ID":"b3e406b8-473a-40f4-b464-e6f9992ab0c5","Type":"ContainerStarted","Data":"5e273671d75f1beb3463080f8f328c7383a02c31bdfc760530eeef9f71a86c8f"} Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.513771 4846 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zbpnp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.513807 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" podUID="34dedc1d-10d1-4ca0-b059-99048ba37464" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.522813 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2984r" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.545905 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s8wzs" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.556619 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-kbvlq" podStartSLOduration=8.556597937 podStartE2EDuration="8.556597937s" podCreationTimestamp="2026-02-02 12:11:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:06.553989091 +0000 UTC m=+157.782575974" watchObservedRunningTime="2026-02-02 12:12:06.556597937 +0000 UTC m=+157.785184800" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.564209 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.564539 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.064517005 +0000 UTC m=+158.293103868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.665532 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.666882 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.166870615 +0000 UTC m=+158.395457478 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.767289 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.767701 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.267672149 +0000 UTC m=+158.496259012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.868604 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.868899 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.368888494 +0000 UTC m=+158.597475357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.881213 4846 patch_prober.go:28] interesting pod/apiserver-76f77b778f-9h9tw container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]log ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]etcd ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]poststarthook/generic-apiserver-start-informers ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]poststarthook/max-in-flight-filter ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 02 12:12:06 crc kubenswrapper[4846]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 02 12:12:06 crc kubenswrapper[4846]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Feb 02 12:12:06 crc kubenswrapper[4846]: [+]poststarthook/project.openshift.io-projectcache ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]poststarthook/openshift.io-startinformers ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 02 12:12:06 crc kubenswrapper[4846]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 02 12:12:06 crc kubenswrapper[4846]: livez check failed Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.881267 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" podUID="459bf85f-e304-4481-a4cb-2c391953d6f5" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.886987 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pjvpk"] Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.887954 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.892896 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.966916 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pjvpk"] Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.969272 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.969476 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.469447861 +0000 UTC m=+158.698034724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.969706 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-utilities\") pod \"certified-operators-pjvpk\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.969929 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt56f\" (UniqueName: \"kubernetes.io/projected/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-kube-api-access-pt56f\") pod \"certified-operators-pjvpk\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.970148 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-catalog-content\") pod \"certified-operators-pjvpk\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:06 crc kubenswrapper[4846]: I0202 12:12:06.970315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:06 crc kubenswrapper[4846]: E0202 12:12:06.970730 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.470710458 +0000 UTC m=+158.699297321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.029378 4846 patch_prober.go:28] interesting pod/router-default-5444994796-h77rj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 12:12:07 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Feb 02 12:12:07 crc kubenswrapper[4846]: [+]process-running ok Feb 02 12:12:07 crc kubenswrapper[4846]: healthz check failed Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.029451 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h77rj" podUID="47a375de-20fc-4cd3-a224-b3fd6cd088c1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.073218 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.073423 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-catalog-content\") pod \"certified-operators-pjvpk\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.073482 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-utilities\") pod \"certified-operators-pjvpk\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.073510 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt56f\" (UniqueName: \"kubernetes.io/projected/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-kube-api-access-pt56f\") pod \"certified-operators-pjvpk\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.074097 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-catalog-content\") pod \"certified-operators-pjvpk\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.074149 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-utilities\") pod \"certified-operators-pjvpk\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:07 crc kubenswrapper[4846]: E0202 12:12:07.074222 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.574198779 +0000 UTC m=+158.802785702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.107672 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gmnp2"] Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.108866 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.113756 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.117862 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt56f\" (UniqueName: \"kubernetes.io/projected/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-kube-api-access-pt56f\") pod \"certified-operators-pjvpk\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.186415 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:07 crc kubenswrapper[4846]: E0202 12:12:07.186860 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.686845565 +0000 UTC m=+158.915432428 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.207468 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.218238 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gmnp2"] Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.287184 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.287449 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8rtj\" (UniqueName: \"kubernetes.io/projected/98f32d89-a132-4cb9-a647-b48848e1b18a-kube-api-access-s8rtj\") pod \"community-operators-gmnp2\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.287521 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-catalog-content\") pod \"community-operators-gmnp2\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.287546 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-utilities\") pod \"community-operators-gmnp2\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: E0202 12:12:07.287716 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.7876904 +0000 UTC m=+159.016277263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.304734 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zn89z"] Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.306584 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.322574 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zn89z"] Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.390356 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8rtj\" (UniqueName: \"kubernetes.io/projected/98f32d89-a132-4cb9-a647-b48848e1b18a-kube-api-access-s8rtj\") pod \"community-operators-gmnp2\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.390438 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.390479 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-catalog-content\") pod \"community-operators-gmnp2\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.390509 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-utilities\") pod \"community-operators-gmnp2\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: E0202 12:12:07.391020 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.891002417 +0000 UTC m=+159.119589280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.391031 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-utilities\") pod \"community-operators-gmnp2\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.391284 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-catalog-content\") pod \"community-operators-gmnp2\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.484618 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8rtj\" (UniqueName: \"kubernetes.io/projected/98f32d89-a132-4cb9-a647-b48848e1b18a-kube-api-access-s8rtj\") pod \"community-operators-gmnp2\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.488472 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ksr5l"] Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.495857 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.493285 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:07 crc kubenswrapper[4846]: E0202 12:12:07.493362 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.993342985 +0000 UTC m=+159.221929858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.496761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.496871 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-catalog-content\") pod \"certified-operators-zn89z\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.496985 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72mv4\" (UniqueName: \"kubernetes.io/projected/99680b5d-496f-4819-a517-1b1ead7e403d-kube-api-access-72mv4\") pod \"certified-operators-zn89z\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.497117 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-utilities\") pod \"certified-operators-zn89z\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: E0202 12:12:07.497980 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:07.997960129 +0000 UTC m=+159.226546992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.513672 4846 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-p8mks container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.513778 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" podUID="07661a94-bbd8-4b80-842c-d9db04206967" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.514328 4846 patch_prober.go:28] interesting pod/console-operator-58897d9998-zgkpl container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.514358 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-zgkpl" podUID="b30d0b23-8b4d-4863-93e9-5afd111cb2c9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.525725 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ksr5l"] Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.542760 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" event={"ID":"b3e406b8-473a-40f4-b464-e6f9992ab0c5","Type":"ContainerStarted","Data":"2724db0a2b4df695409028dc735b85d8bbad1a12d854e0be228381e087e948eb"} Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.543947 4846 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zbpnp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.544001 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" podUID="34dedc1d-10d1-4ca0-b059-99048ba37464" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.599157 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.599361 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-utilities\") pod \"certified-operators-zn89z\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.599418 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj4s2\" (UniqueName: \"kubernetes.io/projected/c8056015-9fef-462c-b839-b2171e1efa4b-kube-api-access-dj4s2\") pod \"community-operators-ksr5l\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.599454 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-utilities\") pod \"community-operators-ksr5l\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.599486 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-catalog-content\") pod \"certified-operators-zn89z\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.599530 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72mv4\" (UniqueName: \"kubernetes.io/projected/99680b5d-496f-4819-a517-1b1ead7e403d-kube-api-access-72mv4\") pod \"certified-operators-zn89z\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.599561 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-catalog-content\") pod \"community-operators-ksr5l\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: E0202 12:12:07.599714 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:08.09969665 +0000 UTC m=+159.328283513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.600166 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-utilities\") pod \"certified-operators-zn89z\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.600436 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-catalog-content\") pod \"certified-operators-zn89z\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.611757 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-zgkpl" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.632872 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72mv4\" (UniqueName: \"kubernetes.io/projected/99680b5d-496f-4819-a517-1b1ead7e403d-kube-api-access-72mv4\") pod \"certified-operators-zn89z\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.638918 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.710705 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj4s2\" (UniqueName: \"kubernetes.io/projected/c8056015-9fef-462c-b839-b2171e1efa4b-kube-api-access-dj4s2\") pod \"community-operators-ksr5l\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.710784 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.710811 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-utilities\") pod \"community-operators-ksr5l\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.711120 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-catalog-content\") pod \"community-operators-ksr5l\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: E0202 12:12:07.712880 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:08.212864051 +0000 UTC m=+159.441450914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.713443 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-utilities\") pod \"community-operators-ksr5l\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.717199 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-catalog-content\") pod \"community-operators-ksr5l\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.724367 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-p8mks" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.759578 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.792949 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj4s2\" (UniqueName: \"kubernetes.io/projected/c8056015-9fef-462c-b839-b2171e1efa4b-kube-api-access-dj4s2\") pod \"community-operators-ksr5l\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.818104 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:07 crc kubenswrapper[4846]: E0202 12:12:07.818598 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:08.318575177 +0000 UTC m=+159.547162040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.833331 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:07 crc kubenswrapper[4846]: I0202 12:12:07.920474 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:07 crc kubenswrapper[4846]: E0202 12:12:07.921036 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:08.421022288 +0000 UTC m=+159.649609151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.022575 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:08 crc kubenswrapper[4846]: E0202 12:12:08.024067 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:08.524044157 +0000 UTC m=+159.752631020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.040343 4846 patch_prober.go:28] interesting pod/router-default-5444994796-h77rj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 12:12:08 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Feb 02 12:12:08 crc kubenswrapper[4846]: [+]process-running ok Feb 02 12:12:08 crc kubenswrapper[4846]: healthz check failed Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.040581 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h77rj" podUID="47a375de-20fc-4cd3-a224-b3fd6cd088c1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.125035 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:08 crc kubenswrapper[4846]: E0202 12:12:08.125374 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:08.625362685 +0000 UTC m=+159.853949548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.162412 4846 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.226282 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:08 crc kubenswrapper[4846]: E0202 12:12:08.229430 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:08.729398813 +0000 UTC m=+159.957985676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.259356 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pjvpk"] Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.330370 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:08 crc kubenswrapper[4846]: E0202 12:12:08.330843 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:08.830822685 +0000 UTC m=+160.059409598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.423464 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zn89z"] Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.430857 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:08 crc kubenswrapper[4846]: E0202 12:12:08.431179 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 12:12:08.931164056 +0000 UTC m=+160.159750919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.491957 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gmnp2"] Feb 02 12:12:08 crc kubenswrapper[4846]: W0202 12:12:08.499099 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98f32d89_a132_4cb9_a647_b48848e1b18a.slice/crio-69be1edd328643017aecd7b67b5a8ff66fe81d623b7c17fd80725a345df3cab3 WatchSource:0}: Error finding container 69be1edd328643017aecd7b67b5a8ff66fe81d623b7c17fd80725a345df3cab3: Status 404 returned error can't find the container with id 69be1edd328643017aecd7b67b5a8ff66fe81d623b7c17fd80725a345df3cab3 Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.533156 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ksr5l"] Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.534384 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:08 crc kubenswrapper[4846]: E0202 12:12:08.534675 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 12:12:09.034664248 +0000 UTC m=+160.263251111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-94n5d" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.560733 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn89z" event={"ID":"99680b5d-496f-4819-a517-1b1ead7e403d","Type":"ContainerStarted","Data":"4be44f2d733444f8872a0cbc5df0e2420d1bc1fecad3b70850ad17c93638f102"} Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.566694 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" event={"ID":"b3e406b8-473a-40f4-b464-e6f9992ab0c5","Type":"ContainerStarted","Data":"d1602b6307e80a0cc3085cff6e93e4604d295170aae84380200bda17a6d4d50d"} Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.569550 4846 generic.go:334] "Generic (PLEG): container finished" podID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerID="ca650255f3ade6088a95b4752a6cb7eddcfd18267a077a6424808a5e59c670ae" exitCode=0 Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.569683 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjvpk" event={"ID":"3b5b4c6f-368c-47ba-87fc-7309cac15d9b","Type":"ContainerDied","Data":"ca650255f3ade6088a95b4752a6cb7eddcfd18267a077a6424808a5e59c670ae"} Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.569717 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjvpk" event={"ID":"3b5b4c6f-368c-47ba-87fc-7309cac15d9b","Type":"ContainerStarted","Data":"4f7f72c8ef672a5c5bd6ac0baa3bd80e9c60b0baad71e44681d854424fda1175"} Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.572170 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.572498 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmnp2" event={"ID":"98f32d89-a132-4cb9-a647-b48848e1b18a","Type":"ContainerStarted","Data":"69be1edd328643017aecd7b67b5a8ff66fe81d623b7c17fd80725a345df3cab3"} Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.606554 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-h6lcv" podStartSLOduration=10.606532565 podStartE2EDuration="10.606532565s" podCreationTimestamp="2026-02-02 12:11:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:08.585038774 +0000 UTC m=+159.813625667" watchObservedRunningTime="2026-02-02 12:12:08.606532565 +0000 UTC m=+159.835119438" Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.611491 4846 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-02T12:12:08.162441747Z","Handler":null,"Name":""} Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.615826 4846 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.615870 4846 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.638410 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.656447 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.739847 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.743176 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.743279 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:08 crc kubenswrapper[4846]: I0202 12:12:08.766476 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-94n5d\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.027276 4846 patch_prober.go:28] interesting pod/router-default-5444994796-h77rj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 12:12:09 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Feb 02 12:12:09 crc kubenswrapper[4846]: [+]process-running ok Feb 02 12:12:09 crc kubenswrapper[4846]: healthz check failed Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.027351 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h77rj" podUID="47a375de-20fc-4cd3-a224-b3fd6cd088c1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.029729 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.065982 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p558w"] Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.069944 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.072503 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.075791 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p558w"] Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.145566 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-utilities\") pod \"redhat-marketplace-p558w\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.145908 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4rgz\" (UniqueName: \"kubernetes.io/projected/14da97a9-4b68-458d-858a-6ba8b67cb749-kube-api-access-b4rgz\") pod \"redhat-marketplace-p558w\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.145947 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-catalog-content\") pod \"redhat-marketplace-p558w\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.248266 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-utilities\") pod \"redhat-marketplace-p558w\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.248324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4rgz\" (UniqueName: \"kubernetes.io/projected/14da97a9-4b68-458d-858a-6ba8b67cb749-kube-api-access-b4rgz\") pod \"redhat-marketplace-p558w\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.248356 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-catalog-content\") pod \"redhat-marketplace-p558w\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.248941 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-catalog-content\") pod \"redhat-marketplace-p558w\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.248956 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-utilities\") pod \"redhat-marketplace-p558w\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.268743 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4rgz\" (UniqueName: \"kubernetes.io/projected/14da97a9-4b68-458d-858a-6ba8b67cb749-kube-api-access-b4rgz\") pod \"redhat-marketplace-p558w\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.406175 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.439380 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.468315 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94n5d"] Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.471145 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cpq4q"] Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.474930 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.494509 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cpq4q"] Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.553827 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-catalog-content\") pod \"redhat-marketplace-cpq4q\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.554156 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-utilities\") pod \"redhat-marketplace-cpq4q\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.554180 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jf5l\" (UniqueName: \"kubernetes.io/projected/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-kube-api-access-8jf5l\") pod \"redhat-marketplace-cpq4q\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.583821 4846 generic.go:334] "Generic (PLEG): container finished" podID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerID="df9511754728df7f993f5220d3ef474477a76645511eb862a22a87ad404493d6" exitCode=0 Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.583885 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmnp2" event={"ID":"98f32d89-a132-4cb9-a647-b48848e1b18a","Type":"ContainerDied","Data":"df9511754728df7f993f5220d3ef474477a76645511eb862a22a87ad404493d6"} Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.587431 4846 generic.go:334] "Generic (PLEG): container finished" podID="c8056015-9fef-462c-b839-b2171e1efa4b" containerID="15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430" exitCode=0 Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.587494 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksr5l" event={"ID":"c8056015-9fef-462c-b839-b2171e1efa4b","Type":"ContainerDied","Data":"15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430"} Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.587523 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksr5l" event={"ID":"c8056015-9fef-462c-b839-b2171e1efa4b","Type":"ContainerStarted","Data":"97c1292ca4a6a65a75204ea33784dda2642129af7f5eb5124ce46bd2e74eb19d"} Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.599573 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" event={"ID":"59df79f3-cb9f-412c-bc01-7f3376620c17","Type":"ContainerStarted","Data":"a0e06192f0cc45669a0c281bbf5530f2d20d6fef56ffc566f04d4af86cd85757"} Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.603193 4846 generic.go:334] "Generic (PLEG): container finished" podID="99680b5d-496f-4819-a517-1b1ead7e403d" containerID="a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2" exitCode=0 Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.604078 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn89z" event={"ID":"99680b5d-496f-4819-a517-1b1ead7e403d","Type":"ContainerDied","Data":"a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2"} Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.656211 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-catalog-content\") pod \"redhat-marketplace-cpq4q\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.656268 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-utilities\") pod \"redhat-marketplace-cpq4q\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.656296 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jf5l\" (UniqueName: \"kubernetes.io/projected/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-kube-api-access-8jf5l\") pod \"redhat-marketplace-cpq4q\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.657053 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-catalog-content\") pod \"redhat-marketplace-cpq4q\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.665753 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-utilities\") pod \"redhat-marketplace-cpq4q\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.678081 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p558w"] Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.701965 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jf5l\" (UniqueName: \"kubernetes.io/projected/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-kube-api-access-8jf5l\") pod \"redhat-marketplace-cpq4q\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:09 crc kubenswrapper[4846]: I0202 12:12:09.835436 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.027387 4846 patch_prober.go:28] interesting pod/router-default-5444994796-h77rj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 12:12:10 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Feb 02 12:12:10 crc kubenswrapper[4846]: [+]process-running ok Feb 02 12:12:10 crc kubenswrapper[4846]: healthz check failed Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.027743 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h77rj" podUID="47a375de-20fc-4cd3-a224-b3fd6cd088c1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.072874 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n44pv"] Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.074033 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.079172 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.088100 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n44pv"] Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.265807 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-utilities\") pod \"redhat-operators-n44pv\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.265874 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhzjh\" (UniqueName: \"kubernetes.io/projected/64a47758-13e5-4a28-b781-2bdf84b870fd-kube-api-access-lhzjh\") pod \"redhat-operators-n44pv\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.265997 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-catalog-content\") pod \"redhat-operators-n44pv\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.307160 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cpq4q"] Feb 02 12:12:10 crc kubenswrapper[4846]: W0202 12:12:10.313264 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e90d3b0_496a_4f62_bdd7_5018fc11da0e.slice/crio-21c50224d365a94bab1a9ea1955a60298626dafc606a716692b0910086e4e30f WatchSource:0}: Error finding container 21c50224d365a94bab1a9ea1955a60298626dafc606a716692b0910086e4e30f: Status 404 returned error can't find the container with id 21c50224d365a94bab1a9ea1955a60298626dafc606a716692b0910086e4e30f Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.367850 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-utilities\") pod \"redhat-operators-n44pv\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.367902 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhzjh\" (UniqueName: \"kubernetes.io/projected/64a47758-13e5-4a28-b781-2bdf84b870fd-kube-api-access-lhzjh\") pod \"redhat-operators-n44pv\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.367964 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-catalog-content\") pod \"redhat-operators-n44pv\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.368488 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-utilities\") pod \"redhat-operators-n44pv\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.368531 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-catalog-content\") pod \"redhat-operators-n44pv\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.389863 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhzjh\" (UniqueName: \"kubernetes.io/projected/64a47758-13e5-4a28-b781-2bdf84b870fd-kube-api-access-lhzjh\") pod \"redhat-operators-n44pv\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.390992 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.489097 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g2t4b"] Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.490384 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.535249 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g2t4b"] Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.632971 4846 generic.go:334] "Generic (PLEG): container finished" podID="7d559728-5463-4a87-b4e7-b767dfe50bd7" containerID="b77000692f2fa0089b2dc53b65e836c076ed0cf6d08245baa4599490ad60038d" exitCode=0 Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.633030 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" event={"ID":"7d559728-5463-4a87-b4e7-b767dfe50bd7","Type":"ContainerDied","Data":"b77000692f2fa0089b2dc53b65e836c076ed0cf6d08245baa4599490ad60038d"} Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.644804 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpq4q" event={"ID":"1e90d3b0-496a-4f62-bdd7-5018fc11da0e","Type":"ContainerStarted","Data":"21c50224d365a94bab1a9ea1955a60298626dafc606a716692b0910086e4e30f"} Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.651027 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" event={"ID":"59df79f3-cb9f-412c-bc01-7f3376620c17","Type":"ContainerStarted","Data":"b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee"} Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.651829 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.653250 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.653975 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.663419 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.663657 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.667460 4846 generic.go:334] "Generic (PLEG): container finished" podID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerID="b9cc2ecbfff30ad71954731617456c9ecef7ae873e3e59fc57abdc4504923c5d" exitCode=0 Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.667499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p558w" event={"ID":"14da97a9-4b68-458d-858a-6ba8b67cb749","Type":"ContainerDied","Data":"b9cc2ecbfff30ad71954731617456c9ecef7ae873e3e59fc57abdc4504923c5d"} Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.667521 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p558w" event={"ID":"14da97a9-4b68-458d-858a-6ba8b67cb749","Type":"ContainerStarted","Data":"a6eb9a138cf55764de1184b906468c972c1de9aa26bd531c2e7a8869bd60da72"} Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.668144 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.692477 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-utilities\") pod \"redhat-operators-g2t4b\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.692595 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-catalog-content\") pod \"redhat-operators-g2t4b\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.692643 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e08e80c6-fdfb-48cc-a752-9650f31f249e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e08e80c6-fdfb-48cc-a752-9650f31f249e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.692695 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e08e80c6-fdfb-48cc-a752-9650f31f249e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e08e80c6-fdfb-48cc-a752-9650f31f249e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.692720 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5skb\" (UniqueName: \"kubernetes.io/projected/04ce875b-335c-4dc9-a1d8-c2bd1f269675-kube-api-access-k5skb\") pod \"redhat-operators-g2t4b\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.719356 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" podStartSLOduration=136.719341641 podStartE2EDuration="2m16.719341641s" podCreationTimestamp="2026-02-02 12:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:10.718930819 +0000 UTC m=+161.947517682" watchObservedRunningTime="2026-02-02 12:12:10.719341641 +0000 UTC m=+161.947928494" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.725840 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.740563 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-9h9tw" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.783323 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-bckqh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.783372 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-bckqh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.783420 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bckqh" podUID="4157225e-a9ff-4cbe-b56b-43d309030abd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.791975 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bckqh" podUID="4157225e-a9ff-4cbe-b56b-43d309030abd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.793367 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-catalog-content\") pod \"redhat-operators-g2t4b\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.793414 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e08e80c6-fdfb-48cc-a752-9650f31f249e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e08e80c6-fdfb-48cc-a752-9650f31f249e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.793496 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e08e80c6-fdfb-48cc-a752-9650f31f249e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e08e80c6-fdfb-48cc-a752-9650f31f249e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.793523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5skb\" (UniqueName: \"kubernetes.io/projected/04ce875b-335c-4dc9-a1d8-c2bd1f269675-kube-api-access-k5skb\") pod \"redhat-operators-g2t4b\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.793551 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-utilities\") pod \"redhat-operators-g2t4b\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.798250 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-utilities\") pod \"redhat-operators-g2t4b\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.798297 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-catalog-content\") pod \"redhat-operators-g2t4b\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.798490 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e08e80c6-fdfb-48cc-a752-9650f31f249e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e08e80c6-fdfb-48cc-a752-9650f31f249e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.840298 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5skb\" (UniqueName: \"kubernetes.io/projected/04ce875b-335c-4dc9-a1d8-c2bd1f269675-kube-api-access-k5skb\") pod \"redhat-operators-g2t4b\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.840440 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e08e80c6-fdfb-48cc-a752-9650f31f249e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e08e80c6-fdfb-48cc-a752-9650f31f249e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.881172 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.962911 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n44pv"] Feb 02 12:12:10 crc kubenswrapper[4846]: I0202 12:12:10.999704 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.027620 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.045143 4846 patch_prober.go:28] interesting pod/router-default-5444994796-h77rj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 12:12:11 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Feb 02 12:12:11 crc kubenswrapper[4846]: [+]process-running ok Feb 02 12:12:11 crc kubenswrapper[4846]: healthz check failed Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.045206 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h77rj" podUID="47a375de-20fc-4cd3-a224-b3fd6cd088c1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.357803 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.361349 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g2t4b"] Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.709797 4846 generic.go:334] "Generic (PLEG): container finished" podID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerID="241687a3f8c362a669e31bf5a7d8ee6c5ad23aed00c529b123b4627ade5737a7" exitCode=0 Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.710142 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n44pv" event={"ID":"64a47758-13e5-4a28-b781-2bdf84b870fd","Type":"ContainerDied","Data":"241687a3f8c362a669e31bf5a7d8ee6c5ad23aed00c529b123b4627ade5737a7"} Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.710168 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n44pv" event={"ID":"64a47758-13e5-4a28-b781-2bdf84b870fd","Type":"ContainerStarted","Data":"7f8e6c51540063226181c7b4194d87405b772d3931aa495b8397174a4e8c5171"} Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.717365 4846 generic.go:334] "Generic (PLEG): container finished" podID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerID="83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b" exitCode=0 Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.717414 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpq4q" event={"ID":"1e90d3b0-496a-4f62-bdd7-5018fc11da0e","Type":"ContainerDied","Data":"83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b"} Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.729585 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2t4b" event={"ID":"04ce875b-335c-4dc9-a1d8-c2bd1f269675","Type":"ContainerStarted","Data":"82fe4b1ee571afcfbf27877e4dad476365e5c20d346e397e02ee74fe024ac827"} Feb 02 12:12:11 crc kubenswrapper[4846]: I0202 12:12:11.759104 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 02 12:12:11 crc kubenswrapper[4846]: W0202 12:12:11.862105 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode08e80c6_fdfb_48cc_a752_9650f31f249e.slice/crio-eaa9f2f224c08420b2e1b741cfa1f01d671d0769adc74917341dd956ac483836 WatchSource:0}: Error finding container eaa9f2f224c08420b2e1b741cfa1f01d671d0769adc74917341dd956ac483836: Status 404 returned error can't find the container with id eaa9f2f224c08420b2e1b741cfa1f01d671d0769adc74917341dd956ac483836 Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.039033 4846 patch_prober.go:28] interesting pod/router-default-5444994796-h77rj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 12:12:12 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Feb 02 12:12:12 crc kubenswrapper[4846]: [+]process-running ok Feb 02 12:12:12 crc kubenswrapper[4846]: healthz check failed Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.039408 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h77rj" podUID="47a375de-20fc-4cd3-a224-b3fd6cd088c1" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.222025 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.272326 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.272371 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.286284 4846 patch_prober.go:28] interesting pod/console-f9d7485db-79skc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.286354 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-79skc" podUID="af51b9e4-0ee3-4bf9-93e5-27b4039be121" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.337576 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d559728-5463-4a87-b4e7-b767dfe50bd7-config-volume\") pod \"7d559728-5463-4a87-b4e7-b767dfe50bd7\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.337652 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pm6v\" (UniqueName: \"kubernetes.io/projected/7d559728-5463-4a87-b4e7-b767dfe50bd7-kube-api-access-6pm6v\") pod \"7d559728-5463-4a87-b4e7-b767dfe50bd7\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.337680 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d559728-5463-4a87-b4e7-b767dfe50bd7-secret-volume\") pod \"7d559728-5463-4a87-b4e7-b767dfe50bd7\" (UID: \"7d559728-5463-4a87-b4e7-b767dfe50bd7\") " Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.338331 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d559728-5463-4a87-b4e7-b767dfe50bd7-config-volume" (OuterVolumeSpecName: "config-volume") pod "7d559728-5463-4a87-b4e7-b767dfe50bd7" (UID: "7d559728-5463-4a87-b4e7-b767dfe50bd7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.347063 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d559728-5463-4a87-b4e7-b767dfe50bd7-kube-api-access-6pm6v" (OuterVolumeSpecName: "kube-api-access-6pm6v") pod "7d559728-5463-4a87-b4e7-b767dfe50bd7" (UID: "7d559728-5463-4a87-b4e7-b767dfe50bd7"). InnerVolumeSpecName "kube-api-access-6pm6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.358837 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d559728-5463-4a87-b4e7-b767dfe50bd7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7d559728-5463-4a87-b4e7-b767dfe50bd7" (UID: "7d559728-5463-4a87-b4e7-b767dfe50bd7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.441496 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d559728-5463-4a87-b4e7-b767dfe50bd7-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.441541 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pm6v\" (UniqueName: \"kubernetes.io/projected/7d559728-5463-4a87-b4e7-b767dfe50bd7-kube-api-access-6pm6v\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.441582 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d559728-5463-4a87-b4e7-b767dfe50bd7-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.752095 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" event={"ID":"7d559728-5463-4a87-b4e7-b767dfe50bd7","Type":"ContainerDied","Data":"942bdc4a37858f977897c029f6418ee44e764beda839c691fc5762adfc207fba"} Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.752562 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="942bdc4a37858f977897c029f6418ee44e764beda839c691fc5762adfc207fba" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.752661 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk" Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.778477 4846 generic.go:334] "Generic (PLEG): container finished" podID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerID="c3a90734817c2f0baa3cbb295141d9b51ab8cbd5d79c74857a530dd4989a1b68" exitCode=0 Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.778541 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2t4b" event={"ID":"04ce875b-335c-4dc9-a1d8-c2bd1f269675","Type":"ContainerDied","Data":"c3a90734817c2f0baa3cbb295141d9b51ab8cbd5d79c74857a530dd4989a1b68"} Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.795943 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e08e80c6-fdfb-48cc-a752-9650f31f249e","Type":"ContainerStarted","Data":"e00ca53596a874c7c98fd6463f98eaeab923b6f5b78264503847cf6cd63b60e1"} Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.795985 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e08e80c6-fdfb-48cc-a752-9650f31f249e","Type":"ContainerStarted","Data":"eaa9f2f224c08420b2e1b741cfa1f01d671d0769adc74917341dd956ac483836"} Feb 02 12:12:12 crc kubenswrapper[4846]: I0202 12:12:12.830041 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.830002404 podStartE2EDuration="2.830002404s" podCreationTimestamp="2026-02-02 12:12:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:12.828547122 +0000 UTC m=+164.057133985" watchObservedRunningTime="2026-02-02 12:12:12.830002404 +0000 UTC m=+164.058589267" Feb 02 12:12:13 crc kubenswrapper[4846]: I0202 12:12:13.032837 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:12:13 crc kubenswrapper[4846]: I0202 12:12:13.038438 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-h77rj" Feb 02 12:12:13 crc kubenswrapper[4846]: I0202 12:12:13.812540 4846 generic.go:334] "Generic (PLEG): container finished" podID="e08e80c6-fdfb-48cc-a752-9650f31f249e" containerID="e00ca53596a874c7c98fd6463f98eaeab923b6f5b78264503847cf6cd63b60e1" exitCode=0 Feb 02 12:12:13 crc kubenswrapper[4846]: I0202 12:12:13.813502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e08e80c6-fdfb-48cc-a752-9650f31f249e","Type":"ContainerDied","Data":"e00ca53596a874c7c98fd6463f98eaeab923b6f5b78264503847cf6cd63b60e1"} Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.045064 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 02 12:12:15 crc kubenswrapper[4846]: E0202 12:12:15.049091 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d559728-5463-4a87-b4e7-b767dfe50bd7" containerName="collect-profiles" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.049114 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d559728-5463-4a87-b4e7-b767dfe50bd7" containerName="collect-profiles" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.049253 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d559728-5463-4a87-b4e7-b767dfe50bd7" containerName="collect-profiles" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.049781 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.055090 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.055326 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.064076 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.202197 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e81af364-0c77-4e24-8f1d-720046401479-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e81af364-0c77-4e24-8f1d-720046401479\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.203676 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e81af364-0c77-4e24-8f1d-720046401479-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e81af364-0c77-4e24-8f1d-720046401479\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.305068 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e81af364-0c77-4e24-8f1d-720046401479-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e81af364-0c77-4e24-8f1d-720046401479\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.305134 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e81af364-0c77-4e24-8f1d-720046401479-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e81af364-0c77-4e24-8f1d-720046401479\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.305199 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e81af364-0c77-4e24-8f1d-720046401479-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e81af364-0c77-4e24-8f1d-720046401479\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.323044 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e81af364-0c77-4e24-8f1d-720046401479-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e81af364-0c77-4e24-8f1d-720046401479\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 12:12:15 crc kubenswrapper[4846]: I0202 12:12:15.410859 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 12:12:16 crc kubenswrapper[4846]: I0202 12:12:16.208196 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-kbvlq" Feb 02 12:12:16 crc kubenswrapper[4846]: I0202 12:12:16.420148 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:12:16 crc kubenswrapper[4846]: I0202 12:12:16.424258 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fe216e68-8814-468d-a6f3-fe8721d94eba-metrics-certs\") pod \"network-metrics-daemon-j5m9k\" (UID: \"fe216e68-8814-468d-a6f3-fe8721d94eba\") " pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:12:16 crc kubenswrapper[4846]: I0202 12:12:16.539786 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-j5m9k" Feb 02 12:12:19 crc kubenswrapper[4846]: I0202 12:12:19.931638 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-px4bj_8abc246a-7ca3-4550-ba36-491f830e5419/cluster-samples-operator/0.log" Feb 02 12:12:19 crc kubenswrapper[4846]: I0202 12:12:19.931902 4846 generic.go:334] "Generic (PLEG): container finished" podID="8abc246a-7ca3-4550-ba36-491f830e5419" containerID="933e0a8863d3a600eca001ad3c12750041ece16fe4d65fee335eb17f0fdb58ac" exitCode=2 Feb 02 12:12:19 crc kubenswrapper[4846]: I0202 12:12:19.931931 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" event={"ID":"8abc246a-7ca3-4550-ba36-491f830e5419","Type":"ContainerDied","Data":"933e0a8863d3a600eca001ad3c12750041ece16fe4d65fee335eb17f0fdb58ac"} Feb 02 12:12:19 crc kubenswrapper[4846]: I0202 12:12:19.932355 4846 scope.go:117] "RemoveContainer" containerID="933e0a8863d3a600eca001ad3c12750041ece16fe4d65fee335eb17f0fdb58ac" Feb 02 12:12:20 crc kubenswrapper[4846]: I0202 12:12:20.797122 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-bckqh" Feb 02 12:12:21 crc kubenswrapper[4846]: I0202 12:12:21.452036 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 12:12:21 crc kubenswrapper[4846]: I0202 12:12:21.617153 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e08e80c6-fdfb-48cc-a752-9650f31f249e-kubelet-dir\") pod \"e08e80c6-fdfb-48cc-a752-9650f31f249e\" (UID: \"e08e80c6-fdfb-48cc-a752-9650f31f249e\") " Feb 02 12:12:21 crc kubenswrapper[4846]: I0202 12:12:21.617235 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e08e80c6-fdfb-48cc-a752-9650f31f249e-kube-api-access\") pod \"e08e80c6-fdfb-48cc-a752-9650f31f249e\" (UID: \"e08e80c6-fdfb-48cc-a752-9650f31f249e\") " Feb 02 12:12:21 crc kubenswrapper[4846]: I0202 12:12:21.617289 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e08e80c6-fdfb-48cc-a752-9650f31f249e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e08e80c6-fdfb-48cc-a752-9650f31f249e" (UID: "e08e80c6-fdfb-48cc-a752-9650f31f249e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:12:21 crc kubenswrapper[4846]: I0202 12:12:21.617584 4846 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e08e80c6-fdfb-48cc-a752-9650f31f249e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:21 crc kubenswrapper[4846]: I0202 12:12:21.624474 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e08e80c6-fdfb-48cc-a752-9650f31f249e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e08e80c6-fdfb-48cc-a752-9650f31f249e" (UID: "e08e80c6-fdfb-48cc-a752-9650f31f249e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:21 crc kubenswrapper[4846]: I0202 12:12:21.719237 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e08e80c6-fdfb-48cc-a752-9650f31f249e-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:21 crc kubenswrapper[4846]: I0202 12:12:21.945461 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e08e80c6-fdfb-48cc-a752-9650f31f249e","Type":"ContainerDied","Data":"eaa9f2f224c08420b2e1b741cfa1f01d671d0769adc74917341dd956ac483836"} Feb 02 12:12:21 crc kubenswrapper[4846]: I0202 12:12:21.945491 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 12:12:21 crc kubenswrapper[4846]: I0202 12:12:21.945499 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaa9f2f224c08420b2e1b741cfa1f01d671d0769adc74917341dd956ac483836" Feb 02 12:12:22 crc kubenswrapper[4846]: I0202 12:12:22.279333 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:12:22 crc kubenswrapper[4846]: I0202 12:12:22.284283 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:12:22 crc kubenswrapper[4846]: I0202 12:12:22.796880 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-j5m9k"] Feb 02 12:12:23 crc kubenswrapper[4846]: I0202 12:12:23.302188 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bbmc8"] Feb 02 12:12:23 crc kubenswrapper[4846]: I0202 12:12:23.302764 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" podUID="eba09ec1-c855-493b-8ab3-b40d4db28cc1" containerName="controller-manager" containerID="cri-o://35b11d93c537c1b9170c391cf4becfe22201a1cb6cdbabdf505aa014394503eb" gracePeriod=30 Feb 02 12:12:23 crc kubenswrapper[4846]: I0202 12:12:23.323677 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g"] Feb 02 12:12:23 crc kubenswrapper[4846]: I0202 12:12:23.323921 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" podUID="ec6797b4-3f50-4496-97a9-eb84dd09d5c9" containerName="route-controller-manager" containerID="cri-o://8f468181088598a609c761679bf1da6cc164283ecc0ec927cdb207c614833cec" gracePeriod=30 Feb 02 12:12:23 crc kubenswrapper[4846]: I0202 12:12:23.957696 4846 generic.go:334] "Generic (PLEG): container finished" podID="ec6797b4-3f50-4496-97a9-eb84dd09d5c9" containerID="8f468181088598a609c761679bf1da6cc164283ecc0ec927cdb207c614833cec" exitCode=0 Feb 02 12:12:23 crc kubenswrapper[4846]: I0202 12:12:23.957785 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" event={"ID":"ec6797b4-3f50-4496-97a9-eb84dd09d5c9","Type":"ContainerDied","Data":"8f468181088598a609c761679bf1da6cc164283ecc0ec927cdb207c614833cec"} Feb 02 12:12:23 crc kubenswrapper[4846]: I0202 12:12:23.959399 4846 generic.go:334] "Generic (PLEG): container finished" podID="eba09ec1-c855-493b-8ab3-b40d4db28cc1" containerID="35b11d93c537c1b9170c391cf4becfe22201a1cb6cdbabdf505aa014394503eb" exitCode=0 Feb 02 12:12:23 crc kubenswrapper[4846]: I0202 12:12:23.959434 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" event={"ID":"eba09ec1-c855-493b-8ab3-b40d4db28cc1","Type":"ContainerDied","Data":"35b11d93c537c1b9170c391cf4becfe22201a1cb6cdbabdf505aa014394503eb"} Feb 02 12:12:26 crc kubenswrapper[4846]: W0202 12:12:26.001083 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe216e68_8814_468d_a6f3_fe8721d94eba.slice/crio-a96fd430adea826223ebbb3d3caad571a4d6450ecc93eab54f30f079d95828bc WatchSource:0}: Error finding container a96fd430adea826223ebbb3d3caad571a4d6450ecc93eab54f30f079d95828bc: Status 404 returned error can't find the container with id a96fd430adea826223ebbb3d3caad571a4d6450ecc93eab54f30f079d95828bc Feb 02 12:12:26 crc kubenswrapper[4846]: I0202 12:12:26.149377 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 02 12:12:26 crc kubenswrapper[4846]: I0202 12:12:26.978141 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" event={"ID":"fe216e68-8814-468d-a6f3-fe8721d94eba","Type":"ContainerStarted","Data":"a96fd430adea826223ebbb3d3caad571a4d6450ecc93eab54f30f079d95828bc"} Feb 02 12:12:28 crc kubenswrapper[4846]: I0202 12:12:28.992291 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e81af364-0c77-4e24-8f1d-720046401479","Type":"ContainerStarted","Data":"cd41135a797e39bf7a76839c137555ea462bafb2c2ddea57e207f724310fce87"} Feb 02 12:12:29 crc kubenswrapper[4846]: I0202 12:12:29.036571 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.225345 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.237372 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-client-ca\") pod \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.237432 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-serving-cert\") pod \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.237527 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzzzd\" (UniqueName: \"kubernetes.io/projected/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-kube-api-access-mzzzd\") pod \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.237609 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-config\") pod \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\" (UID: \"ec6797b4-3f50-4496-97a9-eb84dd09d5c9\") " Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.238822 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-client-ca" (OuterVolumeSpecName: "client-ca") pod "ec6797b4-3f50-4496-97a9-eb84dd09d5c9" (UID: "ec6797b4-3f50-4496-97a9-eb84dd09d5c9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.239206 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-config" (OuterVolumeSpecName: "config") pod "ec6797b4-3f50-4496-97a9-eb84dd09d5c9" (UID: "ec6797b4-3f50-4496-97a9-eb84dd09d5c9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.246850 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-kube-api-access-mzzzd" (OuterVolumeSpecName: "kube-api-access-mzzzd") pod "ec6797b4-3f50-4496-97a9-eb84dd09d5c9" (UID: "ec6797b4-3f50-4496-97a9-eb84dd09d5c9"). InnerVolumeSpecName "kube-api-access-mzzzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.252488 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ec6797b4-3f50-4496-97a9-eb84dd09d5c9" (UID: "ec6797b4-3f50-4496-97a9-eb84dd09d5c9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.271337 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx"] Feb 02 12:12:30 crc kubenswrapper[4846]: E0202 12:12:30.271604 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec6797b4-3f50-4496-97a9-eb84dd09d5c9" containerName="route-controller-manager" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.271632 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec6797b4-3f50-4496-97a9-eb84dd09d5c9" containerName="route-controller-manager" Feb 02 12:12:30 crc kubenswrapper[4846]: E0202 12:12:30.271653 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08e80c6-fdfb-48cc-a752-9650f31f249e" containerName="pruner" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.271659 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08e80c6-fdfb-48cc-a752-9650f31f249e" containerName="pruner" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.271753 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e08e80c6-fdfb-48cc-a752-9650f31f249e" containerName="pruner" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.271764 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec6797b4-3f50-4496-97a9-eb84dd09d5c9" containerName="route-controller-manager" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.272184 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.280058 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx"] Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.339288 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-serving-cert\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.339372 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-client-ca\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.339416 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snphf\" (UniqueName: \"kubernetes.io/projected/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-kube-api-access-snphf\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.339465 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-config\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.339512 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.339527 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.339539 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzzzd\" (UniqueName: \"kubernetes.io/projected/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-kube-api-access-mzzzd\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.339550 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec6797b4-3f50-4496-97a9-eb84dd09d5c9-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.440722 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-client-ca\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.440792 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snphf\" (UniqueName: \"kubernetes.io/projected/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-kube-api-access-snphf\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.440844 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-config\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.440885 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-serving-cert\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.442134 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-config\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.442155 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-client-ca\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.444989 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-serving-cert\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.458260 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snphf\" (UniqueName: \"kubernetes.io/projected/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-kube-api-access-snphf\") pod \"route-controller-manager-6496f877d6-zhcmx\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.479250 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.479318 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:12:30 crc kubenswrapper[4846]: I0202 12:12:30.605750 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:31 crc kubenswrapper[4846]: I0202 12:12:31.003000 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" event={"ID":"ec6797b4-3f50-4496-97a9-eb84dd09d5c9","Type":"ContainerDied","Data":"f4a1b4c888e3cce8ee145c1f1fadcb05a16b3283d275e890e22600768817df26"} Feb 02 12:12:31 crc kubenswrapper[4846]: I0202 12:12:31.003055 4846 scope.go:117] "RemoveContainer" containerID="8f468181088598a609c761679bf1da6cc164283ecc0ec927cdb207c614833cec" Feb 02 12:12:31 crc kubenswrapper[4846]: I0202 12:12:31.003107 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g" Feb 02 12:12:31 crc kubenswrapper[4846]: I0202 12:12:31.033220 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g"] Feb 02 12:12:31 crc kubenswrapper[4846]: I0202 12:12:31.036114 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cfv6g"] Feb 02 12:12:31 crc kubenswrapper[4846]: I0202 12:12:31.431290 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec6797b4-3f50-4496-97a9-eb84dd09d5c9" path="/var/lib/kubelet/pods/ec6797b4-3f50-4496-97a9-eb84dd09d5c9/volumes" Feb 02 12:12:31 crc kubenswrapper[4846]: I0202 12:12:31.974658 4846 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-bbmc8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 12:12:31 crc kubenswrapper[4846]: I0202 12:12:31.974749 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" podUID="eba09ec1-c855-493b-8ab3-b40d4db28cc1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 12:12:35 crc kubenswrapper[4846]: I0202 12:12:35.845596 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:12:35 crc kubenswrapper[4846]: I0202 12:12:35.870429 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-94c47745d-pwfz4"] Feb 02 12:12:35 crc kubenswrapper[4846]: E0202 12:12:35.872582 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eba09ec1-c855-493b-8ab3-b40d4db28cc1" containerName="controller-manager" Feb 02 12:12:35 crc kubenswrapper[4846]: I0202 12:12:35.872601 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eba09ec1-c855-493b-8ab3-b40d4db28cc1" containerName="controller-manager" Feb 02 12:12:35 crc kubenswrapper[4846]: I0202 12:12:35.872719 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eba09ec1-c855-493b-8ab3-b40d4db28cc1" containerName="controller-manager" Feb 02 12:12:35 crc kubenswrapper[4846]: I0202 12:12:35.873149 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:35 crc kubenswrapper[4846]: I0202 12:12:35.875336 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-94c47745d-pwfz4"] Feb 02 12:12:35 crc kubenswrapper[4846]: E0202 12:12:35.883426 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 02 12:12:35 crc kubenswrapper[4846]: E0202 12:12:35.883558 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lhzjh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-n44pv_openshift-marketplace(64a47758-13e5-4a28-b781-2bdf84b870fd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 02 12:12:35 crc kubenswrapper[4846]: E0202 12:12:35.884786 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-n44pv" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.002886 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba09ec1-c855-493b-8ab3-b40d4db28cc1-serving-cert\") pod \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.002949 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p47g5\" (UniqueName: \"kubernetes.io/projected/eba09ec1-c855-493b-8ab3-b40d4db28cc1-kube-api-access-p47g5\") pod \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.002972 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-proxy-ca-bundles\") pod \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.003013 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-config\") pod \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.003043 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-client-ca\") pod \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\" (UID: \"eba09ec1-c855-493b-8ab3-b40d4db28cc1\") " Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.003214 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-proxy-ca-bundles\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.003266 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72b4c639-47cb-45de-b8dd-63ff63c407a2-serving-cert\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.003291 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-client-ca\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.003325 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-config\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.003340 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df8tv\" (UniqueName: \"kubernetes.io/projected/72b4c639-47cb-45de-b8dd-63ff63c407a2-kube-api-access-df8tv\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.004368 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-client-ca" (OuterVolumeSpecName: "client-ca") pod "eba09ec1-c855-493b-8ab3-b40d4db28cc1" (UID: "eba09ec1-c855-493b-8ab3-b40d4db28cc1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.004400 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "eba09ec1-c855-493b-8ab3-b40d4db28cc1" (UID: "eba09ec1-c855-493b-8ab3-b40d4db28cc1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.004417 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-config" (OuterVolumeSpecName: "config") pod "eba09ec1-c855-493b-8ab3-b40d4db28cc1" (UID: "eba09ec1-c855-493b-8ab3-b40d4db28cc1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.011674 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eba09ec1-c855-493b-8ab3-b40d4db28cc1-kube-api-access-p47g5" (OuterVolumeSpecName: "kube-api-access-p47g5") pod "eba09ec1-c855-493b-8ab3-b40d4db28cc1" (UID: "eba09ec1-c855-493b-8ab3-b40d4db28cc1"). InnerVolumeSpecName "kube-api-access-p47g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.018418 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eba09ec1-c855-493b-8ab3-b40d4db28cc1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "eba09ec1-c855-493b-8ab3-b40d4db28cc1" (UID: "eba09ec1-c855-493b-8ab3-b40d4db28cc1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.060692 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.060614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bbmc8" event={"ID":"eba09ec1-c855-493b-8ab3-b40d4db28cc1","Type":"ContainerDied","Data":"b15410055d3c6790086cbd87db8cd26d65a19c5677f1e0efdfca918b8560580d"} Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.060762 4846 scope.go:117] "RemoveContainer" containerID="35b11d93c537c1b9170c391cf4becfe22201a1cb6cdbabdf505aa014394503eb" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.081998 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-px4bj_8abc246a-7ca3-4550-ba36-491f830e5419/cluster-samples-operator/0.log" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.082181 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-px4bj" event={"ID":"8abc246a-7ca3-4550-ba36-491f830e5419","Type":"ContainerStarted","Data":"4276d43eaf46d3a844d03c0715b189700b654dad32a38d2153a3a13d838acfb6"} Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.101411 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bbmc8"] Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.104452 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72b4c639-47cb-45de-b8dd-63ff63c407a2-serving-cert\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.104528 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-client-ca\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.104557 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-config\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.104594 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df8tv\" (UniqueName: \"kubernetes.io/projected/72b4c639-47cb-45de-b8dd-63ff63c407a2-kube-api-access-df8tv\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.104654 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-proxy-ca-bundles\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.104747 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eba09ec1-c855-493b-8ab3-b40d4db28cc1-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.104758 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p47g5\" (UniqueName: \"kubernetes.io/projected/eba09ec1-c855-493b-8ab3-b40d4db28cc1-kube-api-access-p47g5\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.104768 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.104777 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.104805 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eba09ec1-c855-493b-8ab3-b40d4db28cc1-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.105682 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-client-ca\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.105771 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bbmc8"] Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.105977 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-config\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.107009 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-proxy-ca-bundles\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: E0202 12:12:36.109721 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-n44pv" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.116808 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72b4c639-47cb-45de-b8dd-63ff63c407a2-serving-cert\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.133424 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df8tv\" (UniqueName: \"kubernetes.io/projected/72b4c639-47cb-45de-b8dd-63ff63c407a2-kube-api-access-df8tv\") pod \"controller-manager-94c47745d-pwfz4\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.196525 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.543154 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx"] Feb 02 12:12:36 crc kubenswrapper[4846]: I0202 12:12:36.559083 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-94c47745d-pwfz4"] Feb 02 12:12:36 crc kubenswrapper[4846]: W0202 12:12:36.573602 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b7f32b7_c015_4ac5_9c8e_9e17230bc72c.slice/crio-a9e5d8ccfea9f64b3da915f1461d54073ee8f72a9b0d8111c8df81dfba594461 WatchSource:0}: Error finding container a9e5d8ccfea9f64b3da915f1461d54073ee8f72a9b0d8111c8df81dfba594461: Status 404 returned error can't find the container with id a9e5d8ccfea9f64b3da915f1461d54073ee8f72a9b0d8111c8df81dfba594461 Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.090963 4846 generic.go:334] "Generic (PLEG): container finished" podID="c8056015-9fef-462c-b839-b2171e1efa4b" containerID="86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2" exitCode=0 Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.091163 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksr5l" event={"ID":"c8056015-9fef-462c-b839-b2171e1efa4b","Type":"ContainerDied","Data":"86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.095492 4846 generic.go:334] "Generic (PLEG): container finished" podID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerID="72dca24d234e79bc7015f8a7c47a3a2eca66dd8e2535d263cb8d1adb482a4f8b" exitCode=0 Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.095718 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p558w" event={"ID":"14da97a9-4b68-458d-858a-6ba8b67cb749","Type":"ContainerDied","Data":"72dca24d234e79bc7015f8a7c47a3a2eca66dd8e2535d263cb8d1adb482a4f8b"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.099024 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" event={"ID":"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c","Type":"ContainerStarted","Data":"2ca54f025dca99bcf3c1f94a48657d7ad0afcca07f7a5703e8f0e5d3f8165a83"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.099059 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" event={"ID":"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c","Type":"ContainerStarted","Data":"a9e5d8ccfea9f64b3da915f1461d54073ee8f72a9b0d8111c8df81dfba594461"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.099241 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.104772 4846 generic.go:334] "Generic (PLEG): container finished" podID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerID="fb4e7a33b3bfc97dda3a67d3a4c7636d61632d03c7fe43802099e873bc821cdb" exitCode=0 Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.104885 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjvpk" event={"ID":"3b5b4c6f-368c-47ba-87fc-7309cac15d9b","Type":"ContainerDied","Data":"fb4e7a33b3bfc97dda3a67d3a4c7636d61632d03c7fe43802099e873bc821cdb"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.107273 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" event={"ID":"fe216e68-8814-468d-a6f3-fe8721d94eba","Type":"ContainerStarted","Data":"925e61857f7d1ed1a9e5a4df91e22f31ffbe4e2b40263a1655331168946784e8"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.107300 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-j5m9k" event={"ID":"fe216e68-8814-468d-a6f3-fe8721d94eba","Type":"ContainerStarted","Data":"ba5f68c9fd93db2f7d6455dd5f141d8d768fd80824160362b28b661d86234414"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.111044 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e81af364-0c77-4e24-8f1d-720046401479","Type":"ContainerStarted","Data":"5f254e69b15362313d66e929a6eebf7970052238b14180d06b77bf4456dcec1c"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.112616 4846 generic.go:334] "Generic (PLEG): container finished" podID="99680b5d-496f-4819-a517-1b1ead7e403d" containerID="ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc" exitCode=0 Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.112710 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn89z" event={"ID":"99680b5d-496f-4819-a517-1b1ead7e403d","Type":"ContainerDied","Data":"ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.115397 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" event={"ID":"72b4c639-47cb-45de-b8dd-63ff63c407a2","Type":"ContainerStarted","Data":"1c9f3e19987bd9363bdb8a2275fc67854b021c37174824bb9825faea62f9530d"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.115442 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" event={"ID":"72b4c639-47cb-45de-b8dd-63ff63c407a2","Type":"ContainerStarted","Data":"e7f4c3ff177c677b698c0e34e334b02b1bd65b17451e003ae14e76728b50088a"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.120028 4846 generic.go:334] "Generic (PLEG): container finished" podID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerID="56c2215aa72b34204e60b6dc026cd1ed208e3a787331105187d6bb32f67f38ad" exitCode=0 Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.120099 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmnp2" event={"ID":"98f32d89-a132-4cb9-a647-b48848e1b18a","Type":"ContainerDied","Data":"56c2215aa72b34204e60b6dc026cd1ed208e3a787331105187d6bb32f67f38ad"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.127821 4846 generic.go:334] "Generic (PLEG): container finished" podID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerID="43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e" exitCode=0 Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.127895 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpq4q" event={"ID":"1e90d3b0-496a-4f62-bdd7-5018fc11da0e","Type":"ContainerDied","Data":"43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.130264 4846 generic.go:334] "Generic (PLEG): container finished" podID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerID="abdae7a6154bcd7ea3e9755c1211bd93d963b7c46c179ec4ab8ced4048d5ebe7" exitCode=0 Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.130920 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2t4b" event={"ID":"04ce875b-335c-4dc9-a1d8-c2bd1f269675","Type":"ContainerDied","Data":"abdae7a6154bcd7ea3e9755c1211bd93d963b7c46c179ec4ab8ced4048d5ebe7"} Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.183586 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" podStartSLOduration=14.183565023 podStartE2EDuration="14.183565023s" podCreationTimestamp="2026-02-02 12:12:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:37.182602084 +0000 UTC m=+188.411188967" watchObservedRunningTime="2026-02-02 12:12:37.183565023 +0000 UTC m=+188.412151886" Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.208068 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=22.20805032 podStartE2EDuration="22.20805032s" podCreationTimestamp="2026-02-02 12:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:37.206125365 +0000 UTC m=+188.434712228" watchObservedRunningTime="2026-02-02 12:12:37.20805032 +0000 UTC m=+188.436637193" Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.225784 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-j5m9k" podStartSLOduration=164.225763882 podStartE2EDuration="2m44.225763882s" podCreationTimestamp="2026-02-02 12:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:37.222961401 +0000 UTC m=+188.451548274" watchObservedRunningTime="2026-02-02 12:12:37.225763882 +0000 UTC m=+188.454350745" Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.232277 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.271698 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" podStartSLOduration=14.2716819 podStartE2EDuration="14.2716819s" podCreationTimestamp="2026-02-02 12:12:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:37.270149886 +0000 UTC m=+188.498736759" watchObservedRunningTime="2026-02-02 12:12:37.2716819 +0000 UTC m=+188.500268763" Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.431948 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eba09ec1-c855-493b-8ab3-b40d4db28cc1" path="/var/lib/kubelet/pods/eba09ec1-c855-493b-8ab3-b40d4db28cc1/volumes" Feb 02 12:12:37 crc kubenswrapper[4846]: I0202 12:12:37.773538 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.141273 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn89z" event={"ID":"99680b5d-496f-4819-a517-1b1ead7e403d","Type":"ContainerStarted","Data":"6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883"} Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.145936 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p558w" event={"ID":"14da97a9-4b68-458d-858a-6ba8b67cb749","Type":"ContainerStarted","Data":"a174332179e0dd7320d061b3d003c07def4344aad74ef063a34205248ea67d06"} Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.149237 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjvpk" event={"ID":"3b5b4c6f-368c-47ba-87fc-7309cac15d9b","Type":"ContainerStarted","Data":"8e8b83ecde2490950dc76da4e5f92d61cce616b29e52fd40b0bee1afae697296"} Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.152063 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmnp2" event={"ID":"98f32d89-a132-4cb9-a647-b48848e1b18a","Type":"ContainerStarted","Data":"b49bfd2884c82785d674898ef7d77801330d101aaffeef0dc55a7dfd9c854c56"} Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.154340 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksr5l" event={"ID":"c8056015-9fef-462c-b839-b2171e1efa4b","Type":"ContainerStarted","Data":"ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93"} Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.157916 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpq4q" event={"ID":"1e90d3b0-496a-4f62-bdd7-5018fc11da0e","Type":"ContainerStarted","Data":"ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1"} Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.160497 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2t4b" event={"ID":"04ce875b-335c-4dc9-a1d8-c2bd1f269675","Type":"ContainerStarted","Data":"dc5e6070d5c73fef01be64b8e3d030f97b60a0d2af46a8581d9edeb2ef577f43"} Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.164435 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zn89z" podStartSLOduration=3.23951157 podStartE2EDuration="31.164417626s" podCreationTimestamp="2026-02-02 12:12:07 +0000 UTC" firstStartedPulling="2026-02-02 12:12:09.628132517 +0000 UTC m=+160.856719380" lastFinishedPulling="2026-02-02 12:12:37.553038573 +0000 UTC m=+188.781625436" observedRunningTime="2026-02-02 12:12:38.161663687 +0000 UTC m=+189.390250560" watchObservedRunningTime="2026-02-02 12:12:38.164417626 +0000 UTC m=+189.393004479" Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.166096 4846 generic.go:334] "Generic (PLEG): container finished" podID="e81af364-0c77-4e24-8f1d-720046401479" containerID="5f254e69b15362313d66e929a6eebf7970052238b14180d06b77bf4456dcec1c" exitCode=0 Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.166891 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e81af364-0c77-4e24-8f1d-720046401479","Type":"ContainerDied","Data":"5f254e69b15362313d66e929a6eebf7970052238b14180d06b77bf4456dcec1c"} Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.167664 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.172808 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.184799 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cpq4q" podStartSLOduration=3.344188311 podStartE2EDuration="29.184782755s" podCreationTimestamp="2026-02-02 12:12:09 +0000 UTC" firstStartedPulling="2026-02-02 12:12:11.724199489 +0000 UTC m=+162.952786342" lastFinishedPulling="2026-02-02 12:12:37.564793923 +0000 UTC m=+188.793380786" observedRunningTime="2026-02-02 12:12:38.183103237 +0000 UTC m=+189.411690110" watchObservedRunningTime="2026-02-02 12:12:38.184782755 +0000 UTC m=+189.413369618" Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.206665 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gmnp2" podStartSLOduration=3.164091009 podStartE2EDuration="31.206643637s" podCreationTimestamp="2026-02-02 12:12:07 +0000 UTC" firstStartedPulling="2026-02-02 12:12:09.585476483 +0000 UTC m=+160.814063346" lastFinishedPulling="2026-02-02 12:12:37.628029111 +0000 UTC m=+188.856615974" observedRunningTime="2026-02-02 12:12:38.204605728 +0000 UTC m=+189.433192591" watchObservedRunningTime="2026-02-02 12:12:38.206643637 +0000 UTC m=+189.435230500" Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.247351 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p558w" podStartSLOduration=2.242428982 podStartE2EDuration="29.247330454s" podCreationTimestamp="2026-02-02 12:12:09 +0000 UTC" firstStartedPulling="2026-02-02 12:12:10.673004011 +0000 UTC m=+161.901590874" lastFinishedPulling="2026-02-02 12:12:37.677905483 +0000 UTC m=+188.906492346" observedRunningTime="2026-02-02 12:12:38.227083378 +0000 UTC m=+189.455670241" watchObservedRunningTime="2026-02-02 12:12:38.247330454 +0000 UTC m=+189.475917317" Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.282058 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pjvpk" podStartSLOduration=3.109922666 podStartE2EDuration="32.282042226s" podCreationTimestamp="2026-02-02 12:12:06 +0000 UTC" firstStartedPulling="2026-02-02 12:12:08.571875283 +0000 UTC m=+159.800462146" lastFinishedPulling="2026-02-02 12:12:37.743994843 +0000 UTC m=+188.972581706" observedRunningTime="2026-02-02 12:12:38.249434584 +0000 UTC m=+189.478021447" watchObservedRunningTime="2026-02-02 12:12:38.282042226 +0000 UTC m=+189.510629089" Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.309233 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ksr5l" podStartSLOduration=3.376403147 podStartE2EDuration="31.309217512s" podCreationTimestamp="2026-02-02 12:12:07 +0000 UTC" firstStartedPulling="2026-02-02 12:12:09.589577032 +0000 UTC m=+160.818163885" lastFinishedPulling="2026-02-02 12:12:37.522391387 +0000 UTC m=+188.750978250" observedRunningTime="2026-02-02 12:12:38.285392284 +0000 UTC m=+189.513979147" watchObservedRunningTime="2026-02-02 12:12:38.309217512 +0000 UTC m=+189.537804375" Feb 02 12:12:38 crc kubenswrapper[4846]: I0202 12:12:38.310102 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g2t4b" podStartSLOduration=3.267993086 podStartE2EDuration="28.310096678s" podCreationTimestamp="2026-02-02 12:12:10 +0000 UTC" firstStartedPulling="2026-02-02 12:12:12.785758505 +0000 UTC m=+164.014345368" lastFinishedPulling="2026-02-02 12:12:37.827862097 +0000 UTC m=+189.056448960" observedRunningTime="2026-02-02 12:12:38.307634807 +0000 UTC m=+189.536221670" watchObservedRunningTime="2026-02-02 12:12:38.310096678 +0000 UTC m=+189.538683531" Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.407588 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.407888 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.508917 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.565056 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e81af364-0c77-4e24-8f1d-720046401479-kube-api-access\") pod \"e81af364-0c77-4e24-8f1d-720046401479\" (UID: \"e81af364-0c77-4e24-8f1d-720046401479\") " Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.565120 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e81af364-0c77-4e24-8f1d-720046401479-kubelet-dir\") pod \"e81af364-0c77-4e24-8f1d-720046401479\" (UID: \"e81af364-0c77-4e24-8f1d-720046401479\") " Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.565300 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e81af364-0c77-4e24-8f1d-720046401479-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e81af364-0c77-4e24-8f1d-720046401479" (UID: "e81af364-0c77-4e24-8f1d-720046401479"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.565527 4846 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e81af364-0c77-4e24-8f1d-720046401479-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.572462 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e81af364-0c77-4e24-8f1d-720046401479-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e81af364-0c77-4e24-8f1d-720046401479" (UID: "e81af364-0c77-4e24-8f1d-720046401479"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.667143 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e81af364-0c77-4e24-8f1d-720046401479-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.836896 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:39 crc kubenswrapper[4846]: I0202 12:12:39.836941 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:40 crc kubenswrapper[4846]: I0202 12:12:40.178744 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e81af364-0c77-4e24-8f1d-720046401479","Type":"ContainerDied","Data":"cd41135a797e39bf7a76839c137555ea462bafb2c2ddea57e207f724310fce87"} Feb 02 12:12:40 crc kubenswrapper[4846]: I0202 12:12:40.178801 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd41135a797e39bf7a76839c137555ea462bafb2c2ddea57e207f724310fce87" Feb 02 12:12:40 crc kubenswrapper[4846]: I0202 12:12:40.178951 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 12:12:40 crc kubenswrapper[4846]: I0202 12:12:40.594444 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-p558w" podUID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerName="registry-server" probeResult="failure" output=< Feb 02 12:12:40 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 12:12:40 crc kubenswrapper[4846]: > Feb 02 12:12:40 crc kubenswrapper[4846]: I0202 12:12:40.876824 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-cpq4q" podUID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerName="registry-server" probeResult="failure" output=< Feb 02 12:12:40 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 12:12:40 crc kubenswrapper[4846]: > Feb 02 12:12:40 crc kubenswrapper[4846]: I0202 12:12:40.881434 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:40 crc kubenswrapper[4846]: I0202 12:12:40.881948 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:41 crc kubenswrapper[4846]: I0202 12:12:41.080291 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fdjlt" Feb 02 12:12:41 crc kubenswrapper[4846]: I0202 12:12:41.921726 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g2t4b" podUID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerName="registry-server" probeResult="failure" output=< Feb 02 12:12:41 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 12:12:41 crc kubenswrapper[4846]: > Feb 02 12:12:43 crc kubenswrapper[4846]: I0202 12:12:43.265080 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-94c47745d-pwfz4"] Feb 02 12:12:43 crc kubenswrapper[4846]: I0202 12:12:43.265578 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" podUID="72b4c639-47cb-45de-b8dd-63ff63c407a2" containerName="controller-manager" containerID="cri-o://1c9f3e19987bd9363bdb8a2275fc67854b021c37174824bb9825faea62f9530d" gracePeriod=30 Feb 02 12:12:43 crc kubenswrapper[4846]: I0202 12:12:43.360598 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx"] Feb 02 12:12:43 crc kubenswrapper[4846]: I0202 12:12:43.360809 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" podUID="7b7f32b7-c015-4ac5-9c8e-9e17230bc72c" containerName="route-controller-manager" containerID="cri-o://2ca54f025dca99bcf3c1f94a48657d7ad0afcca07f7a5703e8f0e5d3f8165a83" gracePeriod=30 Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.227728 4846 generic.go:334] "Generic (PLEG): container finished" podID="72b4c639-47cb-45de-b8dd-63ff63c407a2" containerID="1c9f3e19987bd9363bdb8a2275fc67854b021c37174824bb9825faea62f9530d" exitCode=0 Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.228059 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" event={"ID":"72b4c639-47cb-45de-b8dd-63ff63c407a2","Type":"ContainerDied","Data":"1c9f3e19987bd9363bdb8a2275fc67854b021c37174824bb9825faea62f9530d"} Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.229526 4846 generic.go:334] "Generic (PLEG): container finished" podID="7b7f32b7-c015-4ac5-9c8e-9e17230bc72c" containerID="2ca54f025dca99bcf3c1f94a48657d7ad0afcca07f7a5703e8f0e5d3f8165a83" exitCode=0 Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.229556 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" event={"ID":"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c","Type":"ContainerDied","Data":"2ca54f025dca99bcf3c1f94a48657d7ad0afcca07f7a5703e8f0e5d3f8165a83"} Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.283848 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.297021 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.423343 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df8tv\" (UniqueName: \"kubernetes.io/projected/72b4c639-47cb-45de-b8dd-63ff63c407a2-kube-api-access-df8tv\") pod \"72b4c639-47cb-45de-b8dd-63ff63c407a2\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.423419 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-client-ca\") pod \"72b4c639-47cb-45de-b8dd-63ff63c407a2\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.423455 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-serving-cert\") pod \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.423495 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72b4c639-47cb-45de-b8dd-63ff63c407a2-serving-cert\") pod \"72b4c639-47cb-45de-b8dd-63ff63c407a2\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.423528 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-client-ca\") pod \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.423553 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-config\") pod \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.423576 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-config\") pod \"72b4c639-47cb-45de-b8dd-63ff63c407a2\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.423709 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snphf\" (UniqueName: \"kubernetes.io/projected/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-kube-api-access-snphf\") pod \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\" (UID: \"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c\") " Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.423742 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-proxy-ca-bundles\") pod \"72b4c639-47cb-45de-b8dd-63ff63c407a2\" (UID: \"72b4c639-47cb-45de-b8dd-63ff63c407a2\") " Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.424878 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-client-ca" (OuterVolumeSpecName: "client-ca") pod "72b4c639-47cb-45de-b8dd-63ff63c407a2" (UID: "72b4c639-47cb-45de-b8dd-63ff63c407a2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.424887 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "72b4c639-47cb-45de-b8dd-63ff63c407a2" (UID: "72b4c639-47cb-45de-b8dd-63ff63c407a2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.424920 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-config" (OuterVolumeSpecName: "config") pod "72b4c639-47cb-45de-b8dd-63ff63c407a2" (UID: "72b4c639-47cb-45de-b8dd-63ff63c407a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.424959 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-client-ca" (OuterVolumeSpecName: "client-ca") pod "7b7f32b7-c015-4ac5-9c8e-9e17230bc72c" (UID: "7b7f32b7-c015-4ac5-9c8e-9e17230bc72c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.426254 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-config" (OuterVolumeSpecName: "config") pod "7b7f32b7-c015-4ac5-9c8e-9e17230bc72c" (UID: "7b7f32b7-c015-4ac5-9c8e-9e17230bc72c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.432788 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7b7f32b7-c015-4ac5-9c8e-9e17230bc72c" (UID: "7b7f32b7-c015-4ac5-9c8e-9e17230bc72c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.432859 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-kube-api-access-snphf" (OuterVolumeSpecName: "kube-api-access-snphf") pod "7b7f32b7-c015-4ac5-9c8e-9e17230bc72c" (UID: "7b7f32b7-c015-4ac5-9c8e-9e17230bc72c"). InnerVolumeSpecName "kube-api-access-snphf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.438253 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw"] Feb 02 12:12:44 crc kubenswrapper[4846]: E0202 12:12:44.438525 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b4c639-47cb-45de-b8dd-63ff63c407a2" containerName="controller-manager" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.438540 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b4c639-47cb-45de-b8dd-63ff63c407a2" containerName="controller-manager" Feb 02 12:12:44 crc kubenswrapper[4846]: E0202 12:12:44.438554 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b7f32b7-c015-4ac5-9c8e-9e17230bc72c" containerName="route-controller-manager" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.438562 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b7f32b7-c015-4ac5-9c8e-9e17230bc72c" containerName="route-controller-manager" Feb 02 12:12:44 crc kubenswrapper[4846]: E0202 12:12:44.438574 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81af364-0c77-4e24-8f1d-720046401479" containerName="pruner" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.438581 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81af364-0c77-4e24-8f1d-720046401479" containerName="pruner" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.441043 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b7f32b7-c015-4ac5-9c8e-9e17230bc72c" containerName="route-controller-manager" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.441069 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="72b4c639-47cb-45de-b8dd-63ff63c407a2" containerName="controller-manager" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.441081 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e81af364-0c77-4e24-8f1d-720046401479" containerName="pruner" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.441790 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.444444 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72b4c639-47cb-45de-b8dd-63ff63c407a2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "72b4c639-47cb-45de-b8dd-63ff63c407a2" (UID: "72b4c639-47cb-45de-b8dd-63ff63c407a2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.444823 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw"] Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.461974 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72b4c639-47cb-45de-b8dd-63ff63c407a2-kube-api-access-df8tv" (OuterVolumeSpecName: "kube-api-access-df8tv") pod "72b4c639-47cb-45de-b8dd-63ff63c407a2" (UID: "72b4c639-47cb-45de-b8dd-63ff63c407a2"). InnerVolumeSpecName "kube-api-access-df8tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.524921 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc0cce5-a456-43d4-b515-ed033c3df5b3-serving-cert\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.524986 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-config\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525008 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd2hd\" (UniqueName: \"kubernetes.io/projected/8dc0cce5-a456-43d4-b515-ed033c3df5b3-kube-api-access-bd2hd\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525038 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-client-ca\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525101 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snphf\" (UniqueName: \"kubernetes.io/projected/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-kube-api-access-snphf\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525121 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525131 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df8tv\" (UniqueName: \"kubernetes.io/projected/72b4c639-47cb-45de-b8dd-63ff63c407a2-kube-api-access-df8tv\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525140 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525149 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525159 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72b4c639-47cb-45de-b8dd-63ff63c407a2-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525167 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525214 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72b4c639-47cb-45de-b8dd-63ff63c407a2-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.525225 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.626558 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc0cce5-a456-43d4-b515-ed033c3df5b3-serving-cert\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.626632 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-config\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.626656 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd2hd\" (UniqueName: \"kubernetes.io/projected/8dc0cce5-a456-43d4-b515-ed033c3df5b3-kube-api-access-bd2hd\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.626688 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-client-ca\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.627639 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-client-ca\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.627779 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-config\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.638736 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc0cce5-a456-43d4-b515-ed033c3df5b3-serving-cert\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.641735 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd2hd\" (UniqueName: \"kubernetes.io/projected/8dc0cce5-a456-43d4-b515-ed033c3df5b3-kube-api-access-bd2hd\") pod \"route-controller-manager-6f99846f7-c6hcw\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:44 crc kubenswrapper[4846]: I0202 12:12:44.799437 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.187883 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw"] Feb 02 12:12:45 crc kubenswrapper[4846]: W0202 12:12:45.189011 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dc0cce5_a456_43d4_b515_ed033c3df5b3.slice/crio-763fd47f849c7b65cc06d884c0d4739545d15b5f106cfb008c8df43f644a6887 WatchSource:0}: Error finding container 763fd47f849c7b65cc06d884c0d4739545d15b5f106cfb008c8df43f644a6887: Status 404 returned error can't find the container with id 763fd47f849c7b65cc06d884c0d4739545d15b5f106cfb008c8df43f644a6887 Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.237916 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" event={"ID":"8dc0cce5-a456-43d4-b515-ed033c3df5b3","Type":"ContainerStarted","Data":"763fd47f849c7b65cc06d884c0d4739545d15b5f106cfb008c8df43f644a6887"} Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.239577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" event={"ID":"72b4c639-47cb-45de-b8dd-63ff63c407a2","Type":"ContainerDied","Data":"e7f4c3ff177c677b698c0e34e334b02b1bd65b17451e003ae14e76728b50088a"} Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.239612 4846 scope.go:117] "RemoveContainer" containerID="1c9f3e19987bd9363bdb8a2275fc67854b021c37174824bb9825faea62f9530d" Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.239747 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-94c47745d-pwfz4" Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.242181 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" event={"ID":"7b7f32b7-c015-4ac5-9c8e-9e17230bc72c","Type":"ContainerDied","Data":"a9e5d8ccfea9f64b3da915f1461d54073ee8f72a9b0d8111c8df81dfba594461"} Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.242265 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx" Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.257355 4846 scope.go:117] "RemoveContainer" containerID="2ca54f025dca99bcf3c1f94a48657d7ad0afcca07f7a5703e8f0e5d3f8165a83" Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.275992 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-94c47745d-pwfz4"] Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.280019 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-94c47745d-pwfz4"] Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.287920 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx"] Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.294731 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6496f877d6-zhcmx"] Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.431648 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72b4c639-47cb-45de-b8dd-63ff63c407a2" path="/var/lib/kubelet/pods/72b4c639-47cb-45de-b8dd-63ff63c407a2/volumes" Feb 02 12:12:45 crc kubenswrapper[4846]: I0202 12:12:45.432320 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b7f32b7-c015-4ac5-9c8e-9e17230bc72c" path="/var/lib/kubelet/pods/7b7f32b7-c015-4ac5-9c8e-9e17230bc72c/volumes" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.248269 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" event={"ID":"8dc0cce5-a456-43d4-b515-ed033c3df5b3","Type":"ContainerStarted","Data":"85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830"} Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.248546 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.254441 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.264868 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" podStartSLOduration=3.264851644 podStartE2EDuration="3.264851644s" podCreationTimestamp="2026-02-02 12:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:46.262889 +0000 UTC m=+197.491475863" watchObservedRunningTime="2026-02-02 12:12:46.264851644 +0000 UTC m=+197.493438507" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.435029 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-697f8fb498-p854p"] Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.438655 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.444657 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.444671 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.444712 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.444815 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.444907 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.445328 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.445692 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-697f8fb498-p854p"] Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.448676 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.455941 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-config\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.455992 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxscd\" (UniqueName: \"kubernetes.io/projected/709d2acc-338f-4f64-8b22-d81215b650cf-kube-api-access-xxscd\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.456103 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-client-ca\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.456220 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/709d2acc-338f-4f64-8b22-d81215b650cf-serving-cert\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.456277 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-proxy-ca-bundles\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.557410 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/709d2acc-338f-4f64-8b22-d81215b650cf-serving-cert\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.557468 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-proxy-ca-bundles\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.557668 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-config\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.557698 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxscd\" (UniqueName: \"kubernetes.io/projected/709d2acc-338f-4f64-8b22-d81215b650cf-kube-api-access-xxscd\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.557747 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-client-ca\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.558491 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-client-ca\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.559747 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-config\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.559935 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-proxy-ca-bundles\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.573529 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/709d2acc-338f-4f64-8b22-d81215b650cf-serving-cert\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.577057 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxscd\" (UniqueName: \"kubernetes.io/projected/709d2acc-338f-4f64-8b22-d81215b650cf-kube-api-access-xxscd\") pod \"controller-manager-697f8fb498-p854p\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:46 crc kubenswrapper[4846]: I0202 12:12:46.760440 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.153610 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-697f8fb498-p854p"] Feb 02 12:12:47 crc kubenswrapper[4846]: W0202 12:12:47.158386 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod709d2acc_338f_4f64_8b22_d81215b650cf.slice/crio-af13caf60f921fbeedf0afcd4b9f6f7f42e82de5ece930f5b372bd0a3f46d139 WatchSource:0}: Error finding container af13caf60f921fbeedf0afcd4b9f6f7f42e82de5ece930f5b372bd0a3f46d139: Status 404 returned error can't find the container with id af13caf60f921fbeedf0afcd4b9f6f7f42e82de5ece930f5b372bd0a3f46d139 Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.208261 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.208301 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.254033 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.257259 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" event={"ID":"709d2acc-338f-4f64-8b22-d81215b650cf","Type":"ContainerStarted","Data":"af13caf60f921fbeedf0afcd4b9f6f7f42e82de5ece930f5b372bd0a3f46d139"} Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.330556 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.639733 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.639802 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.688008 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.760774 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.760824 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.807832 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.834454 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.834489 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:47 crc kubenswrapper[4846]: I0202 12:12:47.875709 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:48 crc kubenswrapper[4846]: I0202 12:12:48.263301 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" event={"ID":"709d2acc-338f-4f64-8b22-d81215b650cf","Type":"ContainerStarted","Data":"723701191e8b206603814678dbc3013de9cfd5efa4736b5daff6a257396f50af"} Feb 02 12:12:48 crc kubenswrapper[4846]: I0202 12:12:48.264466 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:48 crc kubenswrapper[4846]: I0202 12:12:48.271607 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:12:48 crc kubenswrapper[4846]: I0202 12:12:48.307362 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" podStartSLOduration=5.307342567 podStartE2EDuration="5.307342567s" podCreationTimestamp="2026-02-02 12:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:12:48.28613929 +0000 UTC m=+199.514726153" watchObservedRunningTime="2026-02-02 12:12:48.307342567 +0000 UTC m=+199.535929430" Feb 02 12:12:48 crc kubenswrapper[4846]: I0202 12:12:48.307716 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:48 crc kubenswrapper[4846]: I0202 12:12:48.309508 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:12:48 crc kubenswrapper[4846]: I0202 12:12:48.320854 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:49 crc kubenswrapper[4846]: I0202 12:12:49.082970 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zn89z"] Feb 02 12:12:49 crc kubenswrapper[4846]: I0202 12:12:49.455990 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:49 crc kubenswrapper[4846]: I0202 12:12:49.494483 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:12:49 crc kubenswrapper[4846]: I0202 12:12:49.877235 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:49 crc kubenswrapper[4846]: I0202 12:12:49.923434 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.090271 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ksr5l"] Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.273596 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ksr5l" podUID="c8056015-9fef-462c-b839-b2171e1efa4b" containerName="registry-server" containerID="cri-o://ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93" gracePeriod=2 Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.274331 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zn89z" podUID="99680b5d-496f-4819-a517-1b1ead7e403d" containerName="registry-server" containerID="cri-o://6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883" gracePeriod=2 Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.716495 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.721057 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.820119 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj4s2\" (UniqueName: \"kubernetes.io/projected/c8056015-9fef-462c-b839-b2171e1efa4b-kube-api-access-dj4s2\") pod \"c8056015-9fef-462c-b839-b2171e1efa4b\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.820652 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72mv4\" (UniqueName: \"kubernetes.io/projected/99680b5d-496f-4819-a517-1b1ead7e403d-kube-api-access-72mv4\") pod \"99680b5d-496f-4819-a517-1b1ead7e403d\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.820707 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-utilities\") pod \"c8056015-9fef-462c-b839-b2171e1efa4b\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.820746 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-utilities\") pod \"99680b5d-496f-4819-a517-1b1ead7e403d\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.820816 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-catalog-content\") pod \"99680b5d-496f-4819-a517-1b1ead7e403d\" (UID: \"99680b5d-496f-4819-a517-1b1ead7e403d\") " Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.820893 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-catalog-content\") pod \"c8056015-9fef-462c-b839-b2171e1efa4b\" (UID: \"c8056015-9fef-462c-b839-b2171e1efa4b\") " Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.822220 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-utilities" (OuterVolumeSpecName: "utilities") pod "99680b5d-496f-4819-a517-1b1ead7e403d" (UID: "99680b5d-496f-4819-a517-1b1ead7e403d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.822989 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-utilities" (OuterVolumeSpecName: "utilities") pod "c8056015-9fef-462c-b839-b2171e1efa4b" (UID: "c8056015-9fef-462c-b839-b2171e1efa4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.830100 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99680b5d-496f-4819-a517-1b1ead7e403d-kube-api-access-72mv4" (OuterVolumeSpecName: "kube-api-access-72mv4") pod "99680b5d-496f-4819-a517-1b1ead7e403d" (UID: "99680b5d-496f-4819-a517-1b1ead7e403d"). InnerVolumeSpecName "kube-api-access-72mv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.830896 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8056015-9fef-462c-b839-b2171e1efa4b-kube-api-access-dj4s2" (OuterVolumeSpecName: "kube-api-access-dj4s2") pod "c8056015-9fef-462c-b839-b2171e1efa4b" (UID: "c8056015-9fef-462c-b839-b2171e1efa4b"). InnerVolumeSpecName "kube-api-access-dj4s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.875863 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99680b5d-496f-4819-a517-1b1ead7e403d" (UID: "99680b5d-496f-4819-a517-1b1ead7e403d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.883239 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8056015-9fef-462c-b839-b2171e1efa4b" (UID: "c8056015-9fef-462c-b839-b2171e1efa4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.918715 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.922060 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj4s2\" (UniqueName: \"kubernetes.io/projected/c8056015-9fef-462c-b839-b2171e1efa4b-kube-api-access-dj4s2\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.922087 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72mv4\" (UniqueName: \"kubernetes.io/projected/99680b5d-496f-4819-a517-1b1ead7e403d-kube-api-access-72mv4\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.922099 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.922109 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.922169 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99680b5d-496f-4819-a517-1b1ead7e403d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.922181 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8056015-9fef-462c-b839-b2171e1efa4b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:50 crc kubenswrapper[4846]: I0202 12:12:50.970311 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.284080 4846 generic.go:334] "Generic (PLEG): container finished" podID="c8056015-9fef-462c-b839-b2171e1efa4b" containerID="ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93" exitCode=0 Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.284165 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksr5l" event={"ID":"c8056015-9fef-462c-b839-b2171e1efa4b","Type":"ContainerDied","Data":"ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93"} Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.284198 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksr5l" event={"ID":"c8056015-9fef-462c-b839-b2171e1efa4b","Type":"ContainerDied","Data":"97c1292ca4a6a65a75204ea33784dda2642129af7f5eb5124ce46bd2e74eb19d"} Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.284222 4846 scope.go:117] "RemoveContainer" containerID="ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.284374 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksr5l" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.287932 4846 generic.go:334] "Generic (PLEG): container finished" podID="99680b5d-496f-4819-a517-1b1ead7e403d" containerID="6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883" exitCode=0 Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.287995 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zn89z" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.288012 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn89z" event={"ID":"99680b5d-496f-4819-a517-1b1ead7e403d","Type":"ContainerDied","Data":"6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883"} Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.288084 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn89z" event={"ID":"99680b5d-496f-4819-a517-1b1ead7e403d","Type":"ContainerDied","Data":"4be44f2d733444f8872a0cbc5df0e2420d1bc1fecad3b70850ad17c93638f102"} Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.302473 4846 scope.go:117] "RemoveContainer" containerID="86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.318562 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ksr5l"] Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.326472 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ksr5l"] Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.330471 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zn89z"] Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.333539 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zn89z"] Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.336455 4846 scope.go:117] "RemoveContainer" containerID="15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.348963 4846 scope.go:117] "RemoveContainer" containerID="ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93" Feb 02 12:12:51 crc kubenswrapper[4846]: E0202 12:12:51.349342 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93\": container with ID starting with ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93 not found: ID does not exist" containerID="ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.349375 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93"} err="failed to get container status \"ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93\": rpc error: code = NotFound desc = could not find container \"ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93\": container with ID starting with ec957571c8bf0bf05d9aacb462ab2032a32be1869994b15ab57def3f810f8e93 not found: ID does not exist" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.349415 4846 scope.go:117] "RemoveContainer" containerID="86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2" Feb 02 12:12:51 crc kubenswrapper[4846]: E0202 12:12:51.349709 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2\": container with ID starting with 86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2 not found: ID does not exist" containerID="86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.349727 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2"} err="failed to get container status \"86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2\": rpc error: code = NotFound desc = could not find container \"86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2\": container with ID starting with 86b1d477175bb4f6af1278fe3baace409c9cd76aecf7fc29f2cae9bbdac1e6b2 not found: ID does not exist" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.349745 4846 scope.go:117] "RemoveContainer" containerID="15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430" Feb 02 12:12:51 crc kubenswrapper[4846]: E0202 12:12:51.349977 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430\": container with ID starting with 15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430 not found: ID does not exist" containerID="15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.349995 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430"} err="failed to get container status \"15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430\": rpc error: code = NotFound desc = could not find container \"15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430\": container with ID starting with 15eb235d83f55454fb9062b8e84c07ba9445794f0473538a969edabf5787d430 not found: ID does not exist" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.350007 4846 scope.go:117] "RemoveContainer" containerID="6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.362318 4846 scope.go:117] "RemoveContainer" containerID="ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.375658 4846 scope.go:117] "RemoveContainer" containerID="a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.390015 4846 scope.go:117] "RemoveContainer" containerID="6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883" Feb 02 12:12:51 crc kubenswrapper[4846]: E0202 12:12:51.390375 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883\": container with ID starting with 6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883 not found: ID does not exist" containerID="6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.390413 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883"} err="failed to get container status \"6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883\": rpc error: code = NotFound desc = could not find container \"6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883\": container with ID starting with 6cc21b75e23cc3867564cd7351a21ebf1bc298a3d43d0f13001efc165bec4883 not found: ID does not exist" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.390441 4846 scope.go:117] "RemoveContainer" containerID="ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc" Feb 02 12:12:51 crc kubenswrapper[4846]: E0202 12:12:51.390828 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc\": container with ID starting with ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc not found: ID does not exist" containerID="ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.390898 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc"} err="failed to get container status \"ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc\": rpc error: code = NotFound desc = could not find container \"ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc\": container with ID starting with ae449f70df9c1d21ffc58902a1796e5e1ff6b77d347ee25b37483fef24b27fdc not found: ID does not exist" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.390950 4846 scope.go:117] "RemoveContainer" containerID="a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2" Feb 02 12:12:51 crc kubenswrapper[4846]: E0202 12:12:51.391268 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2\": container with ID starting with a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2 not found: ID does not exist" containerID="a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.391298 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2"} err="failed to get container status \"a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2\": rpc error: code = NotFound desc = could not find container \"a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2\": container with ID starting with a173adc06e62fe9ec55ec09a6ff9c01def59345ccf48bb9d77040d234c5411f2 not found: ID does not exist" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.432845 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99680b5d-496f-4819-a517-1b1ead7e403d" path="/var/lib/kubelet/pods/99680b5d-496f-4819-a517-1b1ead7e403d/volumes" Feb 02 12:12:51 crc kubenswrapper[4846]: I0202 12:12:51.433587 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8056015-9fef-462c-b839-b2171e1efa4b" path="/var/lib/kubelet/pods/c8056015-9fef-462c-b839-b2171e1efa4b/volumes" Feb 02 12:12:52 crc kubenswrapper[4846]: I0202 12:12:52.484518 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cpq4q"] Feb 02 12:12:52 crc kubenswrapper[4846]: I0202 12:12:52.485027 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cpq4q" podUID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerName="registry-server" containerID="cri-o://ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1" gracePeriod=2 Feb 02 12:12:52 crc kubenswrapper[4846]: I0202 12:12:52.849939 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:52 crc kubenswrapper[4846]: I0202 12:12:52.947523 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-catalog-content\") pod \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " Feb 02 12:12:52 crc kubenswrapper[4846]: I0202 12:12:52.947632 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-utilities\") pod \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " Feb 02 12:12:52 crc kubenswrapper[4846]: I0202 12:12:52.947691 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jf5l\" (UniqueName: \"kubernetes.io/projected/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-kube-api-access-8jf5l\") pod \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\" (UID: \"1e90d3b0-496a-4f62-bdd7-5018fc11da0e\") " Feb 02 12:12:52 crc kubenswrapper[4846]: I0202 12:12:52.948682 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-utilities" (OuterVolumeSpecName: "utilities") pod "1e90d3b0-496a-4f62-bdd7-5018fc11da0e" (UID: "1e90d3b0-496a-4f62-bdd7-5018fc11da0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:12:52 crc kubenswrapper[4846]: I0202 12:12:52.953083 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-kube-api-access-8jf5l" (OuterVolumeSpecName: "kube-api-access-8jf5l") pod "1e90d3b0-496a-4f62-bdd7-5018fc11da0e" (UID: "1e90d3b0-496a-4f62-bdd7-5018fc11da0e"). InnerVolumeSpecName "kube-api-access-8jf5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:52 crc kubenswrapper[4846]: I0202 12:12:52.990406 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e90d3b0-496a-4f62-bdd7-5018fc11da0e" (UID: "1e90d3b0-496a-4f62-bdd7-5018fc11da0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.049566 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.049613 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.049641 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jf5l\" (UniqueName: \"kubernetes.io/projected/1e90d3b0-496a-4f62-bdd7-5018fc11da0e-kube-api-access-8jf5l\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.300813 4846 generic.go:334] "Generic (PLEG): container finished" podID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerID="e4ea05ba5acc3ccfe68a216c877a5b7c4aa682276bbc4a4999ebd856da4b3e89" exitCode=0 Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.300927 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n44pv" event={"ID":"64a47758-13e5-4a28-b781-2bdf84b870fd","Type":"ContainerDied","Data":"e4ea05ba5acc3ccfe68a216c877a5b7c4aa682276bbc4a4999ebd856da4b3e89"} Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.303164 4846 generic.go:334] "Generic (PLEG): container finished" podID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerID="ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1" exitCode=0 Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.303184 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpq4q" event={"ID":"1e90d3b0-496a-4f62-bdd7-5018fc11da0e","Type":"ContainerDied","Data":"ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1"} Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.303202 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cpq4q" event={"ID":"1e90d3b0-496a-4f62-bdd7-5018fc11da0e","Type":"ContainerDied","Data":"21c50224d365a94bab1a9ea1955a60298626dafc606a716692b0910086e4e30f"} Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.303218 4846 scope.go:117] "RemoveContainer" containerID="ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.303266 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cpq4q" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.317121 4846 scope.go:117] "RemoveContainer" containerID="43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.336048 4846 scope.go:117] "RemoveContainer" containerID="83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.344041 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cpq4q"] Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.350787 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cpq4q"] Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.351803 4846 scope.go:117] "RemoveContainer" containerID="ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.352210 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1\": container with ID starting with ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1 not found: ID does not exist" containerID="ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.352253 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1"} err="failed to get container status \"ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1\": rpc error: code = NotFound desc = could not find container \"ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1\": container with ID starting with ca21c5a0afef30d0407d54b7d07813e47244b96415483e469da3c3e7c229cad1 not found: ID does not exist" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.352284 4846 scope.go:117] "RemoveContainer" containerID="43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.352741 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e\": container with ID starting with 43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e not found: ID does not exist" containerID="43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.352785 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e"} err="failed to get container status \"43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e\": rpc error: code = NotFound desc = could not find container \"43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e\": container with ID starting with 43675803ff92a61bfbc867315a16d8e91101bbc4d5e2ae1d9d252d5b8deb794e not found: ID does not exist" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.352812 4846 scope.go:117] "RemoveContainer" containerID="83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.353100 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b\": container with ID starting with 83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b not found: ID does not exist" containerID="83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.353139 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b"} err="failed to get container status \"83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b\": rpc error: code = NotFound desc = could not find container \"83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b\": container with ID starting with 83e204fce054974f7a9fffc7bb205ab12b593da8ba44aab7f08f551dee00df6b not found: ID does not exist" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.422557 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.422833 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerName="extract-utilities" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.422850 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerName="extract-utilities" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.422872 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8056015-9fef-462c-b839-b2171e1efa4b" containerName="registry-server" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.422880 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8056015-9fef-462c-b839-b2171e1efa4b" containerName="registry-server" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.422892 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99680b5d-496f-4819-a517-1b1ead7e403d" containerName="registry-server" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.422899 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="99680b5d-496f-4819-a517-1b1ead7e403d" containerName="registry-server" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.422911 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerName="extract-content" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.422919 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerName="extract-content" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.422934 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99680b5d-496f-4819-a517-1b1ead7e403d" containerName="extract-content" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.422943 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="99680b5d-496f-4819-a517-1b1ead7e403d" containerName="extract-content" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.422952 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8056015-9fef-462c-b839-b2171e1efa4b" containerName="extract-utilities" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.422959 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8056015-9fef-462c-b839-b2171e1efa4b" containerName="extract-utilities" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.422970 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99680b5d-496f-4819-a517-1b1ead7e403d" containerName="extract-utilities" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.422977 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="99680b5d-496f-4819-a517-1b1ead7e403d" containerName="extract-utilities" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.422986 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerName="registry-server" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.422994 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerName="registry-server" Feb 02 12:12:53 crc kubenswrapper[4846]: E0202 12:12:53.423007 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8056015-9fef-462c-b839-b2171e1efa4b" containerName="extract-content" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.423015 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8056015-9fef-462c-b839-b2171e1efa4b" containerName="extract-content" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.423130 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8056015-9fef-462c-b839-b2171e1efa4b" containerName="registry-server" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.423148 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" containerName="registry-server" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.423161 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="99680b5d-496f-4819-a517-1b1ead7e403d" containerName="registry-server" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.423710 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.426171 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.426371 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.444153 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e90d3b0-496a-4f62-bdd7-5018fc11da0e" path="/var/lib/kubelet/pods/1e90d3b0-496a-4f62-bdd7-5018fc11da0e/volumes" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.444704 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.453645 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ac34faf-2f02-4190-bacf-cc8164f121bc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3ac34faf-2f02-4190-bacf-cc8164f121bc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.453711 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ac34faf-2f02-4190-bacf-cc8164f121bc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3ac34faf-2f02-4190-bacf-cc8164f121bc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.555408 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ac34faf-2f02-4190-bacf-cc8164f121bc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3ac34faf-2f02-4190-bacf-cc8164f121bc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.555469 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ac34faf-2f02-4190-bacf-cc8164f121bc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3ac34faf-2f02-4190-bacf-cc8164f121bc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.555519 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ac34faf-2f02-4190-bacf-cc8164f121bc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3ac34faf-2f02-4190-bacf-cc8164f121bc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.577962 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ac34faf-2f02-4190-bacf-cc8164f121bc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3ac34faf-2f02-4190-bacf-cc8164f121bc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 12:12:53 crc kubenswrapper[4846]: I0202 12:12:53.760149 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 12:12:54 crc kubenswrapper[4846]: I0202 12:12:54.159755 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 02 12:12:54 crc kubenswrapper[4846]: I0202 12:12:54.309550 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n44pv" event={"ID":"64a47758-13e5-4a28-b781-2bdf84b870fd","Type":"ContainerStarted","Data":"ba291467c7fb9876c4109dfe1171cb8179f89c87fdc1e8aa3cbe5780f0dddf04"} Feb 02 12:12:54 crc kubenswrapper[4846]: I0202 12:12:54.312814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3ac34faf-2f02-4190-bacf-cc8164f121bc","Type":"ContainerStarted","Data":"b8c63af7add3e4f2057825cb4b5e4c1276672a4afcadbe52630168478b279dbc"} Feb 02 12:12:54 crc kubenswrapper[4846]: I0202 12:12:54.337824 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n44pv" podStartSLOduration=2.379659749 podStartE2EDuration="44.337779342s" podCreationTimestamp="2026-02-02 12:12:10 +0000 UTC" firstStartedPulling="2026-02-02 12:12:11.724378004 +0000 UTC m=+162.952964857" lastFinishedPulling="2026-02-02 12:12:53.682497597 +0000 UTC m=+204.911084450" observedRunningTime="2026-02-02 12:12:54.333826704 +0000 UTC m=+205.562413567" watchObservedRunningTime="2026-02-02 12:12:54.337779342 +0000 UTC m=+205.566366205" Feb 02 12:12:54 crc kubenswrapper[4846]: I0202 12:12:54.882493 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g2t4b"] Feb 02 12:12:54 crc kubenswrapper[4846]: I0202 12:12:54.882942 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g2t4b" podUID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerName="registry-server" containerID="cri-o://dc5e6070d5c73fef01be64b8e3d030f97b60a0d2af46a8581d9edeb2ef577f43" gracePeriod=2 Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.318804 4846 generic.go:334] "Generic (PLEG): container finished" podID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerID="dc5e6070d5c73fef01be64b8e3d030f97b60a0d2af46a8581d9edeb2ef577f43" exitCode=0 Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.318871 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2t4b" event={"ID":"04ce875b-335c-4dc9-a1d8-c2bd1f269675","Type":"ContainerDied","Data":"dc5e6070d5c73fef01be64b8e3d030f97b60a0d2af46a8581d9edeb2ef577f43"} Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.320119 4846 generic.go:334] "Generic (PLEG): container finished" podID="3ac34faf-2f02-4190-bacf-cc8164f121bc" containerID="cbf47d242582c616cbc65b3dedab29f7976fcbb4fd37729f942229775e01cd5a" exitCode=0 Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.320145 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3ac34faf-2f02-4190-bacf-cc8164f121bc","Type":"ContainerDied","Data":"cbf47d242582c616cbc65b3dedab29f7976fcbb4fd37729f942229775e01cd5a"} Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.441169 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.507135 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-utilities\") pod \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.507208 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5skb\" (UniqueName: \"kubernetes.io/projected/04ce875b-335c-4dc9-a1d8-c2bd1f269675-kube-api-access-k5skb\") pod \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.507283 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-catalog-content\") pod \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\" (UID: \"04ce875b-335c-4dc9-a1d8-c2bd1f269675\") " Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.507912 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-utilities" (OuterVolumeSpecName: "utilities") pod "04ce875b-335c-4dc9-a1d8-c2bd1f269675" (UID: "04ce875b-335c-4dc9-a1d8-c2bd1f269675"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.512722 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04ce875b-335c-4dc9-a1d8-c2bd1f269675-kube-api-access-k5skb" (OuterVolumeSpecName: "kube-api-access-k5skb") pod "04ce875b-335c-4dc9-a1d8-c2bd1f269675" (UID: "04ce875b-335c-4dc9-a1d8-c2bd1f269675"). InnerVolumeSpecName "kube-api-access-k5skb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.608259 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.608301 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5skb\" (UniqueName: \"kubernetes.io/projected/04ce875b-335c-4dc9-a1d8-c2bd1f269675-kube-api-access-k5skb\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.645673 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04ce875b-335c-4dc9-a1d8-c2bd1f269675" (UID: "04ce875b-335c-4dc9-a1d8-c2bd1f269675"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:12:55 crc kubenswrapper[4846]: I0202 12:12:55.708755 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04ce875b-335c-4dc9-a1d8-c2bd1f269675-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.325884 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g2t4b" Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.325852 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g2t4b" event={"ID":"04ce875b-335c-4dc9-a1d8-c2bd1f269675","Type":"ContainerDied","Data":"82fe4b1ee571afcfbf27877e4dad476365e5c20d346e397e02ee74fe024ac827"} Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.326693 4846 scope.go:117] "RemoveContainer" containerID="dc5e6070d5c73fef01be64b8e3d030f97b60a0d2af46a8581d9edeb2ef577f43" Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.346223 4846 scope.go:117] "RemoveContainer" containerID="abdae7a6154bcd7ea3e9755c1211bd93d963b7c46c179ec4ab8ced4048d5ebe7" Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.364195 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g2t4b"] Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.367306 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g2t4b"] Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.405228 4846 scope.go:117] "RemoveContainer" containerID="c3a90734817c2f0baa3cbb295141d9b51ab8cbd5d79c74857a530dd4989a1b68" Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.679552 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.823232 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ac34faf-2f02-4190-bacf-cc8164f121bc-kube-api-access\") pod \"3ac34faf-2f02-4190-bacf-cc8164f121bc\" (UID: \"3ac34faf-2f02-4190-bacf-cc8164f121bc\") " Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.823357 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ac34faf-2f02-4190-bacf-cc8164f121bc-kubelet-dir\") pod \"3ac34faf-2f02-4190-bacf-cc8164f121bc\" (UID: \"3ac34faf-2f02-4190-bacf-cc8164f121bc\") " Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.823491 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3ac34faf-2f02-4190-bacf-cc8164f121bc-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3ac34faf-2f02-4190-bacf-cc8164f121bc" (UID: "3ac34faf-2f02-4190-bacf-cc8164f121bc"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.823927 4846 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ac34faf-2f02-4190-bacf-cc8164f121bc-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.832414 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ac34faf-2f02-4190-bacf-cc8164f121bc-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3ac34faf-2f02-4190-bacf-cc8164f121bc" (UID: "3ac34faf-2f02-4190-bacf-cc8164f121bc"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:12:56 crc kubenswrapper[4846]: I0202 12:12:56.924667 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ac34faf-2f02-4190-bacf-cc8164f121bc-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 12:12:57 crc kubenswrapper[4846]: I0202 12:12:57.333054 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"3ac34faf-2f02-4190-bacf-cc8164f121bc","Type":"ContainerDied","Data":"b8c63af7add3e4f2057825cb4b5e4c1276672a4afcadbe52630168478b279dbc"} Feb 02 12:12:57 crc kubenswrapper[4846]: I0202 12:12:57.334330 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8c63af7add3e4f2057825cb4b5e4c1276672a4afcadbe52630168478b279dbc" Feb 02 12:12:57 crc kubenswrapper[4846]: I0202 12:12:57.333373 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 12:12:57 crc kubenswrapper[4846]: I0202 12:12:57.431500 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" path="/var/lib/kubelet/pods/04ce875b-335c-4dc9-a1d8-c2bd1f269675/volumes" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.022193 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 02 12:13:00 crc kubenswrapper[4846]: E0202 12:13:00.022867 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerName="extract-utilities" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.022890 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerName="extract-utilities" Feb 02 12:13:00 crc kubenswrapper[4846]: E0202 12:13:00.022904 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac34faf-2f02-4190-bacf-cc8164f121bc" containerName="pruner" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.022915 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac34faf-2f02-4190-bacf-cc8164f121bc" containerName="pruner" Feb 02 12:13:00 crc kubenswrapper[4846]: E0202 12:13:00.022935 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerName="extract-content" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.022952 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerName="extract-content" Feb 02 12:13:00 crc kubenswrapper[4846]: E0202 12:13:00.022967 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerName="registry-server" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.022978 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerName="registry-server" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.023168 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="04ce875b-335c-4dc9-a1d8-c2bd1f269675" containerName="registry-server" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.023188 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ac34faf-2f02-4190-bacf-cc8164f121bc" containerName="pruner" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.024007 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.026295 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.028337 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.032232 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.062969 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kube-api-access\") pod \"installer-9-crc\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.063044 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.063113 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-var-lock\") pod \"installer-9-crc\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.164255 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kube-api-access\") pod \"installer-9-crc\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.164300 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.164333 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-var-lock\") pod \"installer-9-crc\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.164383 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.164420 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-var-lock\") pod \"installer-9-crc\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.181526 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kube-api-access\") pod \"installer-9-crc\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.347353 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.392184 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.392243 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.448890 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.479275 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.479324 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.479368 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.479976 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.480033 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f" gracePeriod=600 Feb 02 12:13:00 crc kubenswrapper[4846]: I0202 12:13:00.725026 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 02 12:13:00 crc kubenswrapper[4846]: W0202 12:13:00.731324 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd2a34ae2_14ef_4ca9_b3be_67e43673d5f9.slice/crio-94c7286786e9088871506cf9661003f0bf0ba1b8f6874cbf75743e0dc0d38f54 WatchSource:0}: Error finding container 94c7286786e9088871506cf9661003f0bf0ba1b8f6874cbf75743e0dc0d38f54: Status 404 returned error can't find the container with id 94c7286786e9088871506cf9661003f0bf0ba1b8f6874cbf75743e0dc0d38f54 Feb 02 12:13:01 crc kubenswrapper[4846]: I0202 12:13:01.354825 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f" exitCode=0 Feb 02 12:13:01 crc kubenswrapper[4846]: I0202 12:13:01.354905 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f"} Feb 02 12:13:01 crc kubenswrapper[4846]: I0202 12:13:01.355432 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"e960baa64942401b64faf1ec81d39bc736adab9ff74905ed38161bca5093c8f7"} Feb 02 12:13:01 crc kubenswrapper[4846]: I0202 12:13:01.356880 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9","Type":"ContainerStarted","Data":"ddb62ef78948ff5ae2be9ffb1977122ab48a11cda504dcc660b5631402f71707"} Feb 02 12:13:01 crc kubenswrapper[4846]: I0202 12:13:01.356929 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9","Type":"ContainerStarted","Data":"94c7286786e9088871506cf9661003f0bf0ba1b8f6874cbf75743e0dc0d38f54"} Feb 02 12:13:01 crc kubenswrapper[4846]: I0202 12:13:01.393502 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.393483631 podStartE2EDuration="1.393483631s" podCreationTimestamp="2026-02-02 12:13:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:13:01.391856087 +0000 UTC m=+212.620442950" watchObservedRunningTime="2026-02-02 12:13:01.393483631 +0000 UTC m=+212.622070494" Feb 02 12:13:01 crc kubenswrapper[4846]: I0202 12:13:01.414143 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.236154 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-697f8fb498-p854p"] Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.236824 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" podUID="709d2acc-338f-4f64-8b22-d81215b650cf" containerName="controller-manager" containerID="cri-o://723701191e8b206603814678dbc3013de9cfd5efa4736b5daff6a257396f50af" gracePeriod=30 Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.254478 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw"] Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.254689 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" podUID="8dc0cce5-a456-43d4-b515-ed033c3df5b3" containerName="route-controller-manager" containerID="cri-o://85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830" gracePeriod=30 Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.374679 4846 generic.go:334] "Generic (PLEG): container finished" podID="709d2acc-338f-4f64-8b22-d81215b650cf" containerID="723701191e8b206603814678dbc3013de9cfd5efa4736b5daff6a257396f50af" exitCode=0 Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.374768 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" event={"ID":"709d2acc-338f-4f64-8b22-d81215b650cf","Type":"ContainerDied","Data":"723701191e8b206603814678dbc3013de9cfd5efa4736b5daff6a257396f50af"} Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.791873 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.798334 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.921419 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc0cce5-a456-43d4-b515-ed033c3df5b3-serving-cert\") pod \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.921478 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxscd\" (UniqueName: \"kubernetes.io/projected/709d2acc-338f-4f64-8b22-d81215b650cf-kube-api-access-xxscd\") pod \"709d2acc-338f-4f64-8b22-d81215b650cf\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.921509 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-client-ca\") pod \"709d2acc-338f-4f64-8b22-d81215b650cf\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.921562 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-config\") pod \"709d2acc-338f-4f64-8b22-d81215b650cf\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.921586 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-client-ca\") pod \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.921600 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-config\") pod \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.921615 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd2hd\" (UniqueName: \"kubernetes.io/projected/8dc0cce5-a456-43d4-b515-ed033c3df5b3-kube-api-access-bd2hd\") pod \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\" (UID: \"8dc0cce5-a456-43d4-b515-ed033c3df5b3\") " Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.921675 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-proxy-ca-bundles\") pod \"709d2acc-338f-4f64-8b22-d81215b650cf\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.921697 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/709d2acc-338f-4f64-8b22-d81215b650cf-serving-cert\") pod \"709d2acc-338f-4f64-8b22-d81215b650cf\" (UID: \"709d2acc-338f-4f64-8b22-d81215b650cf\") " Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.922416 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-client-ca" (OuterVolumeSpecName: "client-ca") pod "709d2acc-338f-4f64-8b22-d81215b650cf" (UID: "709d2acc-338f-4f64-8b22-d81215b650cf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.922463 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-config" (OuterVolumeSpecName: "config") pod "8dc0cce5-a456-43d4-b515-ed033c3df5b3" (UID: "8dc0cce5-a456-43d4-b515-ed033c3df5b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.922661 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-client-ca" (OuterVolumeSpecName: "client-ca") pod "8dc0cce5-a456-43d4-b515-ed033c3df5b3" (UID: "8dc0cce5-a456-43d4-b515-ed033c3df5b3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.922734 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-config" (OuterVolumeSpecName: "config") pod "709d2acc-338f-4f64-8b22-d81215b650cf" (UID: "709d2acc-338f-4f64-8b22-d81215b650cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.922903 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "709d2acc-338f-4f64-8b22-d81215b650cf" (UID: "709d2acc-338f-4f64-8b22-d81215b650cf"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.927373 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/709d2acc-338f-4f64-8b22-d81215b650cf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "709d2acc-338f-4f64-8b22-d81215b650cf" (UID: "709d2acc-338f-4f64-8b22-d81215b650cf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.927489 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/709d2acc-338f-4f64-8b22-d81215b650cf-kube-api-access-xxscd" (OuterVolumeSpecName: "kube-api-access-xxscd") pod "709d2acc-338f-4f64-8b22-d81215b650cf" (UID: "709d2acc-338f-4f64-8b22-d81215b650cf"). InnerVolumeSpecName "kube-api-access-xxscd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.927860 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dc0cce5-a456-43d4-b515-ed033c3df5b3-kube-api-access-bd2hd" (OuterVolumeSpecName: "kube-api-access-bd2hd") pod "8dc0cce5-a456-43d4-b515-ed033c3df5b3" (UID: "8dc0cce5-a456-43d4-b515-ed033c3df5b3"). InnerVolumeSpecName "kube-api-access-bd2hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:13:03 crc kubenswrapper[4846]: I0202 12:13:03.928055 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dc0cce5-a456-43d4-b515-ed033c3df5b3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8dc0cce5-a456-43d4-b515-ed033c3df5b3" (UID: "8dc0cce5-a456-43d4-b515-ed033c3df5b3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.023589 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dc0cce5-a456-43d4-b515-ed033c3df5b3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.023659 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxscd\" (UniqueName: \"kubernetes.io/projected/709d2acc-338f-4f64-8b22-d81215b650cf-kube-api-access-xxscd\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.023675 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.023687 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.023699 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd2hd\" (UniqueName: \"kubernetes.io/projected/8dc0cce5-a456-43d4-b515-ed033c3df5b3-kube-api-access-bd2hd\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.023711 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.023723 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dc0cce5-a456-43d4-b515-ed033c3df5b3-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.023734 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/709d2acc-338f-4f64-8b22-d81215b650cf-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.023747 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/709d2acc-338f-4f64-8b22-d81215b650cf-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.381271 4846 generic.go:334] "Generic (PLEG): container finished" podID="8dc0cce5-a456-43d4-b515-ed033c3df5b3" containerID="85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830" exitCode=0 Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.381336 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" event={"ID":"8dc0cce5-a456-43d4-b515-ed033c3df5b3","Type":"ContainerDied","Data":"85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830"} Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.381364 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" event={"ID":"8dc0cce5-a456-43d4-b515-ed033c3df5b3","Type":"ContainerDied","Data":"763fd47f849c7b65cc06d884c0d4739545d15b5f106cfb008c8df43f644a6887"} Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.381386 4846 scope.go:117] "RemoveContainer" containerID="85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.381483 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.386639 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.386574 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-697f8fb498-p854p" event={"ID":"709d2acc-338f-4f64-8b22-d81215b650cf","Type":"ContainerDied","Data":"af13caf60f921fbeedf0afcd4b9f6f7f42e82de5ece930f5b372bd0a3f46d139"} Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.409252 4846 scope.go:117] "RemoveContainer" containerID="85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830" Feb 02 12:13:04 crc kubenswrapper[4846]: E0202 12:13:04.410025 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830\": container with ID starting with 85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830 not found: ID does not exist" containerID="85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.410086 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830"} err="failed to get container status \"85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830\": rpc error: code = NotFound desc = could not find container \"85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830\": container with ID starting with 85d3f3cb16c6fca01949641bc256ac30f0a05d5b287d96797b0e2c04ad2e0830 not found: ID does not exist" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.410114 4846 scope.go:117] "RemoveContainer" containerID="723701191e8b206603814678dbc3013de9cfd5efa4736b5daff6a257396f50af" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.416657 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw"] Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.419490 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f99846f7-c6hcw"] Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.441570 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-697f8fb498-p854p"] Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.444609 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-697f8fb498-p854p"] Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.453209 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-846564888c-4xh29"] Feb 02 12:13:04 crc kubenswrapper[4846]: E0202 12:13:04.453443 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc0cce5-a456-43d4-b515-ed033c3df5b3" containerName="route-controller-manager" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.453454 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc0cce5-a456-43d4-b515-ed033c3df5b3" containerName="route-controller-manager" Feb 02 12:13:04 crc kubenswrapper[4846]: E0202 12:13:04.453466 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709d2acc-338f-4f64-8b22-d81215b650cf" containerName="controller-manager" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.453472 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="709d2acc-338f-4f64-8b22-d81215b650cf" containerName="controller-manager" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.453557 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dc0cce5-a456-43d4-b515-ed033c3df5b3" containerName="route-controller-manager" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.453568 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="709d2acc-338f-4f64-8b22-d81215b650cf" containerName="controller-manager" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.453939 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.457961 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-54c7b79596-bqfw5"] Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.458555 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.458993 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.459080 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.459110 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.459155 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.459167 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.459431 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.464366 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.464477 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.464781 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.465113 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.465226 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.465301 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.468779 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-846564888c-4xh29"] Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.470734 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.472700 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54c7b79596-bqfw5"] Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.528670 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/175b5961-3f99-4091-a844-54341845fdab-serving-cert\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.528751 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srwqw\" (UniqueName: \"kubernetes.io/projected/175b5961-3f99-4091-a844-54341845fdab-kube-api-access-srwqw\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.528778 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg4q6\" (UniqueName: \"kubernetes.io/projected/e22d2363-876b-47b7-a713-c4f089fbc23b-kube-api-access-gg4q6\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.528822 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-proxy-ca-bundles\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.528856 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-config\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.528881 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-config\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.528911 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-client-ca\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.528942 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-client-ca\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.528973 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e22d2363-876b-47b7-a713-c4f089fbc23b-serving-cert\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.629980 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/175b5961-3f99-4091-a844-54341845fdab-serving-cert\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.630043 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg4q6\" (UniqueName: \"kubernetes.io/projected/e22d2363-876b-47b7-a713-c4f089fbc23b-kube-api-access-gg4q6\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.630067 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srwqw\" (UniqueName: \"kubernetes.io/projected/175b5961-3f99-4091-a844-54341845fdab-kube-api-access-srwqw\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.630088 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-proxy-ca-bundles\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.630118 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-config\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.630142 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-config\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.630170 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-client-ca\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.630197 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-client-ca\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.630228 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e22d2363-876b-47b7-a713-c4f089fbc23b-serving-cert\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.631351 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-client-ca\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.631807 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-client-ca\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.631816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-config\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.631836 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-config\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.631925 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-proxy-ca-bundles\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.634597 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e22d2363-876b-47b7-a713-c4f089fbc23b-serving-cert\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.634693 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/175b5961-3f99-4091-a844-54341845fdab-serving-cert\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.646284 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg4q6\" (UniqueName: \"kubernetes.io/projected/e22d2363-876b-47b7-a713-c4f089fbc23b-kube-api-access-gg4q6\") pod \"route-controller-manager-846564888c-4xh29\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.652310 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srwqw\" (UniqueName: \"kubernetes.io/projected/175b5961-3f99-4091-a844-54341845fdab-kube-api-access-srwqw\") pod \"controller-manager-54c7b79596-bqfw5\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.785885 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.792028 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:04 crc kubenswrapper[4846]: I0202 12:13:04.987919 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-846564888c-4xh29"] Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.030355 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54c7b79596-bqfw5"] Feb 02 12:13:05 crc kubenswrapper[4846]: W0202 12:13:05.038937 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod175b5961_3f99_4091_a844_54341845fdab.slice/crio-f78203f363c0680c7c795ad41003d8d73d36644838290f529f01c59b3e4bf917 WatchSource:0}: Error finding container f78203f363c0680c7c795ad41003d8d73d36644838290f529f01c59b3e4bf917: Status 404 returned error can't find the container with id f78203f363c0680c7c795ad41003d8d73d36644838290f529f01c59b3e4bf917 Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.393533 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" event={"ID":"175b5961-3f99-4091-a844-54341845fdab","Type":"ContainerStarted","Data":"22f58c0bc2debdd6847f952eec813cbdfeed39f7781fb4f035b979a51ad8cd8f"} Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.393778 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" event={"ID":"175b5961-3f99-4091-a844-54341845fdab","Type":"ContainerStarted","Data":"f78203f363c0680c7c795ad41003d8d73d36644838290f529f01c59b3e4bf917"} Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.394824 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.401454 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" event={"ID":"e22d2363-876b-47b7-a713-c4f089fbc23b","Type":"ContainerStarted","Data":"0ee6a741b782499189ba1e8b852cbbec8ec0f9c0f3de0f27c15ec6ee1ddf4008"} Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.401499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" event={"ID":"e22d2363-876b-47b7-a713-c4f089fbc23b","Type":"ContainerStarted","Data":"d8bc76004b6c39f3fd3930f7cffb2ea4c8a58f61587ede37a48944aeb8f6c8b0"} Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.401554 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.401662 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.415824 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" podStartSLOduration=2.415804669 podStartE2EDuration="2.415804669s" podCreationTimestamp="2026-02-02 12:13:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:13:05.412136018 +0000 UTC m=+216.640722891" watchObservedRunningTime="2026-02-02 12:13:05.415804669 +0000 UTC m=+216.644391532" Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.431741 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="709d2acc-338f-4f64-8b22-d81215b650cf" path="/var/lib/kubelet/pods/709d2acc-338f-4f64-8b22-d81215b650cf/volumes" Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.432572 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dc0cce5-a456-43d4-b515-ed033c3df5b3" path="/var/lib/kubelet/pods/8dc0cce5-a456-43d4-b515-ed033c3df5b3/volumes" Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.448618 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" podStartSLOduration=2.448597702 podStartE2EDuration="2.448597702s" podCreationTimestamp="2026-02-02 12:13:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:13:05.431448085 +0000 UTC m=+216.660034948" watchObservedRunningTime="2026-02-02 12:13:05.448597702 +0000 UTC m=+216.677184565" Feb 02 12:13:05 crc kubenswrapper[4846]: I0202 12:13:05.528537 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:09 crc kubenswrapper[4846]: I0202 12:13:09.818842 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-q57ps"] Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.239001 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-54c7b79596-bqfw5"] Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.239767 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" podUID="175b5961-3f99-4091-a844-54341845fdab" containerName="controller-manager" containerID="cri-o://22f58c0bc2debdd6847f952eec813cbdfeed39f7781fb4f035b979a51ad8cd8f" gracePeriod=30 Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.336660 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-846564888c-4xh29"] Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.336853 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" podUID="e22d2363-876b-47b7-a713-c4f089fbc23b" containerName="route-controller-manager" containerID="cri-o://0ee6a741b782499189ba1e8b852cbbec8ec0f9c0f3de0f27c15ec6ee1ddf4008" gracePeriod=30 Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.518525 4846 generic.go:334] "Generic (PLEG): container finished" podID="175b5961-3f99-4091-a844-54341845fdab" containerID="22f58c0bc2debdd6847f952eec813cbdfeed39f7781fb4f035b979a51ad8cd8f" exitCode=0 Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.518604 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" event={"ID":"175b5961-3f99-4091-a844-54341845fdab","Type":"ContainerDied","Data":"22f58c0bc2debdd6847f952eec813cbdfeed39f7781fb4f035b979a51ad8cd8f"} Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.519857 4846 generic.go:334] "Generic (PLEG): container finished" podID="e22d2363-876b-47b7-a713-c4f089fbc23b" containerID="0ee6a741b782499189ba1e8b852cbbec8ec0f9c0f3de0f27c15ec6ee1ddf4008" exitCode=0 Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.519884 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" event={"ID":"e22d2363-876b-47b7-a713-c4f089fbc23b","Type":"ContainerDied","Data":"0ee6a741b782499189ba1e8b852cbbec8ec0f9c0f3de0f27c15ec6ee1ddf4008"} Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.741371 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.811974 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.888953 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg4q6\" (UniqueName: \"kubernetes.io/projected/e22d2363-876b-47b7-a713-c4f089fbc23b-kube-api-access-gg4q6\") pod \"e22d2363-876b-47b7-a713-c4f089fbc23b\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.889041 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-client-ca\") pod \"e22d2363-876b-47b7-a713-c4f089fbc23b\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.889115 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e22d2363-876b-47b7-a713-c4f089fbc23b-serving-cert\") pod \"e22d2363-876b-47b7-a713-c4f089fbc23b\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.889142 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-config\") pod \"e22d2363-876b-47b7-a713-c4f089fbc23b\" (UID: \"e22d2363-876b-47b7-a713-c4f089fbc23b\") " Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.890087 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-client-ca" (OuterVolumeSpecName: "client-ca") pod "e22d2363-876b-47b7-a713-c4f089fbc23b" (UID: "e22d2363-876b-47b7-a713-c4f089fbc23b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.890159 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.890300 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-config" (OuterVolumeSpecName: "config") pod "e22d2363-876b-47b7-a713-c4f089fbc23b" (UID: "e22d2363-876b-47b7-a713-c4f089fbc23b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.894888 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e22d2363-876b-47b7-a713-c4f089fbc23b-kube-api-access-gg4q6" (OuterVolumeSpecName: "kube-api-access-gg4q6") pod "e22d2363-876b-47b7-a713-c4f089fbc23b" (UID: "e22d2363-876b-47b7-a713-c4f089fbc23b"). InnerVolumeSpecName "kube-api-access-gg4q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.895705 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22d2363-876b-47b7-a713-c4f089fbc23b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e22d2363-876b-47b7-a713-c4f089fbc23b" (UID: "e22d2363-876b-47b7-a713-c4f089fbc23b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.994269 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/175b5961-3f99-4091-a844-54341845fdab-serving-cert\") pod \"175b5961-3f99-4091-a844-54341845fdab\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.994342 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-client-ca\") pod \"175b5961-3f99-4091-a844-54341845fdab\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.994384 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srwqw\" (UniqueName: \"kubernetes.io/projected/175b5961-3f99-4091-a844-54341845fdab-kube-api-access-srwqw\") pod \"175b5961-3f99-4091-a844-54341845fdab\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.994401 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-proxy-ca-bundles\") pod \"175b5961-3f99-4091-a844-54341845fdab\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.994504 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-config\") pod \"175b5961-3f99-4091-a844-54341845fdab\" (UID: \"175b5961-3f99-4091-a844-54341845fdab\") " Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.994842 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e22d2363-876b-47b7-a713-c4f089fbc23b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.994856 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e22d2363-876b-47b7-a713-c4f089fbc23b-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.994887 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg4q6\" (UniqueName: \"kubernetes.io/projected/e22d2363-876b-47b7-a713-c4f089fbc23b-kube-api-access-gg4q6\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.995533 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-client-ca" (OuterVolumeSpecName: "client-ca") pod "175b5961-3f99-4091-a844-54341845fdab" (UID: "175b5961-3f99-4091-a844-54341845fdab"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.995637 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-config" (OuterVolumeSpecName: "config") pod "175b5961-3f99-4091-a844-54341845fdab" (UID: "175b5961-3f99-4091-a844-54341845fdab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.995654 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "175b5961-3f99-4091-a844-54341845fdab" (UID: "175b5961-3f99-4091-a844-54341845fdab"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.997342 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/175b5961-3f99-4091-a844-54341845fdab-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "175b5961-3f99-4091-a844-54341845fdab" (UID: "175b5961-3f99-4091-a844-54341845fdab"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:23 crc kubenswrapper[4846]: I0202 12:13:23.997782 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/175b5961-3f99-4091-a844-54341845fdab-kube-api-access-srwqw" (OuterVolumeSpecName: "kube-api-access-srwqw") pod "175b5961-3f99-4091-a844-54341845fdab" (UID: "175b5961-3f99-4091-a844-54341845fdab"). InnerVolumeSpecName "kube-api-access-srwqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.097019 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.097074 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srwqw\" (UniqueName: \"kubernetes.io/projected/175b5961-3f99-4091-a844-54341845fdab-kube-api-access-srwqw\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.097096 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.097120 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/175b5961-3f99-4091-a844-54341845fdab-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.097139 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/175b5961-3f99-4091-a844-54341845fdab-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.469534 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf"] Feb 02 12:13:24 crc kubenswrapper[4846]: E0202 12:13:24.469988 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22d2363-876b-47b7-a713-c4f089fbc23b" containerName="route-controller-manager" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.470012 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22d2363-876b-47b7-a713-c4f089fbc23b" containerName="route-controller-manager" Feb 02 12:13:24 crc kubenswrapper[4846]: E0202 12:13:24.470043 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175b5961-3f99-4091-a844-54341845fdab" containerName="controller-manager" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.470057 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="175b5961-3f99-4091-a844-54341845fdab" containerName="controller-manager" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.470269 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22d2363-876b-47b7-a713-c4f089fbc23b" containerName="route-controller-manager" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.470302 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="175b5961-3f99-4091-a844-54341845fdab" containerName="controller-manager" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.471062 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.472439 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-554f5d57bd-4cw9c"] Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.473166 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.481523 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf"] Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.490605 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-554f5d57bd-4cw9c"] Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.530219 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" event={"ID":"e22d2363-876b-47b7-a713-c4f089fbc23b","Type":"ContainerDied","Data":"d8bc76004b6c39f3fd3930f7cffb2ea4c8a58f61587ede37a48944aeb8f6c8b0"} Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.530269 4846 scope.go:117] "RemoveContainer" containerID="0ee6a741b782499189ba1e8b852cbbec8ec0f9c0f3de0f27c15ec6ee1ddf4008" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.530385 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-846564888c-4xh29" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.537614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" event={"ID":"175b5961-3f99-4091-a844-54341845fdab","Type":"ContainerDied","Data":"f78203f363c0680c7c795ad41003d8d73d36644838290f529f01c59b3e4bf917"} Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.537891 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54c7b79596-bqfw5" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.550844 4846 scope.go:117] "RemoveContainer" containerID="22f58c0bc2debdd6847f952eec813cbdfeed39f7781fb4f035b979a51ad8cd8f" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.557061 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-846564888c-4xh29"] Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.560218 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-846564888c-4xh29"] Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.579851 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-54c7b79596-bqfw5"] Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.583001 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-54c7b79596-bqfw5"] Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.605864 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-client-ca\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.605950 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32d8c062-07f7-4074-ad4b-40dc1817f924-config\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.606002 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32d8c062-07f7-4074-ad4b-40dc1817f924-client-ca\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.606023 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32d8c062-07f7-4074-ad4b-40dc1817f924-serving-cert\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.606116 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjgml\" (UniqueName: \"kubernetes.io/projected/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-kube-api-access-wjgml\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.606171 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-proxy-ca-bundles\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.606218 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-serving-cert\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.606276 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-config\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.606312 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhlfc\" (UniqueName: \"kubernetes.io/projected/32d8c062-07f7-4074-ad4b-40dc1817f924-kube-api-access-vhlfc\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.707753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhlfc\" (UniqueName: \"kubernetes.io/projected/32d8c062-07f7-4074-ad4b-40dc1817f924-kube-api-access-vhlfc\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.707807 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-client-ca\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.707838 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32d8c062-07f7-4074-ad4b-40dc1817f924-config\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.707858 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32d8c062-07f7-4074-ad4b-40dc1817f924-serving-cert\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.707873 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32d8c062-07f7-4074-ad4b-40dc1817f924-client-ca\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.707895 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjgml\" (UniqueName: \"kubernetes.io/projected/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-kube-api-access-wjgml\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.707922 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-proxy-ca-bundles\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.707953 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-serving-cert\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.707980 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-config\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.709507 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32d8c062-07f7-4074-ad4b-40dc1817f924-client-ca\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.709672 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32d8c062-07f7-4074-ad4b-40dc1817f924-config\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.709906 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-proxy-ca-bundles\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.710168 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-client-ca\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.710515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-config\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.711910 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32d8c062-07f7-4074-ad4b-40dc1817f924-serving-cert\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.712592 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-serving-cert\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.727723 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjgml\" (UniqueName: \"kubernetes.io/projected/9c5ae3b2-08ee-4c72-affb-9cdc1177b344-kube-api-access-wjgml\") pod \"controller-manager-554f5d57bd-4cw9c\" (UID: \"9c5ae3b2-08ee-4c72-affb-9cdc1177b344\") " pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.730133 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhlfc\" (UniqueName: \"kubernetes.io/projected/32d8c062-07f7-4074-ad4b-40dc1817f924-kube-api-access-vhlfc\") pod \"route-controller-manager-c46bbf868-8g5tf\" (UID: \"32d8c062-07f7-4074-ad4b-40dc1817f924\") " pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.818504 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:24 crc kubenswrapper[4846]: I0202 12:13:24.834960 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.210955 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf"] Feb 02 12:13:25 crc kubenswrapper[4846]: W0202 12:13:25.218760 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32d8c062_07f7_4074_ad4b_40dc1817f924.slice/crio-2eee6851b2afb52c657b1bd77c4b01a0c6ace1d44106cf86cdc2205e4c57b547 WatchSource:0}: Error finding container 2eee6851b2afb52c657b1bd77c4b01a0c6ace1d44106cf86cdc2205e4c57b547: Status 404 returned error can't find the container with id 2eee6851b2afb52c657b1bd77c4b01a0c6ace1d44106cf86cdc2205e4c57b547 Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.250475 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-554f5d57bd-4cw9c"] Feb 02 12:13:25 crc kubenswrapper[4846]: W0202 12:13:25.252419 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c5ae3b2_08ee_4c72_affb_9cdc1177b344.slice/crio-04c60e9b413ac2b4f90efd8b179fb7da179f0e851f78538ecaa0af60ba68c81d WatchSource:0}: Error finding container 04c60e9b413ac2b4f90efd8b179fb7da179f0e851f78538ecaa0af60ba68c81d: Status 404 returned error can't find the container with id 04c60e9b413ac2b4f90efd8b179fb7da179f0e851f78538ecaa0af60ba68c81d Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.431042 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="175b5961-3f99-4091-a844-54341845fdab" path="/var/lib/kubelet/pods/175b5961-3f99-4091-a844-54341845fdab/volumes" Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.432323 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e22d2363-876b-47b7-a713-c4f089fbc23b" path="/var/lib/kubelet/pods/e22d2363-876b-47b7-a713-c4f089fbc23b/volumes" Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.544251 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" event={"ID":"9c5ae3b2-08ee-4c72-affb-9cdc1177b344","Type":"ContainerStarted","Data":"1e837794374e00c79c54419a48a2ad1df8389d9e650f76487a409d25adfea499"} Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.544299 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" event={"ID":"9c5ae3b2-08ee-4c72-affb-9cdc1177b344","Type":"ContainerStarted","Data":"04c60e9b413ac2b4f90efd8b179fb7da179f0e851f78538ecaa0af60ba68c81d"} Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.544687 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.548010 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" event={"ID":"32d8c062-07f7-4074-ad4b-40dc1817f924","Type":"ContainerStarted","Data":"3707b0f6be7838db9907ca1db4e1a503bcb157bbf701ab52b3bf0ae8dd639147"} Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.548047 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" event={"ID":"32d8c062-07f7-4074-ad4b-40dc1817f924","Type":"ContainerStarted","Data":"2eee6851b2afb52c657b1bd77c4b01a0c6ace1d44106cf86cdc2205e4c57b547"} Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.548889 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.554499 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.563146 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-554f5d57bd-4cw9c" podStartSLOduration=2.5631321419999997 podStartE2EDuration="2.563132142s" podCreationTimestamp="2026-02-02 12:13:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:13:25.562410322 +0000 UTC m=+236.790997185" watchObservedRunningTime="2026-02-02 12:13:25.563132142 +0000 UTC m=+236.791719005" Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.593889 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" podStartSLOduration=2.593866559 podStartE2EDuration="2.593866559s" podCreationTimestamp="2026-02-02 12:13:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:13:25.58291903 +0000 UTC m=+236.811505893" watchObservedRunningTime="2026-02-02 12:13:25.593866559 +0000 UTC m=+236.822453412" Feb 02 12:13:25 crc kubenswrapper[4846]: I0202 12:13:25.689852 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c46bbf868-8g5tf" Feb 02 12:13:34 crc kubenswrapper[4846]: I0202 12:13:34.859329 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" podUID="ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" containerName="oauth-openshift" containerID="cri-o://9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf" gracePeriod=15 Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.370634 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.414540 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-8449b79ffb-jjhr8"] Feb 02 12:13:35 crc kubenswrapper[4846]: E0202 12:13:35.414851 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" containerName="oauth-openshift" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.414872 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" containerName="oauth-openshift" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.415008 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" containerName="oauth-openshift" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.415504 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.437353 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-8449b79ffb-jjhr8"] Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543693 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-error\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543736 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-dir\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543761 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-serving-cert\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543810 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-trusted-ca-bundle\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543853 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-session\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543873 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-ocp-branding-template\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543889 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-cliconfig\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543887 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543927 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjccf\" (UniqueName: \"kubernetes.io/projected/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-kube-api-access-tjccf\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543948 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-provider-selection\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.543975 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-idp-0-file-data\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544004 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-login\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544032 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-router-certs\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544055 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-policies\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544071 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-service-ca\") pod \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\" (UID: \"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0\") " Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544179 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544203 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-service-ca\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544234 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-template-error\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544253 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544274 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-audit-policies\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544294 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c2ch\" (UniqueName: \"kubernetes.io/projected/84eed31e-fe16-466d-9d97-186b23d5dde3-kube-api-access-2c2ch\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544314 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544329 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-router-certs\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544344 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544363 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544387 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-session\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544830 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544900 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.544945 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-template-login\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.545021 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/84eed31e-fe16-466d-9d97-186b23d5dde3-audit-dir\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.545077 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.545195 4846 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.545222 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.545384 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.545439 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.550783 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.551247 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.551659 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.552098 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.552149 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-kube-api-access-tjccf" (OuterVolumeSpecName: "kube-api-access-tjccf") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "kube-api-access-tjccf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.557088 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.558968 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.559149 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.559505 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" (UID: "ad8db62d-d0ce-4ee8-9be7-452786e1c1b0"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.599985 4846 generic.go:334] "Generic (PLEG): container finished" podID="ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" containerID="9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf" exitCode=0 Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.600033 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" event={"ID":"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0","Type":"ContainerDied","Data":"9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf"} Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.600280 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" event={"ID":"ad8db62d-d0ce-4ee8-9be7-452786e1c1b0","Type":"ContainerDied","Data":"cefc6f9d60a83ed6f04d9a8e16bd284f7da588841f6173e060f0226e8594a0fd"} Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.600338 4846 scope.go:117] "RemoveContainer" containerID="9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.600054 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-q57ps" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.620869 4846 scope.go:117] "RemoveContainer" containerID="9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf" Feb 02 12:13:35 crc kubenswrapper[4846]: E0202 12:13:35.621363 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf\": container with ID starting with 9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf not found: ID does not exist" containerID="9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.621417 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf"} err="failed to get container status \"9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf\": rpc error: code = NotFound desc = could not find container \"9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf\": container with ID starting with 9b964bda7933209f7e23714a1e9430fa086a47492f15670d88c96d6a48d1aacf not found: ID does not exist" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.627187 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-q57ps"] Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.629679 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-q57ps"] Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646320 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-audit-policies\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646356 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c2ch\" (UniqueName: \"kubernetes.io/projected/84eed31e-fe16-466d-9d97-186b23d5dde3-kube-api-access-2c2ch\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646397 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-router-certs\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646415 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646435 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646458 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-session\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646486 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646507 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-template-login\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646524 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/84eed31e-fe16-466d-9d97-186b23d5dde3-audit-dir\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646557 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646580 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-service-ca\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646607 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-template-error\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646651 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646691 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646701 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646711 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjccf\" (UniqueName: \"kubernetes.io/projected/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-kube-api-access-tjccf\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646721 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646731 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646739 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646748 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646757 4846 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646766 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646775 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646784 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.646794 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.647364 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-audit-policies\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.647368 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/84eed31e-fe16-466d-9d97-186b23d5dde3-audit-dir\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.647544 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.648020 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-service-ca\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.648729 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.650352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.650421 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-router-certs\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.651869 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-template-login\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.652428 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.652438 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.653091 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-user-template-error\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.654007 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.654746 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/84eed31e-fe16-466d-9d97-186b23d5dde3-v4-0-config-system-session\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.666663 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c2ch\" (UniqueName: \"kubernetes.io/projected/84eed31e-fe16-466d-9d97-186b23d5dde3-kube-api-access-2c2ch\") pod \"oauth-openshift-8449b79ffb-jjhr8\" (UID: \"84eed31e-fe16-466d-9d97-186b23d5dde3\") " pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:35 crc kubenswrapper[4846]: I0202 12:13:35.736814 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:36 crc kubenswrapper[4846]: I0202 12:13:36.136366 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-8449b79ffb-jjhr8"] Feb 02 12:13:36 crc kubenswrapper[4846]: I0202 12:13:36.607656 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" event={"ID":"84eed31e-fe16-466d-9d97-186b23d5dde3","Type":"ContainerStarted","Data":"8fa325accdee677db60d53a0de047e87dcdd02575bdd6418c8098a2ba17c400e"} Feb 02 12:13:36 crc kubenswrapper[4846]: I0202 12:13:36.607984 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:36 crc kubenswrapper[4846]: I0202 12:13:36.607998 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" event={"ID":"84eed31e-fe16-466d-9d97-186b23d5dde3","Type":"ContainerStarted","Data":"c9045ab09d0aa6563ffe89b6011ff53a15686c7363a3de5abfdfb84d15671391"} Feb 02 12:13:37 crc kubenswrapper[4846]: I0202 12:13:37.180771 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" Feb 02 12:13:37 crc kubenswrapper[4846]: I0202 12:13:37.197438 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-8449b79ffb-jjhr8" podStartSLOduration=28.197415907 podStartE2EDuration="28.197415907s" podCreationTimestamp="2026-02-02 12:13:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:13:36.638142628 +0000 UTC m=+247.866729521" watchObservedRunningTime="2026-02-02 12:13:37.197415907 +0000 UTC m=+248.426002800" Feb 02 12:13:37 crc kubenswrapper[4846]: I0202 12:13:37.430957 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad8db62d-d0ce-4ee8-9be7-452786e1c1b0" path="/var/lib/kubelet/pods/ad8db62d-d0ce-4ee8-9be7-452786e1c1b0/volumes" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.545158 4846 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.547206 4846 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.547375 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.547768 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138" gracePeriod=15 Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.547848 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75" gracePeriod=15 Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.547851 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979" gracePeriod=15 Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.547965 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a" gracePeriod=15 Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.548005 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b" gracePeriod=15 Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.548818 4846 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 12:13:38 crc kubenswrapper[4846]: E0202 12:13:38.549093 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549116 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 02 12:13:38 crc kubenswrapper[4846]: E0202 12:13:38.549135 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549149 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 02 12:13:38 crc kubenswrapper[4846]: E0202 12:13:38.549167 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549180 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 12:13:38 crc kubenswrapper[4846]: E0202 12:13:38.549195 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549210 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 12:13:38 crc kubenswrapper[4846]: E0202 12:13:38.549239 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549252 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 02 12:13:38 crc kubenswrapper[4846]: E0202 12:13:38.549268 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549281 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 02 12:13:38 crc kubenswrapper[4846]: E0202 12:13:38.549301 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549344 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549563 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549588 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549614 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549671 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549694 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549724 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 02 12:13:38 crc kubenswrapper[4846]: E0202 12:13:38.549933 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.549949 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.550132 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.590695 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.748105 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.748723 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.748757 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.748785 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.748827 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.748858 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.748897 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.748952 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850171 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850279 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850357 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850412 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850444 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850488 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850531 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850676 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850676 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850737 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850775 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850789 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850776 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.850742 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: I0202 12:13:38.880971 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:13:38 crc kubenswrapper[4846]: W0202 12:13:38.908471 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-26266b594f2e7350c86a747ed49b6feb31e80f8ad008bee4be096cfdc673457b WatchSource:0}: Error finding container 26266b594f2e7350c86a747ed49b6feb31e80f8ad008bee4be096cfdc673457b: Status 404 returned error can't find the container with id 26266b594f2e7350c86a747ed49b6feb31e80f8ad008bee4be096cfdc673457b Feb 02 12:13:38 crc kubenswrapper[4846]: E0202 12:13:38.914861 4846 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.223:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18906ced0b87d049 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 12:13:38.913103945 +0000 UTC m=+250.141690818,LastTimestamp:2026-02-02 12:13:38.913103945 +0000 UTC m=+250.141690818,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.426072 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.426913 4846 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:39 crc kubenswrapper[4846]: E0202 12:13:39.447882 4846 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:39 crc kubenswrapper[4846]: E0202 12:13:39.449103 4846 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:39 crc kubenswrapper[4846]: E0202 12:13:39.449588 4846 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:39 crc kubenswrapper[4846]: E0202 12:13:39.450039 4846 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:39 crc kubenswrapper[4846]: E0202 12:13:39.450706 4846 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.450776 4846 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 02 12:13:39 crc kubenswrapper[4846]: E0202 12:13:39.451162 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="200ms" Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.626545 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b"} Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.626607 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"26266b594f2e7350c86a747ed49b6feb31e80f8ad008bee4be096cfdc673457b"} Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.627291 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.630470 4846 generic.go:334] "Generic (PLEG): container finished" podID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" containerID="ddb62ef78948ff5ae2be9ffb1977122ab48a11cda504dcc660b5631402f71707" exitCode=0 Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.630520 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9","Type":"ContainerDied","Data":"ddb62ef78948ff5ae2be9ffb1977122ab48a11cda504dcc660b5631402f71707"} Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.631833 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.632317 4846 status_manager.go:851] "Failed to get status for pod" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.633460 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.635220 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.636175 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b" exitCode=0 Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.636211 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75" exitCode=0 Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.636223 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979" exitCode=0 Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.636236 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a" exitCode=2 Feb 02 12:13:39 crc kubenswrapper[4846]: I0202 12:13:39.636267 4846 scope.go:117] "RemoveContainer" containerID="ae917ef69fccaad5b0844e6af2f1acb24e26acf402f3fa1ecace203abd22aea7" Feb 02 12:13:39 crc kubenswrapper[4846]: E0202 12:13:39.652528 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="400ms" Feb 02 12:13:40 crc kubenswrapper[4846]: E0202 12:13:40.054009 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="800ms" Feb 02 12:13:40 crc kubenswrapper[4846]: E0202 12:13:40.442202 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:13:40Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:13:40Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:13:40Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T12:13:40Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:40 crc kubenswrapper[4846]: E0202 12:13:40.442600 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:40 crc kubenswrapper[4846]: E0202 12:13:40.442831 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:40 crc kubenswrapper[4846]: E0202 12:13:40.443037 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:40 crc kubenswrapper[4846]: E0202 12:13:40.443301 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:40 crc kubenswrapper[4846]: E0202 12:13:40.443328 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 12:13:40 crc kubenswrapper[4846]: I0202 12:13:40.686680 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 12:13:40 crc kubenswrapper[4846]: E0202 12:13:40.855804 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="1.6s" Feb 02 12:13:40 crc kubenswrapper[4846]: I0202 12:13:40.921379 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 12:13:40 crc kubenswrapper[4846]: I0202 12:13:40.922458 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:40 crc kubenswrapper[4846]: I0202 12:13:40.923113 4846 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:40 crc kubenswrapper[4846]: I0202 12:13:40.923497 4846 status_manager.go:851] "Failed to get status for pod" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:40 crc kubenswrapper[4846]: I0202 12:13:40.923858 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.059278 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.059916 4846 status_manager.go:851] "Failed to get status for pod" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.060335 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.060832 4846 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.080232 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.080358 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.080339 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.080385 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.080404 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.080447 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.080978 4846 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.081005 4846 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.081017 4846 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.181927 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kube-api-access\") pod \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.182112 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-var-lock\") pod \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.182141 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kubelet-dir\") pod \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\" (UID: \"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9\") " Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.182261 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-var-lock" (OuterVolumeSpecName: "var-lock") pod "d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" (UID: "d2a34ae2-14ef-4ca9-b3be-67e43673d5f9"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.182310 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" (UID: "d2a34ae2-14ef-4ca9-b3be-67e43673d5f9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.182429 4846 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-var-lock\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.182449 4846 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.192918 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" (UID: "d2a34ae2-14ef-4ca9-b3be-67e43673d5f9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.283265 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2a34ae2-14ef-4ca9-b3be-67e43673d5f9-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.435447 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 02 12:13:41 crc kubenswrapper[4846]: E0202 12:13:41.486084 4846 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.223:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" volumeName="registry-storage" Feb 02 12:13:41 crc kubenswrapper[4846]: E0202 12:13:41.544360 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-podd2a34ae2_14ef_4ca9_b3be_67e43673d5f9.slice/crio-94c7286786e9088871506cf9661003f0bf0ba1b8f6874cbf75743e0dc0d38f54\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-27feb736f4ad9553012cddbd8f1440bdce91f6e4942b2ae8486f0d6b1393d170\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-podd2a34ae2_14ef_4ca9_b3be_67e43673d5f9.slice\": RecentStats: unable to find data in memory cache]" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.697933 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d2a34ae2-14ef-4ca9-b3be-67e43673d5f9","Type":"ContainerDied","Data":"94c7286786e9088871506cf9661003f0bf0ba1b8f6874cbf75743e0dc0d38f54"} Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.697972 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94c7286786e9088871506cf9661003f0bf0ba1b8f6874cbf75743e0dc0d38f54" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.698039 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.703912 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.704083 4846 status_manager.go:851] "Failed to get status for pod" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: E0202 12:13:41.711111 4846 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.223:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18906ced0b87d049 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 12:13:38.913103945 +0000 UTC m=+250.141690818,LastTimestamp:2026-02-02 12:13:38.913103945 +0000 UTC m=+250.141690818,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.713639 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.714486 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138" exitCode=0 Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.714543 4846 scope.go:117] "RemoveContainer" containerID="0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.714750 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.715509 4846 status_manager.go:851] "Failed to get status for pod" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.715842 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.716135 4846 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.718436 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.719157 4846 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.719561 4846 status_manager.go:851] "Failed to get status for pod" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.736765 4846 scope.go:117] "RemoveContainer" containerID="797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.752495 4846 scope.go:117] "RemoveContainer" containerID="e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.768820 4846 scope.go:117] "RemoveContainer" containerID="edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.782576 4846 scope.go:117] "RemoveContainer" containerID="ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.796032 4846 scope.go:117] "RemoveContainer" containerID="e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.823408 4846 scope.go:117] "RemoveContainer" containerID="0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b" Feb 02 12:13:41 crc kubenswrapper[4846]: E0202 12:13:41.823835 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\": container with ID starting with 0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b not found: ID does not exist" containerID="0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.823872 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b"} err="failed to get container status \"0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\": rpc error: code = NotFound desc = could not find container \"0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b\": container with ID starting with 0d16d126736a01d574694d540cfef8e2aeaf1090274b1a4404397ddf641a719b not found: ID does not exist" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.823895 4846 scope.go:117] "RemoveContainer" containerID="797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75" Feb 02 12:13:41 crc kubenswrapper[4846]: E0202 12:13:41.824683 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\": container with ID starting with 797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75 not found: ID does not exist" containerID="797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.824731 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75"} err="failed to get container status \"797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\": rpc error: code = NotFound desc = could not find container \"797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75\": container with ID starting with 797211e36e3c7db38c4482f4fd54ba7b17dbc39af3ca37fc92c88bf6ac5c6b75 not found: ID does not exist" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.824773 4846 scope.go:117] "RemoveContainer" containerID="e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979" Feb 02 12:13:41 crc kubenswrapper[4846]: E0202 12:13:41.825193 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\": container with ID starting with e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979 not found: ID does not exist" containerID="e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.825233 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979"} err="failed to get container status \"e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\": rpc error: code = NotFound desc = could not find container \"e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979\": container with ID starting with e333c3ddb2f53c076caedc5d71099a43eab665f6b55ee404e889f737676d0979 not found: ID does not exist" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.825259 4846 scope.go:117] "RemoveContainer" containerID="edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a" Feb 02 12:13:41 crc kubenswrapper[4846]: E0202 12:13:41.825492 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\": container with ID starting with edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a not found: ID does not exist" containerID="edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.825512 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a"} err="failed to get container status \"edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\": rpc error: code = NotFound desc = could not find container \"edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a\": container with ID starting with edbe0eccae1d2fad3e8e7bb44dcf70f9ad95cdae0c42fe81c9d9422efd39de8a not found: ID does not exist" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.825529 4846 scope.go:117] "RemoveContainer" containerID="ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138" Feb 02 12:13:41 crc kubenswrapper[4846]: E0202 12:13:41.826738 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\": container with ID starting with ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138 not found: ID does not exist" containerID="ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.826774 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138"} err="failed to get container status \"ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\": rpc error: code = NotFound desc = could not find container \"ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138\": container with ID starting with ad48f0fdd30193e90a91008f97ab0e717f1608f2b1e33e19bf6ccb572adb1138 not found: ID does not exist" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.826814 4846 scope.go:117] "RemoveContainer" containerID="e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9" Feb 02 12:13:41 crc kubenswrapper[4846]: E0202 12:13:41.827275 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\": container with ID starting with e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9 not found: ID does not exist" containerID="e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9" Feb 02 12:13:41 crc kubenswrapper[4846]: I0202 12:13:41.827312 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9"} err="failed to get container status \"e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\": rpc error: code = NotFound desc = could not find container \"e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9\": container with ID starting with e6f2d5d6aa73a2f2bf55cb4035cb444e60860c001c660230e9ca7db139239df9 not found: ID does not exist" Feb 02 12:13:42 crc kubenswrapper[4846]: E0202 12:13:42.456921 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="3.2s" Feb 02 12:13:45 crc kubenswrapper[4846]: E0202 12:13:45.658278 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="6.4s" Feb 02 12:13:49 crc kubenswrapper[4846]: I0202 12:13:49.425371 4846 status_manager.go:851] "Failed to get status for pod" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:49 crc kubenswrapper[4846]: I0202 12:13:49.426201 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:50 crc kubenswrapper[4846]: I0202 12:13:50.910321 4846 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 02 12:13:50 crc kubenswrapper[4846]: I0202 12:13:50.910825 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.422676 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.423648 4846 status_manager.go:851] "Failed to get status for pod" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.424139 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.436830 4846 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e61dd041-dcc8-4102-a956-9e7d1a4ce0b2" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.437378 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e61dd041-dcc8-4102-a956-9e7d1a4ce0b2" Feb 02 12:13:51 crc kubenswrapper[4846]: E0202 12:13:51.438092 4846 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.438679 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:51 crc kubenswrapper[4846]: E0202 12:13:51.714108 4846 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.223:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18906ced0b87d049 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 12:13:38.913103945 +0000 UTC m=+250.141690818,LastTimestamp:2026-02-02 12:13:38.913103945 +0000 UTC m=+250.141690818,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.772684 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.772738 4846 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406" exitCode=1 Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.772801 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406"} Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.773276 4846 scope.go:117] "RemoveContainer" containerID="e5031688a225da918e5e97fd0733e19fc0ec4ca5a8c1f3081e59031f38b72406" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.773686 4846 status_manager.go:851] "Failed to get status for pod" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.774004 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.774328 4846 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.775915 4846 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="e5f195987bf745d43855247a5fed9ab71d6b16654737a9f2604cd8243452218b" exitCode=0 Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.775954 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"e5f195987bf745d43855247a5fed9ab71d6b16654737a9f2604cd8243452218b"} Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.775983 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"92320814ec595f27500f92435abe7e3706f3fa1eb51a6f03739b9c8f7f134ba7"} Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.776435 4846 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e61dd041-dcc8-4102-a956-9e7d1a4ce0b2" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.776486 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e61dd041-dcc8-4102-a956-9e7d1a4ce0b2" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.776712 4846 status_manager.go:851] "Failed to get status for pod" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.777000 4846 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:51 crc kubenswrapper[4846]: E0202 12:13:51.777013 4846 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:51 crc kubenswrapper[4846]: I0202 12:13:51.777432 4846 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.223:6443: connect: connection refused" Feb 02 12:13:52 crc kubenswrapper[4846]: E0202 12:13:52.060389 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.223:6443: connect: connection refused" interval="7s" Feb 02 12:13:52 crc kubenswrapper[4846]: I0202 12:13:52.786643 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 02 12:13:52 crc kubenswrapper[4846]: I0202 12:13:52.786840 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b902d39d8833d855dad1cb67b50c50743240c485d1229f7d4d4652016d1bf6a6"} Feb 02 12:13:52 crc kubenswrapper[4846]: I0202 12:13:52.789976 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"678707918c24c3a28fa054a713846cf0e52d937e45da948550172aa4902b1315"} Feb 02 12:13:52 crc kubenswrapper[4846]: I0202 12:13:52.790013 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"28ac58c1ef337d06e78d60d53f21a85a41c0cfda4009463a5f7f3ea922b94aa8"} Feb 02 12:13:52 crc kubenswrapper[4846]: I0202 12:13:52.790027 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"17ac47434cc0bcfead479aec9ba46f0a8d3a615ad09027eeb00c537f3eefee07"} Feb 02 12:13:52 crc kubenswrapper[4846]: I0202 12:13:52.790038 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e21f8e8c79e943788881574ec0a7394d577eb9f7951e33995b85fc8c36d8efcb"} Feb 02 12:13:53 crc kubenswrapper[4846]: I0202 12:13:53.797986 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bf8521c7409a57da0b28f6e533352e21c3117d7b37ab474ba9b3ecd99b1f4cc6"} Feb 02 12:13:53 crc kubenswrapper[4846]: I0202 12:13:53.798425 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:53 crc kubenswrapper[4846]: I0202 12:13:53.798700 4846 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e61dd041-dcc8-4102-a956-9e7d1a4ce0b2" Feb 02 12:13:53 crc kubenswrapper[4846]: I0202 12:13:53.798785 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e61dd041-dcc8-4102-a956-9e7d1a4ce0b2" Feb 02 12:13:56 crc kubenswrapper[4846]: I0202 12:13:56.439868 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:56 crc kubenswrapper[4846]: I0202 12:13:56.440212 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:56 crc kubenswrapper[4846]: I0202 12:13:56.444715 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:56 crc kubenswrapper[4846]: I0202 12:13:56.559496 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:13:56 crc kubenswrapper[4846]: I0202 12:13:56.564391 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:13:56 crc kubenswrapper[4846]: I0202 12:13:56.812412 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:13:58 crc kubenswrapper[4846]: I0202 12:13:58.817732 4846 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:13:59 crc kubenswrapper[4846]: I0202 12:13:59.443324 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d1db8b3d-841d-4250-a0fc-76cd9db65f27" Feb 02 12:13:59 crc kubenswrapper[4846]: I0202 12:13:59.829480 4846 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e61dd041-dcc8-4102-a956-9e7d1a4ce0b2" Feb 02 12:13:59 crc kubenswrapper[4846]: I0202 12:13:59.829515 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e61dd041-dcc8-4102-a956-9e7d1a4ce0b2" Feb 02 12:13:59 crc kubenswrapper[4846]: I0202 12:13:59.834042 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d1db8b3d-841d-4250-a0fc-76cd9db65f27" Feb 02 12:13:59 crc kubenswrapper[4846]: I0202 12:13:59.835104 4846 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://e21f8e8c79e943788881574ec0a7394d577eb9f7951e33995b85fc8c36d8efcb" Feb 02 12:13:59 crc kubenswrapper[4846]: I0202 12:13:59.835124 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:14:00 crc kubenswrapper[4846]: I0202 12:14:00.835214 4846 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e61dd041-dcc8-4102-a956-9e7d1a4ce0b2" Feb 02 12:14:00 crc kubenswrapper[4846]: I0202 12:14:00.835531 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e61dd041-dcc8-4102-a956-9e7d1a4ce0b2" Feb 02 12:14:00 crc kubenswrapper[4846]: I0202 12:14:00.838998 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d1db8b3d-841d-4250-a0fc-76cd9db65f27" Feb 02 12:14:08 crc kubenswrapper[4846]: I0202 12:14:08.105476 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 02 12:14:09 crc kubenswrapper[4846]: I0202 12:14:09.014042 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 02 12:14:09 crc kubenswrapper[4846]: I0202 12:14:09.159114 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 02 12:14:09 crc kubenswrapper[4846]: I0202 12:14:09.375540 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 02 12:14:09 crc kubenswrapper[4846]: I0202 12:14:09.632723 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 02 12:14:09 crc kubenswrapper[4846]: I0202 12:14:09.874503 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 02 12:14:09 crc kubenswrapper[4846]: I0202 12:14:09.982950 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.190280 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.330977 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.440540 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.524133 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.638576 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.647549 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.655822 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.815924 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.891445 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.913997 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 12:14:10 crc kubenswrapper[4846]: I0202 12:14:10.992981 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.059465 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.073442 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.182745 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.259066 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.272199 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.288474 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.452131 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.462162 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.493594 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.785408 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.794616 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.804693 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.904281 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.907348 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.974469 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 02 12:14:11 crc kubenswrapper[4846]: I0202 12:14:11.993864 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.035896 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.051015 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.100327 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.102347 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.246906 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.287435 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.304917 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.327392 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.335335 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.362382 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.403526 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.431755 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.432867 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.486736 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.550358 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.600571 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.736776 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.738980 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.797493 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.799400 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.845136 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.849041 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 02 12:14:12 crc kubenswrapper[4846]: I0202 12:14:12.879507 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.310133 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.326678 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.337959 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.429378 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.547142 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.561067 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.589066 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.601339 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.639820 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.862107 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.922311 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.964504 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 02 12:14:13 crc kubenswrapper[4846]: I0202 12:14:13.992259 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.044866 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.070920 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.141722 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.235415 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.268267 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.269767 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.298788 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.309705 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.471481 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.490558 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.511583 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.584456 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.670185 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.677950 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.721451 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.741282 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.786348 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.797441 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.814298 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.833390 4846 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.833551 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=36.833516234 podStartE2EDuration="36.833516234s" podCreationTimestamp="2026-02-02 12:13:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:13:58.373951341 +0000 UTC m=+269.602538204" watchObservedRunningTime="2026-02-02 12:14:14.833516234 +0000 UTC m=+286.062103107" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.840853 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.840918 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.846064 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 12:14:14 crc kubenswrapper[4846]: I0202 12:14:14.860108 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.860091748 podStartE2EDuration="16.860091748s" podCreationTimestamp="2026-02-02 12:13:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:14:14.858945748 +0000 UTC m=+286.087532611" watchObservedRunningTime="2026-02-02 12:14:14.860091748 +0000 UTC m=+286.088678611" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.001392 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.046234 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.096165 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.327872 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.334669 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.360465 4846 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.553883 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.561722 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.575209 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.658090 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.663256 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.724600 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.772502 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 02 12:14:15 crc kubenswrapper[4846]: I0202 12:14:15.779279 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.069238 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.136594 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.376075 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.507871 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.508131 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.524467 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.569920 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.679350 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.723344 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.759976 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.808500 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.904288 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 02 12:14:16 crc kubenswrapper[4846]: I0202 12:14:16.930294 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.007220 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.007296 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.020253 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.149285 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.161254 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.181672 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.282937 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.305837 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.341465 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.405209 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.582530 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.584552 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.585192 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.631990 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.638755 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.670233 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.716853 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.754136 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.771429 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.824065 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.848704 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 02 12:14:17 crc kubenswrapper[4846]: I0202 12:14:17.927454 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.120336 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.201475 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.211862 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.219321 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.233229 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.244243 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.343861 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.350678 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.416975 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.418489 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.490100 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.526685 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.592557 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.752394 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.772557 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.816030 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 12:14:18 crc kubenswrapper[4846]: I0202 12:14:18.954732 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.005775 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.070272 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.149807 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.362307 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.452012 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.456686 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.576558 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.621898 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.659541 4846 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.678136 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.693859 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.717718 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.770515 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.839286 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.933613 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 02 12:14:19 crc kubenswrapper[4846]: I0202 12:14:19.952510 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.001656 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.002206 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.055603 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.057691 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.184211 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.200114 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.495004 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.504614 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.525738 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.653649 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.937766 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.938041 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.965923 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.985843 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 02 12:14:20 crc kubenswrapper[4846]: I0202 12:14:20.994511 4846 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.002952 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.040448 4846 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.040770 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b" gracePeriod=5 Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.082745 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.156680 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.215006 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.262116 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.328824 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.332788 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.489647 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.502494 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.578802 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.637266 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.756399 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.765745 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.780864 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.904323 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.912401 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.913219 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 02 12:14:21 crc kubenswrapper[4846]: I0202 12:14:21.994222 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.030417 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.192292 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.222175 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.262077 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.331961 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.334218 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.365610 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.414138 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.484287 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.556141 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.559026 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.722787 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.789643 4846 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.797335 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 02 12:14:22 crc kubenswrapper[4846]: I0202 12:14:22.870317 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 02 12:14:23 crc kubenswrapper[4846]: I0202 12:14:23.208822 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 02 12:14:23 crc kubenswrapper[4846]: I0202 12:14:23.288963 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 02 12:14:23 crc kubenswrapper[4846]: I0202 12:14:23.346053 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 02 12:14:23 crc kubenswrapper[4846]: I0202 12:14:23.368673 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 02 12:14:23 crc kubenswrapper[4846]: I0202 12:14:23.424301 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 02 12:14:23 crc kubenswrapper[4846]: I0202 12:14:23.461775 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 02 12:14:23 crc kubenswrapper[4846]: I0202 12:14:23.684726 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 02 12:14:23 crc kubenswrapper[4846]: I0202 12:14:23.801022 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 02 12:14:23 crc kubenswrapper[4846]: I0202 12:14:23.846649 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 02 12:14:23 crc kubenswrapper[4846]: I0202 12:14:23.902804 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 02 12:14:24 crc kubenswrapper[4846]: I0202 12:14:24.007402 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 12:14:24 crc kubenswrapper[4846]: I0202 12:14:24.151137 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 02 12:14:24 crc kubenswrapper[4846]: I0202 12:14:24.261074 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 02 12:14:24 crc kubenswrapper[4846]: I0202 12:14:24.457306 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 02 12:14:24 crc kubenswrapper[4846]: I0202 12:14:24.522729 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 02 12:14:24 crc kubenswrapper[4846]: I0202 12:14:24.725523 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 02 12:14:24 crc kubenswrapper[4846]: I0202 12:14:24.902404 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 02 12:14:24 crc kubenswrapper[4846]: I0202 12:14:24.907761 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 02 12:14:24 crc kubenswrapper[4846]: I0202 12:14:24.972404 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 02 12:14:25 crc kubenswrapper[4846]: I0202 12:14:25.013359 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 02 12:14:25 crc kubenswrapper[4846]: I0202 12:14:25.058078 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 02 12:14:25 crc kubenswrapper[4846]: I0202 12:14:25.176182 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 02 12:14:25 crc kubenswrapper[4846]: I0202 12:14:25.193474 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 02 12:14:25 crc kubenswrapper[4846]: I0202 12:14:25.367821 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 02 12:14:25 crc kubenswrapper[4846]: I0202 12:14:25.485225 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 02 12:14:25 crc kubenswrapper[4846]: I0202 12:14:25.653206 4846 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 02 12:14:25 crc kubenswrapper[4846]: I0202 12:14:25.726600 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 02 12:14:25 crc kubenswrapper[4846]: I0202 12:14:25.866095 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.616260 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.616639 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.711823 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.768709 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.768759 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.768796 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.768826 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.768848 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.768848 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.768895 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.768905 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.769019 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.769112 4846 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.769125 4846 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.769133 4846 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.769140 4846 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.775667 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.870027 4846 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.972999 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.973052 4846 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b" exitCode=137 Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.973105 4846 scope.go:117] "RemoveContainer" containerID="adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.973226 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.989006 4846 scope.go:117] "RemoveContainer" containerID="adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b" Feb 02 12:14:26 crc kubenswrapper[4846]: E0202 12:14:26.989435 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b\": container with ID starting with adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b not found: ID does not exist" containerID="adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b" Feb 02 12:14:26 crc kubenswrapper[4846]: I0202 12:14:26.989491 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b"} err="failed to get container status \"adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b\": rpc error: code = NotFound desc = could not find container \"adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b\": container with ID starting with adb4ae958f27da839f95cc1b3e57d8ab87b7b7692282fccd4274733e5b1a4a4b not found: ID does not exist" Feb 02 12:14:27 crc kubenswrapper[4846]: I0202 12:14:27.165882 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 02 12:14:27 crc kubenswrapper[4846]: I0202 12:14:27.248385 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 02 12:14:27 crc kubenswrapper[4846]: I0202 12:14:27.290801 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 02 12:14:27 crc kubenswrapper[4846]: I0202 12:14:27.430379 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 02 12:14:27 crc kubenswrapper[4846]: I0202 12:14:27.430686 4846 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 02 12:14:27 crc kubenswrapper[4846]: I0202 12:14:27.439324 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 02 12:14:27 crc kubenswrapper[4846]: I0202 12:14:27.439362 4846 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="08a299bc-47bd-4e91-9138-7430a9cfb5c1" Feb 02 12:14:27 crc kubenswrapper[4846]: I0202 12:14:27.442938 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 02 12:14:27 crc kubenswrapper[4846]: I0202 12:14:27.442983 4846 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="08a299bc-47bd-4e91-9138-7430a9cfb5c1" Feb 02 12:14:27 crc kubenswrapper[4846]: I0202 12:14:27.801040 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 02 12:14:29 crc kubenswrapper[4846]: I0202 12:14:29.220935 4846 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.167190 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb"] Feb 02 12:15:00 crc kubenswrapper[4846]: E0202 12:15:00.167920 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" containerName="installer" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.167936 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" containerName="installer" Feb 02 12:15:00 crc kubenswrapper[4846]: E0202 12:15:00.167951 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.167957 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.168067 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2a34ae2-14ef-4ca9-b3be-67e43673d5f9" containerName="installer" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.168080 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.168530 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.172911 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.180851 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.184335 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb"] Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.249452 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65fb3012-a93f-44f3-b880-06d0c0bdc839-config-volume\") pod \"collect-profiles-29500575-zrjwb\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.249510 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65fb3012-a93f-44f3-b880-06d0c0bdc839-secret-volume\") pod \"collect-profiles-29500575-zrjwb\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.249551 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rqfx\" (UniqueName: \"kubernetes.io/projected/65fb3012-a93f-44f3-b880-06d0c0bdc839-kube-api-access-2rqfx\") pod \"collect-profiles-29500575-zrjwb\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.351002 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rqfx\" (UniqueName: \"kubernetes.io/projected/65fb3012-a93f-44f3-b880-06d0c0bdc839-kube-api-access-2rqfx\") pod \"collect-profiles-29500575-zrjwb\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.351134 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65fb3012-a93f-44f3-b880-06d0c0bdc839-config-volume\") pod \"collect-profiles-29500575-zrjwb\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.351161 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65fb3012-a93f-44f3-b880-06d0c0bdc839-secret-volume\") pod \"collect-profiles-29500575-zrjwb\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.352069 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65fb3012-a93f-44f3-b880-06d0c0bdc839-config-volume\") pod \"collect-profiles-29500575-zrjwb\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.360193 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65fb3012-a93f-44f3-b880-06d0c0bdc839-secret-volume\") pod \"collect-profiles-29500575-zrjwb\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.378198 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rqfx\" (UniqueName: \"kubernetes.io/projected/65fb3012-a93f-44f3-b880-06d0c0bdc839-kube-api-access-2rqfx\") pod \"collect-profiles-29500575-zrjwb\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.478972 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.479032 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.486565 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:00 crc kubenswrapper[4846]: I0202 12:15:00.878414 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb"] Feb 02 12:15:01 crc kubenswrapper[4846]: I0202 12:15:01.170068 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" event={"ID":"65fb3012-a93f-44f3-b880-06d0c0bdc839","Type":"ContainerStarted","Data":"b7907a8b432b7c421a1995997c6296b92a4d5ecb0be3e314160efb0b1dd0ff0b"} Feb 02 12:15:01 crc kubenswrapper[4846]: I0202 12:15:01.170414 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" event={"ID":"65fb3012-a93f-44f3-b880-06d0c0bdc839","Type":"ContainerStarted","Data":"ebb4fcc4a2115009a5e58488dce8590cb1d76149586eed51f93128bfcbb00124"} Feb 02 12:15:01 crc kubenswrapper[4846]: I0202 12:15:01.188249 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" podStartSLOduration=1.188227066 podStartE2EDuration="1.188227066s" podCreationTimestamp="2026-02-02 12:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:15:01.185040632 +0000 UTC m=+332.413627495" watchObservedRunningTime="2026-02-02 12:15:01.188227066 +0000 UTC m=+332.416813929" Feb 02 12:15:02 crc kubenswrapper[4846]: I0202 12:15:02.176438 4846 generic.go:334] "Generic (PLEG): container finished" podID="65fb3012-a93f-44f3-b880-06d0c0bdc839" containerID="b7907a8b432b7c421a1995997c6296b92a4d5ecb0be3e314160efb0b1dd0ff0b" exitCode=0 Feb 02 12:15:02 crc kubenswrapper[4846]: I0202 12:15:02.176477 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" event={"ID":"65fb3012-a93f-44f3-b880-06d0c0bdc839","Type":"ContainerDied","Data":"b7907a8b432b7c421a1995997c6296b92a4d5ecb0be3e314160efb0b1dd0ff0b"} Feb 02 12:15:03 crc kubenswrapper[4846]: I0202 12:15:03.416168 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:03 crc kubenswrapper[4846]: I0202 12:15:03.489925 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65fb3012-a93f-44f3-b880-06d0c0bdc839-secret-volume\") pod \"65fb3012-a93f-44f3-b880-06d0c0bdc839\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " Feb 02 12:15:03 crc kubenswrapper[4846]: I0202 12:15:03.490091 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65fb3012-a93f-44f3-b880-06d0c0bdc839-config-volume\") pod \"65fb3012-a93f-44f3-b880-06d0c0bdc839\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " Feb 02 12:15:03 crc kubenswrapper[4846]: I0202 12:15:03.490130 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rqfx\" (UniqueName: \"kubernetes.io/projected/65fb3012-a93f-44f3-b880-06d0c0bdc839-kube-api-access-2rqfx\") pod \"65fb3012-a93f-44f3-b880-06d0c0bdc839\" (UID: \"65fb3012-a93f-44f3-b880-06d0c0bdc839\") " Feb 02 12:15:03 crc kubenswrapper[4846]: I0202 12:15:03.491151 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65fb3012-a93f-44f3-b880-06d0c0bdc839-config-volume" (OuterVolumeSpecName: "config-volume") pod "65fb3012-a93f-44f3-b880-06d0c0bdc839" (UID: "65fb3012-a93f-44f3-b880-06d0c0bdc839"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:15:03 crc kubenswrapper[4846]: I0202 12:15:03.495020 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65fb3012-a93f-44f3-b880-06d0c0bdc839-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "65fb3012-a93f-44f3-b880-06d0c0bdc839" (UID: "65fb3012-a93f-44f3-b880-06d0c0bdc839"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:15:03 crc kubenswrapper[4846]: I0202 12:15:03.495292 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65fb3012-a93f-44f3-b880-06d0c0bdc839-kube-api-access-2rqfx" (OuterVolumeSpecName: "kube-api-access-2rqfx") pod "65fb3012-a93f-44f3-b880-06d0c0bdc839" (UID: "65fb3012-a93f-44f3-b880-06d0c0bdc839"). InnerVolumeSpecName "kube-api-access-2rqfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:15:03 crc kubenswrapper[4846]: I0202 12:15:03.591567 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65fb3012-a93f-44f3-b880-06d0c0bdc839-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 12:15:03 crc kubenswrapper[4846]: I0202 12:15:03.591614 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rqfx\" (UniqueName: \"kubernetes.io/projected/65fb3012-a93f-44f3-b880-06d0c0bdc839-kube-api-access-2rqfx\") on node \"crc\" DevicePath \"\"" Feb 02 12:15:03 crc kubenswrapper[4846]: I0202 12:15:03.591645 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65fb3012-a93f-44f3-b880-06d0c0bdc839-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 12:15:04 crc kubenswrapper[4846]: I0202 12:15:04.188230 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" event={"ID":"65fb3012-a93f-44f3-b880-06d0c0bdc839","Type":"ContainerDied","Data":"ebb4fcc4a2115009a5e58488dce8590cb1d76149586eed51f93128bfcbb00124"} Feb 02 12:15:04 crc kubenswrapper[4846]: I0202 12:15:04.188265 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebb4fcc4a2115009a5e58488dce8590cb1d76149586eed51f93128bfcbb00124" Feb 02 12:15:04 crc kubenswrapper[4846]: I0202 12:15:04.188288 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb" Feb 02 12:15:30 crc kubenswrapper[4846]: I0202 12:15:30.478703 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:15:30 crc kubenswrapper[4846]: I0202 12:15:30.479326 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.435571 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-scvzn"] Feb 02 12:15:34 crc kubenswrapper[4846]: E0202 12:15:34.436388 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65fb3012-a93f-44f3-b880-06d0c0bdc839" containerName="collect-profiles" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.436407 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="65fb3012-a93f-44f3-b880-06d0c0bdc839" containerName="collect-profiles" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.436559 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="65fb3012-a93f-44f3-b880-06d0c0bdc839" containerName="collect-profiles" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.437220 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.451058 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-scvzn"] Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.582483 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ab470b6f-aaf6-46fb-9fb1-505034807a11-ca-trust-extracted\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.582555 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ab470b6f-aaf6-46fb-9fb1-505034807a11-registry-certificates\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.582613 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ab470b6f-aaf6-46fb-9fb1-505034807a11-bound-sa-token\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.582660 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ab470b6f-aaf6-46fb-9fb1-505034807a11-registry-tls\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.582695 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ab470b6f-aaf6-46fb-9fb1-505034807a11-trusted-ca\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.582717 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr289\" (UniqueName: \"kubernetes.io/projected/ab470b6f-aaf6-46fb-9fb1-505034807a11-kube-api-access-lr289\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.582962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.583041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ab470b6f-aaf6-46fb-9fb1-505034807a11-installation-pull-secrets\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.607578 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.684342 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ab470b6f-aaf6-46fb-9fb1-505034807a11-installation-pull-secrets\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.684391 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ab470b6f-aaf6-46fb-9fb1-505034807a11-ca-trust-extracted\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.684427 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ab470b6f-aaf6-46fb-9fb1-505034807a11-registry-certificates\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.684460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ab470b6f-aaf6-46fb-9fb1-505034807a11-bound-sa-token\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.684479 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ab470b6f-aaf6-46fb-9fb1-505034807a11-registry-tls\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.684504 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ab470b6f-aaf6-46fb-9fb1-505034807a11-trusted-ca\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.684526 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr289\" (UniqueName: \"kubernetes.io/projected/ab470b6f-aaf6-46fb-9fb1-505034807a11-kube-api-access-lr289\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.685050 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ab470b6f-aaf6-46fb-9fb1-505034807a11-ca-trust-extracted\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.686006 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ab470b6f-aaf6-46fb-9fb1-505034807a11-registry-certificates\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.686298 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ab470b6f-aaf6-46fb-9fb1-505034807a11-trusted-ca\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.689904 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ab470b6f-aaf6-46fb-9fb1-505034807a11-installation-pull-secrets\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.690152 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ab470b6f-aaf6-46fb-9fb1-505034807a11-registry-tls\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.701480 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ab470b6f-aaf6-46fb-9fb1-505034807a11-bound-sa-token\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.701613 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr289\" (UniqueName: \"kubernetes.io/projected/ab470b6f-aaf6-46fb-9fb1-505034807a11-kube-api-access-lr289\") pod \"image-registry-66df7c8f76-scvzn\" (UID: \"ab470b6f-aaf6-46fb-9fb1-505034807a11\") " pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:34 crc kubenswrapper[4846]: I0202 12:15:34.752888 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:35 crc kubenswrapper[4846]: I0202 12:15:35.124230 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-scvzn"] Feb 02 12:15:35 crc kubenswrapper[4846]: I0202 12:15:35.349796 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" event={"ID":"ab470b6f-aaf6-46fb-9fb1-505034807a11","Type":"ContainerStarted","Data":"f963642c1adb672159f992d911a47711023bbfe8bcad100b597c7efa561c1017"} Feb 02 12:15:35 crc kubenswrapper[4846]: I0202 12:15:35.350368 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:35 crc kubenswrapper[4846]: I0202 12:15:35.350448 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" event={"ID":"ab470b6f-aaf6-46fb-9fb1-505034807a11","Type":"ContainerStarted","Data":"4309d2bc377e82aa70871868359f81b6b3cb71fcd0a8cebcdc311357669f9086"} Feb 02 12:15:35 crc kubenswrapper[4846]: I0202 12:15:35.372516 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" podStartSLOduration=1.372490969 podStartE2EDuration="1.372490969s" podCreationTimestamp="2026-02-02 12:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:15:35.36836155 +0000 UTC m=+366.596948463" watchObservedRunningTime="2026-02-02 12:15:35.372490969 +0000 UTC m=+366.601077832" Feb 02 12:15:54 crc kubenswrapper[4846]: I0202 12:15:54.758464 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-scvzn" Feb 02 12:15:54 crc kubenswrapper[4846]: I0202 12:15:54.815226 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94n5d"] Feb 02 12:16:00 crc kubenswrapper[4846]: I0202 12:16:00.479221 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:16:00 crc kubenswrapper[4846]: I0202 12:16:00.479853 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:16:00 crc kubenswrapper[4846]: I0202 12:16:00.479904 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:16:00 crc kubenswrapper[4846]: I0202 12:16:00.480583 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e960baa64942401b64faf1ec81d39bc736adab9ff74905ed38161bca5093c8f7"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:16:00 crc kubenswrapper[4846]: I0202 12:16:00.480659 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://e960baa64942401b64faf1ec81d39bc736adab9ff74905ed38161bca5093c8f7" gracePeriod=600 Feb 02 12:16:01 crc kubenswrapper[4846]: I0202 12:16:01.489716 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="e960baa64942401b64faf1ec81d39bc736adab9ff74905ed38161bca5093c8f7" exitCode=0 Feb 02 12:16:01 crc kubenswrapper[4846]: I0202 12:16:01.489844 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"e960baa64942401b64faf1ec81d39bc736adab9ff74905ed38161bca5093c8f7"} Feb 02 12:16:01 crc kubenswrapper[4846]: I0202 12:16:01.490114 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"a4395520eea0717a00d2525b7efc36d173bd70de14fcc8d8b3411afe2fc41cae"} Feb 02 12:16:01 crc kubenswrapper[4846]: I0202 12:16:01.490140 4846 scope.go:117] "RemoveContainer" containerID="b661c0b1ca8aba9d784557cf40d8c20de3e22ce2ceb0c5c3743471a23604805f" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.307111 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pjvpk"] Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.308555 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pjvpk" podUID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerName="registry-server" containerID="cri-o://8e8b83ecde2490950dc76da4e5f92d61cce616b29e52fd40b0bee1afae697296" gracePeriod=30 Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.325794 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gmnp2"] Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.326116 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gmnp2" podUID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerName="registry-server" containerID="cri-o://b49bfd2884c82785d674898ef7d77801330d101aaffeef0dc55a7dfd9c854c56" gracePeriod=30 Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.335730 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zbpnp"] Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.335941 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" podUID="34dedc1d-10d1-4ca0-b059-99048ba37464" containerName="marketplace-operator" containerID="cri-o://ceeb99133435ec0fb619d2010f5a3b2136022c9011c8ee1f62129edd65db609a" gracePeriod=30 Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.343105 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p558w"] Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.343336 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p558w" podUID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerName="registry-server" containerID="cri-o://a174332179e0dd7320d061b3d003c07def4344aad74ef063a34205248ea67d06" gracePeriod=30 Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.354317 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n44pv"] Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.354845 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n44pv" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerName="registry-server" containerID="cri-o://ba291467c7fb9876c4109dfe1171cb8179f89c87fdc1e8aa3cbe5780f0dddf04" gracePeriod=30 Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.379723 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kgbzq"] Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.380912 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.383933 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kgbzq"] Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.562565 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17938097-4a7b-4e18-9e26-4ada5f07b343-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kgbzq\" (UID: \"17938097-4a7b-4e18-9e26-4ada5f07b343\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.562949 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17938097-4a7b-4e18-9e26-4ada5f07b343-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kgbzq\" (UID: \"17938097-4a7b-4e18-9e26-4ada5f07b343\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.562979 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpl8t\" (UniqueName: \"kubernetes.io/projected/17938097-4a7b-4e18-9e26-4ada5f07b343-kube-api-access-fpl8t\") pod \"marketplace-operator-79b997595-kgbzq\" (UID: \"17938097-4a7b-4e18-9e26-4ada5f07b343\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.581518 4846 generic.go:334] "Generic (PLEG): container finished" podID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerID="a174332179e0dd7320d061b3d003c07def4344aad74ef063a34205248ea67d06" exitCode=0 Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.581577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p558w" event={"ID":"14da97a9-4b68-458d-858a-6ba8b67cb749","Type":"ContainerDied","Data":"a174332179e0dd7320d061b3d003c07def4344aad74ef063a34205248ea67d06"} Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.585335 4846 generic.go:334] "Generic (PLEG): container finished" podID="34dedc1d-10d1-4ca0-b059-99048ba37464" containerID="ceeb99133435ec0fb619d2010f5a3b2136022c9011c8ee1f62129edd65db609a" exitCode=0 Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.585586 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" event={"ID":"34dedc1d-10d1-4ca0-b059-99048ba37464","Type":"ContainerDied","Data":"ceeb99133435ec0fb619d2010f5a3b2136022c9011c8ee1f62129edd65db609a"} Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.597099 4846 generic.go:334] "Generic (PLEG): container finished" podID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerID="8e8b83ecde2490950dc76da4e5f92d61cce616b29e52fd40b0bee1afae697296" exitCode=0 Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.597198 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjvpk" event={"ID":"3b5b4c6f-368c-47ba-87fc-7309cac15d9b","Type":"ContainerDied","Data":"8e8b83ecde2490950dc76da4e5f92d61cce616b29e52fd40b0bee1afae697296"} Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.608938 4846 generic.go:334] "Generic (PLEG): container finished" podID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerID="b49bfd2884c82785d674898ef7d77801330d101aaffeef0dc55a7dfd9c854c56" exitCode=0 Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.609010 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmnp2" event={"ID":"98f32d89-a132-4cb9-a647-b48848e1b18a","Type":"ContainerDied","Data":"b49bfd2884c82785d674898ef7d77801330d101aaffeef0dc55a7dfd9c854c56"} Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.624702 4846 generic.go:334] "Generic (PLEG): container finished" podID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerID="ba291467c7fb9876c4109dfe1171cb8179f89c87fdc1e8aa3cbe5780f0dddf04" exitCode=0 Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.624749 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n44pv" event={"ID":"64a47758-13e5-4a28-b781-2bdf84b870fd","Type":"ContainerDied","Data":"ba291467c7fb9876c4109dfe1171cb8179f89c87fdc1e8aa3cbe5780f0dddf04"} Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.664031 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17938097-4a7b-4e18-9e26-4ada5f07b343-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kgbzq\" (UID: \"17938097-4a7b-4e18-9e26-4ada5f07b343\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.664085 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17938097-4a7b-4e18-9e26-4ada5f07b343-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kgbzq\" (UID: \"17938097-4a7b-4e18-9e26-4ada5f07b343\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.664159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpl8t\" (UniqueName: \"kubernetes.io/projected/17938097-4a7b-4e18-9e26-4ada5f07b343-kube-api-access-fpl8t\") pod \"marketplace-operator-79b997595-kgbzq\" (UID: \"17938097-4a7b-4e18-9e26-4ada5f07b343\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.667325 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17938097-4a7b-4e18-9e26-4ada5f07b343-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kgbzq\" (UID: \"17938097-4a7b-4e18-9e26-4ada5f07b343\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.678038 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/17938097-4a7b-4e18-9e26-4ada5f07b343-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kgbzq\" (UID: \"17938097-4a7b-4e18-9e26-4ada5f07b343\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.685486 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpl8t\" (UniqueName: \"kubernetes.io/projected/17938097-4a7b-4e18-9e26-4ada5f07b343-kube-api-access-fpl8t\") pod \"marketplace-operator-79b997595-kgbzq\" (UID: \"17938097-4a7b-4e18-9e26-4ada5f07b343\") " pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.756482 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.763548 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.769313 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.778007 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.790111 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.798435 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.865867 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-utilities\") pod \"98f32d89-a132-4cb9-a647-b48848e1b18a\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.866074 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-utilities\") pod \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.866201 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4rgz\" (UniqueName: \"kubernetes.io/projected/14da97a9-4b68-458d-858a-6ba8b67cb749-kube-api-access-b4rgz\") pod \"14da97a9-4b68-458d-858a-6ba8b67cb749\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.866302 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-utilities\") pod \"14da97a9-4b68-458d-858a-6ba8b67cb749\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.866397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-catalog-content\") pod \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.866486 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8rtj\" (UniqueName: \"kubernetes.io/projected/98f32d89-a132-4cb9-a647-b48848e1b18a-kube-api-access-s8rtj\") pod \"98f32d89-a132-4cb9-a647-b48848e1b18a\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.866737 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-catalog-content\") pod \"98f32d89-a132-4cb9-a647-b48848e1b18a\" (UID: \"98f32d89-a132-4cb9-a647-b48848e1b18a\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.866905 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-catalog-content\") pod \"14da97a9-4b68-458d-858a-6ba8b67cb749\" (UID: \"14da97a9-4b68-458d-858a-6ba8b67cb749\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.867050 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt56f\" (UniqueName: \"kubernetes.io/projected/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-kube-api-access-pt56f\") pod \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\" (UID: \"3b5b4c6f-368c-47ba-87fc-7309cac15d9b\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.870129 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-utilities" (OuterVolumeSpecName: "utilities") pod "98f32d89-a132-4cb9-a647-b48848e1b18a" (UID: "98f32d89-a132-4cb9-a647-b48848e1b18a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.870922 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-utilities" (OuterVolumeSpecName: "utilities") pod "3b5b4c6f-368c-47ba-87fc-7309cac15d9b" (UID: "3b5b4c6f-368c-47ba-87fc-7309cac15d9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.871913 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14da97a9-4b68-458d-858a-6ba8b67cb749-kube-api-access-b4rgz" (OuterVolumeSpecName: "kube-api-access-b4rgz") pod "14da97a9-4b68-458d-858a-6ba8b67cb749" (UID: "14da97a9-4b68-458d-858a-6ba8b67cb749"). InnerVolumeSpecName "kube-api-access-b4rgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.872812 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-utilities" (OuterVolumeSpecName: "utilities") pod "14da97a9-4b68-458d-858a-6ba8b67cb749" (UID: "14da97a9-4b68-458d-858a-6ba8b67cb749"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.876415 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98f32d89-a132-4cb9-a647-b48848e1b18a-kube-api-access-s8rtj" (OuterVolumeSpecName: "kube-api-access-s8rtj") pod "98f32d89-a132-4cb9-a647-b48848e1b18a" (UID: "98f32d89-a132-4cb9-a647-b48848e1b18a"). InnerVolumeSpecName "kube-api-access-s8rtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.880889 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-kube-api-access-pt56f" (OuterVolumeSpecName: "kube-api-access-pt56f") pod "3b5b4c6f-368c-47ba-87fc-7309cac15d9b" (UID: "3b5b4c6f-368c-47ba-87fc-7309cac15d9b"). InnerVolumeSpecName "kube-api-access-pt56f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.935172 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14da97a9-4b68-458d-858a-6ba8b67cb749" (UID: "14da97a9-4b68-458d-858a-6ba8b67cb749"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.963042 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b5b4c6f-368c-47ba-87fc-7309cac15d9b" (UID: "3b5b4c6f-368c-47ba-87fc-7309cac15d9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.971484 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-trusted-ca\") pod \"34dedc1d-10d1-4ca0-b059-99048ba37464\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.971601 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-operator-metrics\") pod \"34dedc1d-10d1-4ca0-b059-99048ba37464\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.971644 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m795\" (UniqueName: \"kubernetes.io/projected/34dedc1d-10d1-4ca0-b059-99048ba37464-kube-api-access-6m795\") pod \"34dedc1d-10d1-4ca0-b059-99048ba37464\" (UID: \"34dedc1d-10d1-4ca0-b059-99048ba37464\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.971669 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-utilities\") pod \"64a47758-13e5-4a28-b781-2bdf84b870fd\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.971696 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-catalog-content\") pod \"64a47758-13e5-4a28-b781-2bdf84b870fd\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.971717 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhzjh\" (UniqueName: \"kubernetes.io/projected/64a47758-13e5-4a28-b781-2bdf84b870fd-kube-api-access-lhzjh\") pod \"64a47758-13e5-4a28-b781-2bdf84b870fd\" (UID: \"64a47758-13e5-4a28-b781-2bdf84b870fd\") " Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.971980 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.971996 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt56f\" (UniqueName: \"kubernetes.io/projected/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-kube-api-access-pt56f\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.972010 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.972021 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.972031 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4rgz\" (UniqueName: \"kubernetes.io/projected/14da97a9-4b68-458d-858a-6ba8b67cb749-kube-api-access-b4rgz\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.972042 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14da97a9-4b68-458d-858a-6ba8b67cb749-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.972052 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5b4c6f-368c-47ba-87fc-7309cac15d9b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.972065 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8rtj\" (UniqueName: \"kubernetes.io/projected/98f32d89-a132-4cb9-a647-b48848e1b18a-kube-api-access-s8rtj\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.973364 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "34dedc1d-10d1-4ca0-b059-99048ba37464" (UID: "34dedc1d-10d1-4ca0-b059-99048ba37464"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.973415 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-utilities" (OuterVolumeSpecName: "utilities") pod "64a47758-13e5-4a28-b781-2bdf84b870fd" (UID: "64a47758-13e5-4a28-b781-2bdf84b870fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.976166 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "34dedc1d-10d1-4ca0-b059-99048ba37464" (UID: "34dedc1d-10d1-4ca0-b059-99048ba37464"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.979197 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64a47758-13e5-4a28-b781-2bdf84b870fd-kube-api-access-lhzjh" (OuterVolumeSpecName: "kube-api-access-lhzjh") pod "64a47758-13e5-4a28-b781-2bdf84b870fd" (UID: "64a47758-13e5-4a28-b781-2bdf84b870fd"). InnerVolumeSpecName "kube-api-access-lhzjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.980797 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98f32d89-a132-4cb9-a647-b48848e1b18a" (UID: "98f32d89-a132-4cb9-a647-b48848e1b18a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:16:16 crc kubenswrapper[4846]: I0202 12:16:16.982841 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34dedc1d-10d1-4ca0-b059-99048ba37464-kube-api-access-6m795" (OuterVolumeSpecName: "kube-api-access-6m795") pod "34dedc1d-10d1-4ca0-b059-99048ba37464" (UID: "34dedc1d-10d1-4ca0-b059-99048ba37464"). InnerVolumeSpecName "kube-api-access-6m795". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.039548 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kgbzq"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.073933 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhzjh\" (UniqueName: \"kubernetes.io/projected/64a47758-13e5-4a28-b781-2bdf84b870fd-kube-api-access-lhzjh\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.073968 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.073983 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98f32d89-a132-4cb9-a647-b48848e1b18a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.073995 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/34dedc1d-10d1-4ca0-b059-99048ba37464-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.074008 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m795\" (UniqueName: \"kubernetes.io/projected/34dedc1d-10d1-4ca0-b059-99048ba37464-kube-api-access-6m795\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.074020 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.128223 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64a47758-13e5-4a28-b781-2bdf84b870fd" (UID: "64a47758-13e5-4a28-b781-2bdf84b870fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.176270 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64a47758-13e5-4a28-b781-2bdf84b870fd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.631004 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pjvpk" event={"ID":"3b5b4c6f-368c-47ba-87fc-7309cac15d9b","Type":"ContainerDied","Data":"4f7f72c8ef672a5c5bd6ac0baa3bd80e9c60b0baad71e44681d854424fda1175"} Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.631035 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pjvpk" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.631069 4846 scope.go:117] "RemoveContainer" containerID="8e8b83ecde2490950dc76da4e5f92d61cce616b29e52fd40b0bee1afae697296" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.637395 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmnp2" event={"ID":"98f32d89-a132-4cb9-a647-b48848e1b18a","Type":"ContainerDied","Data":"69be1edd328643017aecd7b67b5a8ff66fe81d623b7c17fd80725a345df3cab3"} Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.637523 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmnp2" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.641023 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n44pv" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.641115 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n44pv" event={"ID":"64a47758-13e5-4a28-b781-2bdf84b870fd","Type":"ContainerDied","Data":"7f8e6c51540063226181c7b4194d87405b772d3931aa495b8397174a4e8c5171"} Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.642863 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" event={"ID":"17938097-4a7b-4e18-9e26-4ada5f07b343","Type":"ContainerStarted","Data":"a236b6b7650a164a39605ec41d843d42accf8fe3fa9f91bc9c20bd60c1f03580"} Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.642901 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" event={"ID":"17938097-4a7b-4e18-9e26-4ada5f07b343","Type":"ContainerStarted","Data":"afd550643af8eb1380affed02d645c976bd44f77e38b6d09c6ddb9d611cd2514"} Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.643468 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.645779 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.656768 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pjvpk"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.660513 4846 scope.go:117] "RemoveContainer" containerID="fb4e7a33b3bfc97dda3a67d3a4c7636d61632d03c7fe43802099e873bc821cdb" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.662658 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p558w" event={"ID":"14da97a9-4b68-458d-858a-6ba8b67cb749","Type":"ContainerDied","Data":"a6eb9a138cf55764de1184b906468c972c1de9aa26bd531c2e7a8869bd60da72"} Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.662675 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p558w" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.666454 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" event={"ID":"34dedc1d-10d1-4ca0-b059-99048ba37464","Type":"ContainerDied","Data":"0bf34f2ea3c6d0ba73912a7216abc2920b974c47d3be95e3e3b3bc61a2b205fc"} Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.666565 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zbpnp" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.668854 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pjvpk"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.674828 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gmnp2"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.682690 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gmnp2"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.691779 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kgbzq" podStartSLOduration=1.691761394 podStartE2EDuration="1.691761394s" podCreationTimestamp="2026-02-02 12:16:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:16:17.679635413 +0000 UTC m=+408.908222306" watchObservedRunningTime="2026-02-02 12:16:17.691761394 +0000 UTC m=+408.920348257" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.694899 4846 scope.go:117] "RemoveContainer" containerID="ca650255f3ade6088a95b4752a6cb7eddcfd18267a077a6424808a5e59c670ae" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.706551 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n44pv"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.712270 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n44pv"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.715563 4846 scope.go:117] "RemoveContainer" containerID="b49bfd2884c82785d674898ef7d77801330d101aaffeef0dc55a7dfd9c854c56" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.717001 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p558w"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.721517 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p558w"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.727265 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zbpnp"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.727305 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zbpnp"] Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.730927 4846 scope.go:117] "RemoveContainer" containerID="56c2215aa72b34204e60b6dc026cd1ed208e3a787331105187d6bb32f67f38ad" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.744802 4846 scope.go:117] "RemoveContainer" containerID="df9511754728df7f993f5220d3ef474477a76645511eb862a22a87ad404493d6" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.762151 4846 scope.go:117] "RemoveContainer" containerID="ba291467c7fb9876c4109dfe1171cb8179f89c87fdc1e8aa3cbe5780f0dddf04" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.781984 4846 scope.go:117] "RemoveContainer" containerID="e4ea05ba5acc3ccfe68a216c877a5b7c4aa682276bbc4a4999ebd856da4b3e89" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.799862 4846 scope.go:117] "RemoveContainer" containerID="241687a3f8c362a669e31bf5a7d8ee6c5ad23aed00c529b123b4627ade5737a7" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.819451 4846 scope.go:117] "RemoveContainer" containerID="a174332179e0dd7320d061b3d003c07def4344aad74ef063a34205248ea67d06" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.829762 4846 scope.go:117] "RemoveContainer" containerID="72dca24d234e79bc7015f8a7c47a3a2eca66dd8e2535d263cb8d1adb482a4f8b" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.840754 4846 scope.go:117] "RemoveContainer" containerID="b9cc2ecbfff30ad71954731617456c9ecef7ae873e3e59fc57abdc4504923c5d" Feb 02 12:16:17 crc kubenswrapper[4846]: I0202 12:16:17.852744 4846 scope.go:117] "RemoveContainer" containerID="ceeb99133435ec0fb619d2010f5a3b2136022c9011c8ee1f62129edd65db609a" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.506892 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r7jc8"] Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.507899 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.507998 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.508072 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerName="extract-utilities" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.508157 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerName="extract-utilities" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.508231 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.508301 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.508366 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerName="extract-content" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.508561 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerName="extract-content" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.508645 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerName="extract-content" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.508712 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerName="extract-content" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.508770 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerName="extract-content" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.508824 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerName="extract-content" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.508874 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34dedc1d-10d1-4ca0-b059-99048ba37464" containerName="marketplace-operator" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.508932 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="34dedc1d-10d1-4ca0-b059-99048ba37464" containerName="marketplace-operator" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.509022 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerName="extract-utilities" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.509108 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerName="extract-utilities" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.509171 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerName="extract-content" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.509222 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerName="extract-content" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.509278 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerName="extract-utilities" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.509329 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerName="extract-utilities" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.509383 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.509438 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.509623 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerName="extract-utilities" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.509727 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerName="extract-utilities" Feb 02 12:16:18 crc kubenswrapper[4846]: E0202 12:16:18.509809 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.509885 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.510125 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="14da97a9-4b68-458d-858a-6ba8b67cb749" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.510254 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="98f32d89-a132-4cb9-a647-b48848e1b18a" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.510345 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="34dedc1d-10d1-4ca0-b059-99048ba37464" containerName="marketplace-operator" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.510431 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.510557 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" containerName="registry-server" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.511867 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.515562 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.523175 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7jc8"] Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.695805 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf-catalog-content\") pod \"redhat-marketplace-r7jc8\" (UID: \"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf\") " pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.695930 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpj67\" (UniqueName: \"kubernetes.io/projected/de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf-kube-api-access-lpj67\") pod \"redhat-marketplace-r7jc8\" (UID: \"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf\") " pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.695964 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf-utilities\") pod \"redhat-marketplace-r7jc8\" (UID: \"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf\") " pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.706874 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7hsfb"] Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.707909 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.709825 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.716181 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7hsfb"] Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.797527 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpj67\" (UniqueName: \"kubernetes.io/projected/de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf-kube-api-access-lpj67\") pod \"redhat-marketplace-r7jc8\" (UID: \"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf\") " pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.797590 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf-utilities\") pod \"redhat-marketplace-r7jc8\" (UID: \"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf\") " pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.797673 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-catalog-content\") pod \"redhat-operators-7hsfb\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.797739 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf-catalog-content\") pod \"redhat-marketplace-r7jc8\" (UID: \"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf\") " pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.797785 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56fw9\" (UniqueName: \"kubernetes.io/projected/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-kube-api-access-56fw9\") pod \"redhat-operators-7hsfb\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.798288 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf-catalog-content\") pod \"redhat-marketplace-r7jc8\" (UID: \"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf\") " pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.798535 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-utilities\") pod \"redhat-operators-7hsfb\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.799261 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf-utilities\") pod \"redhat-marketplace-r7jc8\" (UID: \"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf\") " pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.819869 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpj67\" (UniqueName: \"kubernetes.io/projected/de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf-kube-api-access-lpj67\") pod \"redhat-marketplace-r7jc8\" (UID: \"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf\") " pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.833441 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.899899 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-catalog-content\") pod \"redhat-operators-7hsfb\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.900328 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56fw9\" (UniqueName: \"kubernetes.io/projected/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-kube-api-access-56fw9\") pod \"redhat-operators-7hsfb\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.900362 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-utilities\") pod \"redhat-operators-7hsfb\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.900796 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-utilities\") pod \"redhat-operators-7hsfb\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.900993 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-catalog-content\") pod \"redhat-operators-7hsfb\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:18 crc kubenswrapper[4846]: I0202 12:16:18.920742 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56fw9\" (UniqueName: \"kubernetes.io/projected/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-kube-api-access-56fw9\") pod \"redhat-operators-7hsfb\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.023227 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7jc8"] Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.033933 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.443486 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14da97a9-4b68-458d-858a-6ba8b67cb749" path="/var/lib/kubelet/pods/14da97a9-4b68-458d-858a-6ba8b67cb749/volumes" Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.445468 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34dedc1d-10d1-4ca0-b059-99048ba37464" path="/var/lib/kubelet/pods/34dedc1d-10d1-4ca0-b059-99048ba37464/volumes" Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.446711 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b5b4c6f-368c-47ba-87fc-7309cac15d9b" path="/var/lib/kubelet/pods/3b5b4c6f-368c-47ba-87fc-7309cac15d9b/volumes" Feb 02 12:16:19 crc kubenswrapper[4846]: W0202 12:16:19.446931 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fe8aed5_574d_47f9_8822_f7a42dbccb5a.slice/crio-868509b99a7e621fcc156f8b60c01efa5454728c7a23c49018ae95adfa6c982a WatchSource:0}: Error finding container 868509b99a7e621fcc156f8b60c01efa5454728c7a23c49018ae95adfa6c982a: Status 404 returned error can't find the container with id 868509b99a7e621fcc156f8b60c01efa5454728c7a23c49018ae95adfa6c982a Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.449120 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64a47758-13e5-4a28-b781-2bdf84b870fd" path="/var/lib/kubelet/pods/64a47758-13e5-4a28-b781-2bdf84b870fd/volumes" Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.453220 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98f32d89-a132-4cb9-a647-b48848e1b18a" path="/var/lib/kubelet/pods/98f32d89-a132-4cb9-a647-b48848e1b18a/volumes" Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.454974 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7hsfb"] Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.681550 4846 generic.go:334] "Generic (PLEG): container finished" podID="de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf" containerID="1c686331c8126ab184042abd0d7f984429525cd61d620dd507322b45e721ee10" exitCode=0 Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.681665 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7jc8" event={"ID":"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf","Type":"ContainerDied","Data":"1c686331c8126ab184042abd0d7f984429525cd61d620dd507322b45e721ee10"} Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.681712 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7jc8" event={"ID":"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf","Type":"ContainerStarted","Data":"6c6202acfd8c521a28aa07ffbe90f7f40116033c9fe7e25cace52f367ca5f9f1"} Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.683002 4846 generic.go:334] "Generic (PLEG): container finished" podID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerID="68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc" exitCode=0 Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.683723 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hsfb" event={"ID":"2fe8aed5-574d-47f9-8822-f7a42dbccb5a","Type":"ContainerDied","Data":"68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc"} Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.683766 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hsfb" event={"ID":"2fe8aed5-574d-47f9-8822-f7a42dbccb5a","Type":"ContainerStarted","Data":"868509b99a7e621fcc156f8b60c01efa5454728c7a23c49018ae95adfa6c982a"} Feb 02 12:16:19 crc kubenswrapper[4846]: I0202 12:16:19.857258 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" podUID="59df79f3-cb9f-412c-bc01-7f3376620c17" containerName="registry" containerID="cri-o://b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee" gracePeriod=30 Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.196677 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.319223 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/59df79f3-cb9f-412c-bc01-7f3376620c17-installation-pull-secrets\") pod \"59df79f3-cb9f-412c-bc01-7f3376620c17\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.319297 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-certificates\") pod \"59df79f3-cb9f-412c-bc01-7f3376620c17\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.319325 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-trusted-ca\") pod \"59df79f3-cb9f-412c-bc01-7f3376620c17\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.319357 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-td668\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-kube-api-access-td668\") pod \"59df79f3-cb9f-412c-bc01-7f3376620c17\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.319409 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-bound-sa-token\") pod \"59df79f3-cb9f-412c-bc01-7f3376620c17\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.319441 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/59df79f3-cb9f-412c-bc01-7f3376620c17-ca-trust-extracted\") pod \"59df79f3-cb9f-412c-bc01-7f3376620c17\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.319476 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-tls\") pod \"59df79f3-cb9f-412c-bc01-7f3376620c17\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.319604 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"59df79f3-cb9f-412c-bc01-7f3376620c17\" (UID: \"59df79f3-cb9f-412c-bc01-7f3376620c17\") " Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.320321 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "59df79f3-cb9f-412c-bc01-7f3376620c17" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.320389 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "59df79f3-cb9f-412c-bc01-7f3376620c17" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.325167 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59df79f3-cb9f-412c-bc01-7f3376620c17-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "59df79f3-cb9f-412c-bc01-7f3376620c17" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.325376 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "59df79f3-cb9f-412c-bc01-7f3376620c17" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.325605 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-kube-api-access-td668" (OuterVolumeSpecName: "kube-api-access-td668") pod "59df79f3-cb9f-412c-bc01-7f3376620c17" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17"). InnerVolumeSpecName "kube-api-access-td668". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.325832 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "59df79f3-cb9f-412c-bc01-7f3376620c17" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.333059 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "59df79f3-cb9f-412c-bc01-7f3376620c17" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.336931 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59df79f3-cb9f-412c-bc01-7f3376620c17-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "59df79f3-cb9f-412c-bc01-7f3376620c17" (UID: "59df79f3-cb9f-412c-bc01-7f3376620c17"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.421129 4846 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/59df79f3-cb9f-412c-bc01-7f3376620c17-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.421170 4846 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.421182 4846 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/59df79f3-cb9f-412c-bc01-7f3376620c17-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.421194 4846 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.421203 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/59df79f3-cb9f-412c-bc01-7f3376620c17-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.421212 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-td668\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-kube-api-access-td668\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.421220 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59df79f3-cb9f-412c-bc01-7f3376620c17-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.699241 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7jc8" event={"ID":"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf","Type":"ContainerStarted","Data":"cd1ad1b5d23f53b781d6d4dbd6b9200cbd40c58d35e51ed6d94ce14d14fbf9eb"} Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.701251 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hsfb" event={"ID":"2fe8aed5-574d-47f9-8822-f7a42dbccb5a","Type":"ContainerStarted","Data":"db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee"} Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.703683 4846 generic.go:334] "Generic (PLEG): container finished" podID="59df79f3-cb9f-412c-bc01-7f3376620c17" containerID="b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee" exitCode=0 Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.703723 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" event={"ID":"59df79f3-cb9f-412c-bc01-7f3376620c17","Type":"ContainerDied","Data":"b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee"} Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.703745 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" event={"ID":"59df79f3-cb9f-412c-bc01-7f3376620c17","Type":"ContainerDied","Data":"a0e06192f0cc45669a0c281bbf5530f2d20d6fef56ffc566f04d4af86cd85757"} Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.703768 4846 scope.go:117] "RemoveContainer" containerID="b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.703881 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-94n5d" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.727843 4846 scope.go:117] "RemoveContainer" containerID="b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee" Feb 02 12:16:20 crc kubenswrapper[4846]: E0202 12:16:20.728318 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee\": container with ID starting with b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee not found: ID does not exist" containerID="b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.728444 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee"} err="failed to get container status \"b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee\": rpc error: code = NotFound desc = could not find container \"b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee\": container with ID starting with b220e87fbba2841b6abb824769f627e648f7780b3a74dfb61fe6e3f9c15d67ee not found: ID does not exist" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.755513 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94n5d"] Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.760719 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-94n5d"] Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.903670 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vjszt"] Feb 02 12:16:20 crc kubenswrapper[4846]: E0202 12:16:20.904008 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59df79f3-cb9f-412c-bc01-7f3376620c17" containerName="registry" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.904023 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="59df79f3-cb9f-412c-bc01-7f3376620c17" containerName="registry" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.904159 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="59df79f3-cb9f-412c-bc01-7f3376620c17" containerName="registry" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.905163 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.907867 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 02 12:16:20 crc kubenswrapper[4846]: I0202 12:16:20.924016 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vjszt"] Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.030679 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfx7p\" (UniqueName: \"kubernetes.io/projected/5e7333ab-9306-45aa-a3aa-2e637120e4f9-kube-api-access-nfx7p\") pod \"certified-operators-vjszt\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.030755 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-utilities\") pod \"certified-operators-vjszt\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.030828 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-catalog-content\") pod \"certified-operators-vjszt\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.108587 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dwjk4"] Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.110206 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.113362 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.117390 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dwjk4"] Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.131548 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-catalog-content\") pod \"certified-operators-vjszt\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.131612 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfx7p\" (UniqueName: \"kubernetes.io/projected/5e7333ab-9306-45aa-a3aa-2e637120e4f9-kube-api-access-nfx7p\") pod \"certified-operators-vjszt\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.131679 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-utilities\") pod \"certified-operators-vjszt\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.132055 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-catalog-content\") pod \"certified-operators-vjszt\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.132184 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-utilities\") pod \"certified-operators-vjszt\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.152779 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfx7p\" (UniqueName: \"kubernetes.io/projected/5e7333ab-9306-45aa-a3aa-2e637120e4f9-kube-api-access-nfx7p\") pod \"certified-operators-vjszt\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.229636 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.232699 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn8wv\" (UniqueName: \"kubernetes.io/projected/5ec82720-fb76-468d-88d2-682ab80aab08-kube-api-access-zn8wv\") pod \"community-operators-dwjk4\" (UID: \"5ec82720-fb76-468d-88d2-682ab80aab08\") " pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.232759 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ec82720-fb76-468d-88d2-682ab80aab08-catalog-content\") pod \"community-operators-dwjk4\" (UID: \"5ec82720-fb76-468d-88d2-682ab80aab08\") " pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.232927 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ec82720-fb76-468d-88d2-682ab80aab08-utilities\") pod \"community-operators-dwjk4\" (UID: \"5ec82720-fb76-468d-88d2-682ab80aab08\") " pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.334896 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn8wv\" (UniqueName: \"kubernetes.io/projected/5ec82720-fb76-468d-88d2-682ab80aab08-kube-api-access-zn8wv\") pod \"community-operators-dwjk4\" (UID: \"5ec82720-fb76-468d-88d2-682ab80aab08\") " pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.335196 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ec82720-fb76-468d-88d2-682ab80aab08-catalog-content\") pod \"community-operators-dwjk4\" (UID: \"5ec82720-fb76-468d-88d2-682ab80aab08\") " pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.335253 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ec82720-fb76-468d-88d2-682ab80aab08-utilities\") pod \"community-operators-dwjk4\" (UID: \"5ec82720-fb76-468d-88d2-682ab80aab08\") " pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.335990 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ec82720-fb76-468d-88d2-682ab80aab08-utilities\") pod \"community-operators-dwjk4\" (UID: \"5ec82720-fb76-468d-88d2-682ab80aab08\") " pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.336135 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ec82720-fb76-468d-88d2-682ab80aab08-catalog-content\") pod \"community-operators-dwjk4\" (UID: \"5ec82720-fb76-468d-88d2-682ab80aab08\") " pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.355268 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn8wv\" (UniqueName: \"kubernetes.io/projected/5ec82720-fb76-468d-88d2-682ab80aab08-kube-api-access-zn8wv\") pod \"community-operators-dwjk4\" (UID: \"5ec82720-fb76-468d-88d2-682ab80aab08\") " pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.403325 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vjszt"] Feb 02 12:16:21 crc kubenswrapper[4846]: W0202 12:16:21.412497 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e7333ab_9306_45aa_a3aa_2e637120e4f9.slice/crio-3e698eda7cf2d210e118f4a7d054964de326997059739704953aa14b8e4e46ca WatchSource:0}: Error finding container 3e698eda7cf2d210e118f4a7d054964de326997059739704953aa14b8e4e46ca: Status 404 returned error can't find the container with id 3e698eda7cf2d210e118f4a7d054964de326997059739704953aa14b8e4e46ca Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.425035 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.429415 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59df79f3-cb9f-412c-bc01-7f3376620c17" path="/var/lib/kubelet/pods/59df79f3-cb9f-412c-bc01-7f3376620c17/volumes" Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.591263 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dwjk4"] Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.710010 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerID="2366afb59437b127afdd2301b249d90f1061f32bcee73e6893dfddfb5ff3e5d3" exitCode=0 Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.710119 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vjszt" event={"ID":"5e7333ab-9306-45aa-a3aa-2e637120e4f9","Type":"ContainerDied","Data":"2366afb59437b127afdd2301b249d90f1061f32bcee73e6893dfddfb5ff3e5d3"} Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.710687 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vjszt" event={"ID":"5e7333ab-9306-45aa-a3aa-2e637120e4f9","Type":"ContainerStarted","Data":"3e698eda7cf2d210e118f4a7d054964de326997059739704953aa14b8e4e46ca"} Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.716182 4846 generic.go:334] "Generic (PLEG): container finished" podID="de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf" containerID="cd1ad1b5d23f53b781d6d4dbd6b9200cbd40c58d35e51ed6d94ce14d14fbf9eb" exitCode=0 Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.716234 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7jc8" event={"ID":"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf","Type":"ContainerDied","Data":"cd1ad1b5d23f53b781d6d4dbd6b9200cbd40c58d35e51ed6d94ce14d14fbf9eb"} Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.722836 4846 generic.go:334] "Generic (PLEG): container finished" podID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerID="db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee" exitCode=0 Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.722897 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hsfb" event={"ID":"2fe8aed5-574d-47f9-8822-f7a42dbccb5a","Type":"ContainerDied","Data":"db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee"} Feb 02 12:16:21 crc kubenswrapper[4846]: I0202 12:16:21.729322 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwjk4" event={"ID":"5ec82720-fb76-468d-88d2-682ab80aab08","Type":"ContainerStarted","Data":"360a8ec6030562aa8841472ba5d54c2f84b918dfb11a2a46f1fee7797a25528a"} Feb 02 12:16:22 crc kubenswrapper[4846]: I0202 12:16:22.736552 4846 generic.go:334] "Generic (PLEG): container finished" podID="5ec82720-fb76-468d-88d2-682ab80aab08" containerID="89e1a5a0ff662520904cddb1693be0f4b047b1cdfffd0fe8778d51ba117091be" exitCode=0 Feb 02 12:16:22 crc kubenswrapper[4846]: I0202 12:16:22.736614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwjk4" event={"ID":"5ec82720-fb76-468d-88d2-682ab80aab08","Type":"ContainerDied","Data":"89e1a5a0ff662520904cddb1693be0f4b047b1cdfffd0fe8778d51ba117091be"} Feb 02 12:16:22 crc kubenswrapper[4846]: I0202 12:16:22.738921 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vjszt" event={"ID":"5e7333ab-9306-45aa-a3aa-2e637120e4f9","Type":"ContainerStarted","Data":"caa9fff96d26773cfc441e827143ea5faa16713881df451cd94a4badac58653a"} Feb 02 12:16:22 crc kubenswrapper[4846]: I0202 12:16:22.741208 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7jc8" event={"ID":"de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf","Type":"ContainerStarted","Data":"88a9ede37f5dc36b1481461db68be19ce2be44db6eab77a7ba290267f44e40b9"} Feb 02 12:16:22 crc kubenswrapper[4846]: I0202 12:16:22.745165 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hsfb" event={"ID":"2fe8aed5-574d-47f9-8822-f7a42dbccb5a","Type":"ContainerStarted","Data":"3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a"} Feb 02 12:16:22 crc kubenswrapper[4846]: I0202 12:16:22.782381 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r7jc8" podStartSLOduration=2.207303626 podStartE2EDuration="4.782360842s" podCreationTimestamp="2026-02-02 12:16:18 +0000 UTC" firstStartedPulling="2026-02-02 12:16:19.682870746 +0000 UTC m=+410.911457609" lastFinishedPulling="2026-02-02 12:16:22.257927962 +0000 UTC m=+413.486514825" observedRunningTime="2026-02-02 12:16:22.782057475 +0000 UTC m=+414.010644338" watchObservedRunningTime="2026-02-02 12:16:22.782360842 +0000 UTC m=+414.010947705" Feb 02 12:16:22 crc kubenswrapper[4846]: I0202 12:16:22.827967 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7hsfb" podStartSLOduration=2.282599981 podStartE2EDuration="4.82794911s" podCreationTimestamp="2026-02-02 12:16:18 +0000 UTC" firstStartedPulling="2026-02-02 12:16:19.684845785 +0000 UTC m=+410.913432648" lastFinishedPulling="2026-02-02 12:16:22.230194924 +0000 UTC m=+413.458781777" observedRunningTime="2026-02-02 12:16:22.824875145 +0000 UTC m=+414.053462008" watchObservedRunningTime="2026-02-02 12:16:22.82794911 +0000 UTC m=+414.056535973" Feb 02 12:16:23 crc kubenswrapper[4846]: I0202 12:16:23.752518 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwjk4" event={"ID":"5ec82720-fb76-468d-88d2-682ab80aab08","Type":"ContainerStarted","Data":"4e08e065a39f8aedba9499f94325b8bc63f425d731981fe5016d0df3a8be6f45"} Feb 02 12:16:23 crc kubenswrapper[4846]: I0202 12:16:23.755245 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vjszt" event={"ID":"5e7333ab-9306-45aa-a3aa-2e637120e4f9","Type":"ContainerDied","Data":"caa9fff96d26773cfc441e827143ea5faa16713881df451cd94a4badac58653a"} Feb 02 12:16:23 crc kubenswrapper[4846]: I0202 12:16:23.755255 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerID="caa9fff96d26773cfc441e827143ea5faa16713881df451cd94a4badac58653a" exitCode=0 Feb 02 12:16:24 crc kubenswrapper[4846]: I0202 12:16:24.763626 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vjszt" event={"ID":"5e7333ab-9306-45aa-a3aa-2e637120e4f9","Type":"ContainerStarted","Data":"3f97a13e10729c233dc9f25bffab25cd76b892ca2736cf199e8382bae879cb0e"} Feb 02 12:16:24 crc kubenswrapper[4846]: I0202 12:16:24.765664 4846 generic.go:334] "Generic (PLEG): container finished" podID="5ec82720-fb76-468d-88d2-682ab80aab08" containerID="4e08e065a39f8aedba9499f94325b8bc63f425d731981fe5016d0df3a8be6f45" exitCode=0 Feb 02 12:16:24 crc kubenswrapper[4846]: I0202 12:16:24.765702 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwjk4" event={"ID":"5ec82720-fb76-468d-88d2-682ab80aab08","Type":"ContainerDied","Data":"4e08e065a39f8aedba9499f94325b8bc63f425d731981fe5016d0df3a8be6f45"} Feb 02 12:16:24 crc kubenswrapper[4846]: I0202 12:16:24.782216 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vjszt" podStartSLOduration=2.342875277 podStartE2EDuration="4.782194303s" podCreationTimestamp="2026-02-02 12:16:20 +0000 UTC" firstStartedPulling="2026-02-02 12:16:21.71143973 +0000 UTC m=+412.940026593" lastFinishedPulling="2026-02-02 12:16:24.150758756 +0000 UTC m=+415.379345619" observedRunningTime="2026-02-02 12:16:24.781242651 +0000 UTC m=+416.009829514" watchObservedRunningTime="2026-02-02 12:16:24.782194303 +0000 UTC m=+416.010781176" Feb 02 12:16:25 crc kubenswrapper[4846]: I0202 12:16:25.773429 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwjk4" event={"ID":"5ec82720-fb76-468d-88d2-682ab80aab08","Type":"ContainerStarted","Data":"41e6a70995412281238fecaa20563fae86701f3ce2982064f2aeba4ff48e5d92"} Feb 02 12:16:25 crc kubenswrapper[4846]: I0202 12:16:25.794340 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dwjk4" podStartSLOduration=2.289799082 podStartE2EDuration="4.794325098s" podCreationTimestamp="2026-02-02 12:16:21 +0000 UTC" firstStartedPulling="2026-02-02 12:16:22.737729157 +0000 UTC m=+413.966316010" lastFinishedPulling="2026-02-02 12:16:25.242255163 +0000 UTC m=+416.470842026" observedRunningTime="2026-02-02 12:16:25.792351161 +0000 UTC m=+417.020938014" watchObservedRunningTime="2026-02-02 12:16:25.794325098 +0000 UTC m=+417.022911961" Feb 02 12:16:28 crc kubenswrapper[4846]: I0202 12:16:28.834121 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:28 crc kubenswrapper[4846]: I0202 12:16:28.834693 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:28 crc kubenswrapper[4846]: I0202 12:16:28.881696 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:29 crc kubenswrapper[4846]: I0202 12:16:29.034573 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:29 crc kubenswrapper[4846]: I0202 12:16:29.034868 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:29 crc kubenswrapper[4846]: I0202 12:16:29.075292 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:29 crc kubenswrapper[4846]: I0202 12:16:29.837979 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r7jc8" Feb 02 12:16:29 crc kubenswrapper[4846]: I0202 12:16:29.842239 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 12:16:31 crc kubenswrapper[4846]: I0202 12:16:31.230534 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:31 crc kubenswrapper[4846]: I0202 12:16:31.231557 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:31 crc kubenswrapper[4846]: I0202 12:16:31.266793 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:31 crc kubenswrapper[4846]: I0202 12:16:31.430390 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:31 crc kubenswrapper[4846]: I0202 12:16:31.430432 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:31 crc kubenswrapper[4846]: I0202 12:16:31.464138 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:16:31 crc kubenswrapper[4846]: I0202 12:16:31.839648 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vjszt" Feb 02 12:16:31 crc kubenswrapper[4846]: I0202 12:16:31.840044 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dwjk4" Feb 02 12:18:00 crc kubenswrapper[4846]: I0202 12:18:00.479268 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:18:00 crc kubenswrapper[4846]: I0202 12:18:00.479938 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:18:30 crc kubenswrapper[4846]: I0202 12:18:30.479331 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:18:30 crc kubenswrapper[4846]: I0202 12:18:30.479997 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:19:00 crc kubenswrapper[4846]: I0202 12:19:00.479077 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:19:00 crc kubenswrapper[4846]: I0202 12:19:00.480509 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:19:00 crc kubenswrapper[4846]: I0202 12:19:00.480656 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:19:00 crc kubenswrapper[4846]: I0202 12:19:00.481150 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a4395520eea0717a00d2525b7efc36d173bd70de14fcc8d8b3411afe2fc41cae"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:19:00 crc kubenswrapper[4846]: I0202 12:19:00.481288 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://a4395520eea0717a00d2525b7efc36d173bd70de14fcc8d8b3411afe2fc41cae" gracePeriod=600 Feb 02 12:19:01 crc kubenswrapper[4846]: I0202 12:19:01.602373 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="a4395520eea0717a00d2525b7efc36d173bd70de14fcc8d8b3411afe2fc41cae" exitCode=0 Feb 02 12:19:01 crc kubenswrapper[4846]: I0202 12:19:01.602467 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"a4395520eea0717a00d2525b7efc36d173bd70de14fcc8d8b3411afe2fc41cae"} Feb 02 12:19:01 crc kubenswrapper[4846]: I0202 12:19:01.602980 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"025493c8107e9680bfbd487f264aead8432b3b590da7fb314c8c0774a547036f"} Feb 02 12:19:01 crc kubenswrapper[4846]: I0202 12:19:01.603004 4846 scope.go:117] "RemoveContainer" containerID="e960baa64942401b64faf1ec81d39bc736adab9ff74905ed38161bca5093c8f7" Feb 02 12:21:00 crc kubenswrapper[4846]: I0202 12:21:00.478866 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:21:00 crc kubenswrapper[4846]: I0202 12:21:00.479408 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:21:30 crc kubenswrapper[4846]: I0202 12:21:30.478765 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:21:30 crc kubenswrapper[4846]: I0202 12:21:30.479272 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:22:00 crc kubenswrapper[4846]: I0202 12:22:00.479821 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:22:00 crc kubenswrapper[4846]: I0202 12:22:00.480410 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:22:00 crc kubenswrapper[4846]: I0202 12:22:00.480459 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:22:00 crc kubenswrapper[4846]: I0202 12:22:00.481140 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"025493c8107e9680bfbd487f264aead8432b3b590da7fb314c8c0774a547036f"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:22:00 crc kubenswrapper[4846]: I0202 12:22:00.481266 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://025493c8107e9680bfbd487f264aead8432b3b590da7fb314c8c0774a547036f" gracePeriod=600 Feb 02 12:22:01 crc kubenswrapper[4846]: I0202 12:22:01.554767 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="025493c8107e9680bfbd487f264aead8432b3b590da7fb314c8c0774a547036f" exitCode=0 Feb 02 12:22:01 crc kubenswrapper[4846]: I0202 12:22:01.554861 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"025493c8107e9680bfbd487f264aead8432b3b590da7fb314c8c0774a547036f"} Feb 02 12:22:01 crc kubenswrapper[4846]: I0202 12:22:01.555409 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"06b5f4ddd20a5b1b0ef8c848baafd57c963bd52aab6eccb2c47d76fc10efc10c"} Feb 02 12:22:01 crc kubenswrapper[4846]: I0202 12:22:01.555432 4846 scope.go:117] "RemoveContainer" containerID="a4395520eea0717a00d2525b7efc36d173bd70de14fcc8d8b3411afe2fc41cae" Feb 02 12:22:14 crc kubenswrapper[4846]: I0202 12:22:14.615847 4846 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 02 12:22:27 crc kubenswrapper[4846]: I0202 12:22:27.914938 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ftvcw"] Feb 02 12:22:27 crc kubenswrapper[4846]: I0202 12:22:27.915874 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovn-controller" containerID="cri-o://535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c" gracePeriod=30 Feb 02 12:22:27 crc kubenswrapper[4846]: I0202 12:22:27.916015 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="sbdb" containerID="cri-o://efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78" gracePeriod=30 Feb 02 12:22:27 crc kubenswrapper[4846]: I0202 12:22:27.916047 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="nbdb" containerID="cri-o://3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3" gracePeriod=30 Feb 02 12:22:27 crc kubenswrapper[4846]: I0202 12:22:27.916077 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="northd" containerID="cri-o://c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03" gracePeriod=30 Feb 02 12:22:27 crc kubenswrapper[4846]: I0202 12:22:27.916106 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1" gracePeriod=30 Feb 02 12:22:27 crc kubenswrapper[4846]: I0202 12:22:27.916134 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="kube-rbac-proxy-node" containerID="cri-o://01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096" gracePeriod=30 Feb 02 12:22:27 crc kubenswrapper[4846]: I0202 12:22:27.916164 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovn-acl-logging" containerID="cri-o://4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a" gracePeriod=30 Feb 02 12:22:27 crc kubenswrapper[4846]: I0202 12:22:27.949027 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" containerID="cri-o://ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa" gracePeriod=30 Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.017985 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ea83b34_bfee_43e5_8b72_5a59fcf0fc8e.slice/crio-conmon-4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ea83b34_bfee_43e5_8b72_5a59fcf0fc8e.slice/crio-conmon-535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ea83b34_bfee_43e5_8b72_5a59fcf0fc8e.slice/crio-conmon-01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096.scope\": RecentStats: unable to find data in memory cache]" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.281829 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/3.log" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.284900 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovn-acl-logging/0.log" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.285406 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovn-controller/0.log" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.285880 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.343940 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mvtvc"] Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344143 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="northd" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344154 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="northd" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344166 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovn-acl-logging" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344174 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovn-acl-logging" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344195 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="kube-rbac-proxy-ovn-metrics" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344204 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="kube-rbac-proxy-ovn-metrics" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344213 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344219 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344226 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344231 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344240 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344247 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344258 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="kubecfg-setup" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344266 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="kubecfg-setup" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344277 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="sbdb" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344284 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="sbdb" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344299 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="nbdb" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344304 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="nbdb" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344313 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovn-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344319 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovn-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344327 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344333 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344342 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="kube-rbac-proxy-node" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344348 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="kube-rbac-proxy-node" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344430 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344438 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovn-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344446 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="kube-rbac-proxy-node" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344452 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovn-acl-logging" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344459 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="northd" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344465 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344473 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="nbdb" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344480 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344487 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="kube-rbac-proxy-ovn-metrics" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344494 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="sbdb" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.344575 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344582 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344695 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.344857 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerName="ovnkube-controller" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.346216 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.434612 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-var-lib-openvswitch\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.434747 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-node-log\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.434761 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.434789 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-node-log" (OuterVolumeSpecName: "node-log") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435061 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-log-socket\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435091 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-etc-openvswitch\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435118 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-script-lib\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435139 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-ovn\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435138 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-log-socket" (OuterVolumeSpecName: "log-socket") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435171 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435152 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-netd\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435192 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435203 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435226 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87zlr\" (UniqueName: \"kubernetes.io/projected/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-kube-api-access-87zlr\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435278 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-config\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435317 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-kubelet\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435363 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435488 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-slash\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435558 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-bin\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435647 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-netns\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435711 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-openvswitch\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435763 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-systemd-units\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435822 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-ovn-kubernetes\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435391 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435432 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435685 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435788 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-slash" (OuterVolumeSpecName: "host-slash") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435811 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435818 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435832 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435851 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435856 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.435903 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-env-overrides\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436048 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-systemd\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436073 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovn-node-metrics-cert\") pod \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\" (UID: \"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e\") " Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436161 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436191 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-cni-bin\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436217 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-cni-netd\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436267 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-run-netns\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436294 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-etc-openvswitch\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436292 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436314 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7bb8803f-6196-4141-9105-21eb89069a28-ovnkube-script-lib\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436354 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-run-ovn\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436382 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-kubelet\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436407 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-systemd-units\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436429 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7bb8803f-6196-4141-9105-21eb89069a28-ovnkube-config\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436455 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-slash\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436484 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7bb8803f-6196-4141-9105-21eb89069a28-ovn-node-metrics-cert\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436505 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-var-lib-openvswitch\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436547 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-log-socket\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436570 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436592 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-node-log\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436612 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7bb8803f-6196-4141-9105-21eb89069a28-env-overrides\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436653 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvlfg\" (UniqueName: \"kubernetes.io/projected/7bb8803f-6196-4141-9105-21eb89069a28-kube-api-access-bvlfg\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436695 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-run-openvswitch\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436732 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-run-ovn-kubernetes\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436760 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-run-systemd\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436925 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436961 4846 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.436985 4846 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437005 4846 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-slash\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437022 4846 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437040 4846 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437059 4846 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437076 4846 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437093 4846 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437114 4846 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437130 4846 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437147 4846 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-node-log\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437164 4846 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-log-socket\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437180 4846 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437198 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437215 4846 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.437232 4846 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.441862 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.441986 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-kube-api-access-87zlr" (OuterVolumeSpecName: "kube-api-access-87zlr") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "kube-api-access-87zlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.449309 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" (UID: "6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538496 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-run-systemd\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538589 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-cni-bin\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538641 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-cni-netd\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538669 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-run-netns\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538675 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-run-systemd\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538728 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-cni-netd\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538746 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-etc-openvswitch\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538701 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-etc-openvswitch\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538753 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-run-netns\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538743 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-cni-bin\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538796 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7bb8803f-6196-4141-9105-21eb89069a28-ovnkube-script-lib\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-run-ovn\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538882 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-kubelet\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538905 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-systemd-units\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538921 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-run-ovn\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538960 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-systemd-units\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538941 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-kubelet\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.538937 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7bb8803f-6196-4141-9105-21eb89069a28-ovnkube-config\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539080 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-slash\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539104 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7bb8803f-6196-4141-9105-21eb89069a28-ovn-node-metrics-cert\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539124 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-var-lib-openvswitch\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539148 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-log-socket\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539172 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539197 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-node-log\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539217 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7bb8803f-6196-4141-9105-21eb89069a28-env-overrides\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539239 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvlfg\" (UniqueName: \"kubernetes.io/projected/7bb8803f-6196-4141-9105-21eb89069a28-kube-api-access-bvlfg\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539267 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-run-openvswitch\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539290 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-run-ovn-kubernetes\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539337 4846 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539353 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539366 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87zlr\" (UniqueName: \"kubernetes.io/projected/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e-kube-api-access-87zlr\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539397 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-run-ovn-kubernetes\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539427 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-slash\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539534 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7bb8803f-6196-4141-9105-21eb89069a28-ovnkube-script-lib\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539574 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7bb8803f-6196-4141-9105-21eb89069a28-ovnkube-config\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539588 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-node-log\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539615 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-log-socket\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539648 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539663 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-var-lib-openvswitch\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539979 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bb8803f-6196-4141-9105-21eb89069a28-run-openvswitch\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.539981 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7bb8803f-6196-4141-9105-21eb89069a28-env-overrides\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.542986 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7bb8803f-6196-4141-9105-21eb89069a28-ovn-node-metrics-cert\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.557432 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvlfg\" (UniqueName: \"kubernetes.io/projected/7bb8803f-6196-4141-9105-21eb89069a28-kube-api-access-bvlfg\") pod \"ovnkube-node-mvtvc\" (UID: \"7bb8803f-6196-4141-9105-21eb89069a28\") " pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.660813 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.693369 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovnkube-controller/3.log" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.695592 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovn-acl-logging/0.log" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696376 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ftvcw_6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/ovn-controller/0.log" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696782 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa" exitCode=0 Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696811 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78" exitCode=0 Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696820 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3" exitCode=0 Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696829 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03" exitCode=0 Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696836 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1" exitCode=0 Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696845 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096" exitCode=0 Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696853 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a" exitCode=143 Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696862 4846 generic.go:334] "Generic (PLEG): container finished" podID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" containerID="535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c" exitCode=143 Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696862 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696888 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696903 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696914 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696925 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696934 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696943 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696954 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696963 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696968 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696973 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696985 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696991 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696993 4846 scope.go:117] "RemoveContainer" containerID="ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.696997 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697074 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697083 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697094 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697106 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697114 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697121 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697128 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697135 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697146 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697165 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697173 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697180 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697189 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697199 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697212 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697219 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697224 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697231 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697236 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697241 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697246 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697252 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697257 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697262 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697269 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ftvcw" event={"ID":"6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e","Type":"ContainerDied","Data":"f2ae9005ccde81fc6c8ba26dfc3337492496c8bae1840dfbdd1186fbde5cec48"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697277 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697284 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697291 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697304 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697311 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697318 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697324 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697329 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697334 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.697339 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.698806 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7jshv_24fc18fa-9739-480c-8f43-ab30989164c2/kube-multus/2.log" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.699391 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7jshv_24fc18fa-9739-480c-8f43-ab30989164c2/kube-multus/1.log" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.699426 4846 generic.go:334] "Generic (PLEG): container finished" podID="24fc18fa-9739-480c-8f43-ab30989164c2" containerID="4c7034f84bb93b8fc083ef27a05cac5d8f31339db2de4fe421ee29ac6ea3417e" exitCode=2 Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.699493 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7jshv" event={"ID":"24fc18fa-9739-480c-8f43-ab30989164c2","Type":"ContainerDied","Data":"4c7034f84bb93b8fc083ef27a05cac5d8f31339db2de4fe421ee29ac6ea3417e"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.699529 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.699927 4846 scope.go:117] "RemoveContainer" containerID="4c7034f84bb93b8fc083ef27a05cac5d8f31339db2de4fe421ee29ac6ea3417e" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.701227 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" event={"ID":"7bb8803f-6196-4141-9105-21eb89069a28","Type":"ContainerStarted","Data":"1bfbd1c682bc658c99ad2d56be893845fce4ec563967e1e931fb12b0c391f762"} Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.720241 4846 scope.go:117] "RemoveContainer" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.747336 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ftvcw"] Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.749121 4846 scope.go:117] "RemoveContainer" containerID="efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.758050 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ftvcw"] Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.771373 4846 scope.go:117] "RemoveContainer" containerID="3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.788610 4846 scope.go:117] "RemoveContainer" containerID="c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.807872 4846 scope.go:117] "RemoveContainer" containerID="45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.867364 4846 scope.go:117] "RemoveContainer" containerID="01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.884411 4846 scope.go:117] "RemoveContainer" containerID="4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.906425 4846 scope.go:117] "RemoveContainer" containerID="535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.920690 4846 scope.go:117] "RemoveContainer" containerID="f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.935076 4846 scope.go:117] "RemoveContainer" containerID="ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.935727 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": container with ID starting with ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa not found: ID does not exist" containerID="ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.935781 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa"} err="failed to get container status \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": rpc error: code = NotFound desc = could not find container \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": container with ID starting with ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.935814 4846 scope.go:117] "RemoveContainer" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.936303 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\": container with ID starting with a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c not found: ID does not exist" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.936341 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c"} err="failed to get container status \"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\": rpc error: code = NotFound desc = could not find container \"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\": container with ID starting with a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.936365 4846 scope.go:117] "RemoveContainer" containerID="efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.936838 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\": container with ID starting with efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78 not found: ID does not exist" containerID="efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.936869 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78"} err="failed to get container status \"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\": rpc error: code = NotFound desc = could not find container \"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\": container with ID starting with efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.936888 4846 scope.go:117] "RemoveContainer" containerID="3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.937188 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\": container with ID starting with 3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3 not found: ID does not exist" containerID="3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.937215 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3"} err="failed to get container status \"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\": rpc error: code = NotFound desc = could not find container \"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\": container with ID starting with 3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.937234 4846 scope.go:117] "RemoveContainer" containerID="c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.937578 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\": container with ID starting with c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03 not found: ID does not exist" containerID="c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.937604 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03"} err="failed to get container status \"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\": rpc error: code = NotFound desc = could not find container \"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\": container with ID starting with c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.937699 4846 scope.go:117] "RemoveContainer" containerID="45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.938045 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\": container with ID starting with 45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1 not found: ID does not exist" containerID="45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.938086 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1"} err="failed to get container status \"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\": rpc error: code = NotFound desc = could not find container \"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\": container with ID starting with 45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.938112 4846 scope.go:117] "RemoveContainer" containerID="01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.938786 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\": container with ID starting with 01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096 not found: ID does not exist" containerID="01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.938812 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096"} err="failed to get container status \"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\": rpc error: code = NotFound desc = could not find container \"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\": container with ID starting with 01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.938826 4846 scope.go:117] "RemoveContainer" containerID="4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.939928 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\": container with ID starting with 4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a not found: ID does not exist" containerID="4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.939963 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a"} err="failed to get container status \"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\": rpc error: code = NotFound desc = could not find container \"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\": container with ID starting with 4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.939979 4846 scope.go:117] "RemoveContainer" containerID="535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.940236 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\": container with ID starting with 535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c not found: ID does not exist" containerID="535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.940256 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c"} err="failed to get container status \"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\": rpc error: code = NotFound desc = could not find container \"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\": container with ID starting with 535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.940270 4846 scope.go:117] "RemoveContainer" containerID="f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c" Feb 02 12:22:28 crc kubenswrapper[4846]: E0202 12:22:28.940520 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\": container with ID starting with f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c not found: ID does not exist" containerID="f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.940540 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c"} err="failed to get container status \"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\": rpc error: code = NotFound desc = could not find container \"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\": container with ID starting with f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.940552 4846 scope.go:117] "RemoveContainer" containerID="ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.940793 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa"} err="failed to get container status \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": rpc error: code = NotFound desc = could not find container \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": container with ID starting with ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.940814 4846 scope.go:117] "RemoveContainer" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941006 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c"} err="failed to get container status \"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\": rpc error: code = NotFound desc = could not find container \"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\": container with ID starting with a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941024 4846 scope.go:117] "RemoveContainer" containerID="efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941195 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78"} err="failed to get container status \"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\": rpc error: code = NotFound desc = could not find container \"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\": container with ID starting with efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941215 4846 scope.go:117] "RemoveContainer" containerID="3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941367 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3"} err="failed to get container status \"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\": rpc error: code = NotFound desc = could not find container \"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\": container with ID starting with 3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941385 4846 scope.go:117] "RemoveContainer" containerID="c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941544 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03"} err="failed to get container status \"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\": rpc error: code = NotFound desc = could not find container \"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\": container with ID starting with c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941563 4846 scope.go:117] "RemoveContainer" containerID="45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941734 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1"} err="failed to get container status \"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\": rpc error: code = NotFound desc = could not find container \"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\": container with ID starting with 45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941752 4846 scope.go:117] "RemoveContainer" containerID="01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941901 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096"} err="failed to get container status \"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\": rpc error: code = NotFound desc = could not find container \"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\": container with ID starting with 01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.941920 4846 scope.go:117] "RemoveContainer" containerID="4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.942087 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a"} err="failed to get container status \"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\": rpc error: code = NotFound desc = could not find container \"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\": container with ID starting with 4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.942114 4846 scope.go:117] "RemoveContainer" containerID="535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.942270 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c"} err="failed to get container status \"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\": rpc error: code = NotFound desc = could not find container \"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\": container with ID starting with 535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.942288 4846 scope.go:117] "RemoveContainer" containerID="f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.942568 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c"} err="failed to get container status \"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\": rpc error: code = NotFound desc = could not find container \"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\": container with ID starting with f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.942586 4846 scope.go:117] "RemoveContainer" containerID="ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.942766 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa"} err="failed to get container status \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": rpc error: code = NotFound desc = could not find container \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": container with ID starting with ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.942786 4846 scope.go:117] "RemoveContainer" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.943071 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c"} err="failed to get container status \"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\": rpc error: code = NotFound desc = could not find container \"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\": container with ID starting with a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.943119 4846 scope.go:117] "RemoveContainer" containerID="efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.943381 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78"} err="failed to get container status \"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\": rpc error: code = NotFound desc = could not find container \"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\": container with ID starting with efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.943401 4846 scope.go:117] "RemoveContainer" containerID="3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.943611 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3"} err="failed to get container status \"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\": rpc error: code = NotFound desc = could not find container \"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\": container with ID starting with 3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.943645 4846 scope.go:117] "RemoveContainer" containerID="c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.943867 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03"} err="failed to get container status \"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\": rpc error: code = NotFound desc = could not find container \"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\": container with ID starting with c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.943889 4846 scope.go:117] "RemoveContainer" containerID="45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.944823 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1"} err="failed to get container status \"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\": rpc error: code = NotFound desc = could not find container \"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\": container with ID starting with 45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.944844 4846 scope.go:117] "RemoveContainer" containerID="01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.945060 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096"} err="failed to get container status \"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\": rpc error: code = NotFound desc = could not find container \"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\": container with ID starting with 01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.945080 4846 scope.go:117] "RemoveContainer" containerID="4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.945311 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a"} err="failed to get container status \"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\": rpc error: code = NotFound desc = could not find container \"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\": container with ID starting with 4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.945346 4846 scope.go:117] "RemoveContainer" containerID="535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.945570 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c"} err="failed to get container status \"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\": rpc error: code = NotFound desc = could not find container \"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\": container with ID starting with 535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.945589 4846 scope.go:117] "RemoveContainer" containerID="f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.945812 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c"} err="failed to get container status \"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\": rpc error: code = NotFound desc = could not find container \"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\": container with ID starting with f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.945833 4846 scope.go:117] "RemoveContainer" containerID="ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.946011 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa"} err="failed to get container status \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": rpc error: code = NotFound desc = could not find container \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": container with ID starting with ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.946027 4846 scope.go:117] "RemoveContainer" containerID="a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.946184 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c"} err="failed to get container status \"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\": rpc error: code = NotFound desc = could not find container \"a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c\": container with ID starting with a0336b2a987a536ed7b77f04f9392e0caf67ed5509dabc377df15c32eee2ca6c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.946201 4846 scope.go:117] "RemoveContainer" containerID="efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.946389 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78"} err="failed to get container status \"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\": rpc error: code = NotFound desc = could not find container \"efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78\": container with ID starting with efac5c48c57c1c5cde4eb927e65dbdf7e26f11678aa2ef439e749535f588ef78 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.946402 4846 scope.go:117] "RemoveContainer" containerID="3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.946581 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3"} err="failed to get container status \"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\": rpc error: code = NotFound desc = could not find container \"3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3\": container with ID starting with 3afb8f152f8e84ba9fcbc86c929bb5b1049fd45a15354f88d7b7596996fab6b3 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.946604 4846 scope.go:117] "RemoveContainer" containerID="c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.946840 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03"} err="failed to get container status \"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\": rpc error: code = NotFound desc = could not find container \"c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03\": container with ID starting with c14d1b4e17cb8b3bc8e05641a2798a67c0c7984a87a010202e6c852118a04d03 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.946867 4846 scope.go:117] "RemoveContainer" containerID="45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.947099 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1"} err="failed to get container status \"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\": rpc error: code = NotFound desc = could not find container \"45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1\": container with ID starting with 45df5ea494b17117c17d6dc075db03aa1a1d688e62bf3cdea68e3eccfb2091d1 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.947118 4846 scope.go:117] "RemoveContainer" containerID="01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.947352 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096"} err="failed to get container status \"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\": rpc error: code = NotFound desc = could not find container \"01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096\": container with ID starting with 01b0b7ee12d764ccfeb64765428ae43086f37c3578e9121bfca333ceaa1a1096 not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.947380 4846 scope.go:117] "RemoveContainer" containerID="4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.947590 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a"} err="failed to get container status \"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\": rpc error: code = NotFound desc = could not find container \"4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a\": container with ID starting with 4512ec2890842ce59b88c591a4d0b13c6f086d24549fa8a494a6b1b62ccf069a not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.947609 4846 scope.go:117] "RemoveContainer" containerID="535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.947866 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c"} err="failed to get container status \"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\": rpc error: code = NotFound desc = could not find container \"535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c\": container with ID starting with 535a2ef0f1c14c0555c36bf38265382b07c5140a76dd4a85a0e975c5f559479c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.947884 4846 scope.go:117] "RemoveContainer" containerID="f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.948062 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c"} err="failed to get container status \"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\": rpc error: code = NotFound desc = could not find container \"f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c\": container with ID starting with f8a05ff3fb878d427d204b577efa2253d7fa0200ed2cd37552c01d3add84605c not found: ID does not exist" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.948079 4846 scope.go:117] "RemoveContainer" containerID="ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa" Feb 02 12:22:28 crc kubenswrapper[4846]: I0202 12:22:28.948245 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa"} err="failed to get container status \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": rpc error: code = NotFound desc = could not find container \"ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa\": container with ID starting with ae2355e4423083332eb3e1a1127d471e66f0f456438eda4d0bf2fdc9d0d0a1fa not found: ID does not exist" Feb 02 12:22:29 crc kubenswrapper[4846]: I0202 12:22:29.432425 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e" path="/var/lib/kubelet/pods/6ea83b34-bfee-43e5-8b72-5a59fcf0fc8e/volumes" Feb 02 12:22:29 crc kubenswrapper[4846]: I0202 12:22:29.698646 4846 scope.go:117] "RemoveContainer" containerID="abe9450562a4837f44ba7f309722104d4f4320dfbdfe89c5d405c397d2b5af92" Feb 02 12:22:29 crc kubenswrapper[4846]: I0202 12:22:29.708009 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7jshv_24fc18fa-9739-480c-8f43-ab30989164c2/kube-multus/2.log" Feb 02 12:22:29 crc kubenswrapper[4846]: I0202 12:22:29.708822 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7jshv" event={"ID":"24fc18fa-9739-480c-8f43-ab30989164c2","Type":"ContainerStarted","Data":"71e2fe1cbf35c81f797163f927da25570a154956142f50a47ea96caf1ac8a0f3"} Feb 02 12:22:29 crc kubenswrapper[4846]: I0202 12:22:29.711900 4846 generic.go:334] "Generic (PLEG): container finished" podID="7bb8803f-6196-4141-9105-21eb89069a28" containerID="9df09b5f0b9b8c3a08f2a24672ab83511d7376e134685799e24ed7b3eddb40c0" exitCode=0 Feb 02 12:22:29 crc kubenswrapper[4846]: I0202 12:22:29.711930 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" event={"ID":"7bb8803f-6196-4141-9105-21eb89069a28","Type":"ContainerDied","Data":"9df09b5f0b9b8c3a08f2a24672ab83511d7376e134685799e24ed7b3eddb40c0"} Feb 02 12:22:30 crc kubenswrapper[4846]: I0202 12:22:30.718736 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7jshv_24fc18fa-9739-480c-8f43-ab30989164c2/kube-multus/2.log" Feb 02 12:22:30 crc kubenswrapper[4846]: I0202 12:22:30.723494 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" event={"ID":"7bb8803f-6196-4141-9105-21eb89069a28","Type":"ContainerStarted","Data":"178721cc14e246cf8ceb82379ecd954405d4e044c721e46bbdae737350f09ee9"} Feb 02 12:22:30 crc kubenswrapper[4846]: I0202 12:22:30.723536 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" event={"ID":"7bb8803f-6196-4141-9105-21eb89069a28","Type":"ContainerStarted","Data":"6602131f2a842dde52ff685e20b943f014b13ae6ce9444d96bd900448fb562e0"} Feb 02 12:22:30 crc kubenswrapper[4846]: I0202 12:22:30.723549 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" event={"ID":"7bb8803f-6196-4141-9105-21eb89069a28","Type":"ContainerStarted","Data":"d690f7caee3824a44f46f66ab54ee582d56c42b7299d964518aa19dd2ba56ec1"} Feb 02 12:22:30 crc kubenswrapper[4846]: I0202 12:22:30.723558 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" event={"ID":"7bb8803f-6196-4141-9105-21eb89069a28","Type":"ContainerStarted","Data":"582e9b024000e56bb0dddbff58e621bd84fde8d90a8abaaae60013dcdb59d20b"} Feb 02 12:22:30 crc kubenswrapper[4846]: I0202 12:22:30.723567 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" event={"ID":"7bb8803f-6196-4141-9105-21eb89069a28","Type":"ContainerStarted","Data":"bb98775de336461eb769a911414aa3d85b476178139efbeca816545ae4832de2"} Feb 02 12:22:30 crc kubenswrapper[4846]: I0202 12:22:30.723576 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" event={"ID":"7bb8803f-6196-4141-9105-21eb89069a28","Type":"ContainerStarted","Data":"4f6fbe57449f00dd63223d7aa75bea46d4890afba150e228361100d65f7c17cc"} Feb 02 12:22:31 crc kubenswrapper[4846]: I0202 12:22:31.790085 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-md5nm"] Feb 02 12:22:31 crc kubenswrapper[4846]: I0202 12:22:31.790713 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:31 crc kubenswrapper[4846]: I0202 12:22:31.792342 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Feb 02 12:22:31 crc kubenswrapper[4846]: I0202 12:22:31.792441 4846 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-l8clt" Feb 02 12:22:31 crc kubenswrapper[4846]: I0202 12:22:31.793257 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Feb 02 12:22:31 crc kubenswrapper[4846]: I0202 12:22:31.794083 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Feb 02 12:22:31 crc kubenswrapper[4846]: I0202 12:22:31.975664 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-node-mnt\") pod \"crc-storage-crc-md5nm\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:31 crc kubenswrapper[4846]: I0202 12:22:31.975770 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-crc-storage\") pod \"crc-storage-crc-md5nm\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:31 crc kubenswrapper[4846]: I0202 12:22:31.975820 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hhf2\" (UniqueName: \"kubernetes.io/projected/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-kube-api-access-8hhf2\") pod \"crc-storage-crc-md5nm\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:32 crc kubenswrapper[4846]: I0202 12:22:32.076865 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-crc-storage\") pod \"crc-storage-crc-md5nm\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:32 crc kubenswrapper[4846]: I0202 12:22:32.076931 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hhf2\" (UniqueName: \"kubernetes.io/projected/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-kube-api-access-8hhf2\") pod \"crc-storage-crc-md5nm\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:32 crc kubenswrapper[4846]: I0202 12:22:32.076970 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-node-mnt\") pod \"crc-storage-crc-md5nm\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:32 crc kubenswrapper[4846]: I0202 12:22:32.077203 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-node-mnt\") pod \"crc-storage-crc-md5nm\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:32 crc kubenswrapper[4846]: I0202 12:22:32.077776 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-crc-storage\") pod \"crc-storage-crc-md5nm\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:32 crc kubenswrapper[4846]: I0202 12:22:32.099307 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hhf2\" (UniqueName: \"kubernetes.io/projected/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-kube-api-access-8hhf2\") pod \"crc-storage-crc-md5nm\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:32 crc kubenswrapper[4846]: I0202 12:22:32.104768 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:32 crc kubenswrapper[4846]: E0202 12:22:32.139782 4846 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-md5nm_crc-storage_4686b9f2-68b9-4de6-a3ce-aa95d6730abe_0(d521bc8427644261d31180285877b5a81a31acad30436c48f98f8cd0ed7f0379): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 12:22:32 crc kubenswrapper[4846]: E0202 12:22:32.140152 4846 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-md5nm_crc-storage_4686b9f2-68b9-4de6-a3ce-aa95d6730abe_0(d521bc8427644261d31180285877b5a81a31acad30436c48f98f8cd0ed7f0379): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:32 crc kubenswrapper[4846]: E0202 12:22:32.140185 4846 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-md5nm_crc-storage_4686b9f2-68b9-4de6-a3ce-aa95d6730abe_0(d521bc8427644261d31180285877b5a81a31acad30436c48f98f8cd0ed7f0379): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:32 crc kubenswrapper[4846]: E0202 12:22:32.140243 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-md5nm_crc-storage(4686b9f2-68b9-4de6-a3ce-aa95d6730abe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-md5nm_crc-storage(4686b9f2-68b9-4de6-a3ce-aa95d6730abe)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-md5nm_crc-storage_4686b9f2-68b9-4de6-a3ce-aa95d6730abe_0(d521bc8427644261d31180285877b5a81a31acad30436c48f98f8cd0ed7f0379): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-md5nm" podUID="4686b9f2-68b9-4de6-a3ce-aa95d6730abe" Feb 02 12:22:32 crc kubenswrapper[4846]: I0202 12:22:32.735294 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" event={"ID":"7bb8803f-6196-4141-9105-21eb89069a28","Type":"ContainerStarted","Data":"a83c2bab5a9669fff231e4de5e7abbcdca33c487bdfc209e35fa90fe9c962385"} Feb 02 12:22:34 crc kubenswrapper[4846]: I0202 12:22:34.762574 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" event={"ID":"7bb8803f-6196-4141-9105-21eb89069a28","Type":"ContainerStarted","Data":"6424c2293ab490455a3e81d292355a8545a99b5f5f806948cdaa29d6eb160d26"} Feb 02 12:22:34 crc kubenswrapper[4846]: I0202 12:22:34.763508 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:34 crc kubenswrapper[4846]: I0202 12:22:34.763683 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:34 crc kubenswrapper[4846]: I0202 12:22:34.790049 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:34 crc kubenswrapper[4846]: I0202 12:22:34.791146 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" podStartSLOduration=6.791121577 podStartE2EDuration="6.791121577s" podCreationTimestamp="2026-02-02 12:22:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:22:34.790274856 +0000 UTC m=+786.018861719" watchObservedRunningTime="2026-02-02 12:22:34.791121577 +0000 UTC m=+786.019708450" Feb 02 12:22:35 crc kubenswrapper[4846]: I0202 12:22:35.768518 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:35 crc kubenswrapper[4846]: I0202 12:22:35.832927 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:35 crc kubenswrapper[4846]: I0202 12:22:35.960560 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-md5nm"] Feb 02 12:22:35 crc kubenswrapper[4846]: I0202 12:22:35.960692 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:35 crc kubenswrapper[4846]: I0202 12:22:35.961096 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:35 crc kubenswrapper[4846]: E0202 12:22:35.984237 4846 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-md5nm_crc-storage_4686b9f2-68b9-4de6-a3ce-aa95d6730abe_0(ed0406214f20ff427487e6f548fab3da1b51bd6182cae3c15ccc976802f4720e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 12:22:35 crc kubenswrapper[4846]: E0202 12:22:35.984431 4846 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-md5nm_crc-storage_4686b9f2-68b9-4de6-a3ce-aa95d6730abe_0(ed0406214f20ff427487e6f548fab3da1b51bd6182cae3c15ccc976802f4720e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:35 crc kubenswrapper[4846]: E0202 12:22:35.984523 4846 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-md5nm_crc-storage_4686b9f2-68b9-4de6-a3ce-aa95d6730abe_0(ed0406214f20ff427487e6f548fab3da1b51bd6182cae3c15ccc976802f4720e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:35 crc kubenswrapper[4846]: E0202 12:22:35.984680 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-md5nm_crc-storage(4686b9f2-68b9-4de6-a3ce-aa95d6730abe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-md5nm_crc-storage(4686b9f2-68b9-4de6-a3ce-aa95d6730abe)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-md5nm_crc-storage_4686b9f2-68b9-4de6-a3ce-aa95d6730abe_0(ed0406214f20ff427487e6f548fab3da1b51bd6182cae3c15ccc976802f4720e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-md5nm" podUID="4686b9f2-68b9-4de6-a3ce-aa95d6730abe" Feb 02 12:22:48 crc kubenswrapper[4846]: I0202 12:22:48.423137 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:48 crc kubenswrapper[4846]: I0202 12:22:48.424056 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:48 crc kubenswrapper[4846]: I0202 12:22:48.809448 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-md5nm"] Feb 02 12:22:48 crc kubenswrapper[4846]: W0202 12:22:48.814229 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4686b9f2_68b9_4de6_a3ce_aa95d6730abe.slice/crio-169833a469c3fa573a7ede1b05c05c421100ec53d68d325976df50180eb4cccd WatchSource:0}: Error finding container 169833a469c3fa573a7ede1b05c05c421100ec53d68d325976df50180eb4cccd: Status 404 returned error can't find the container with id 169833a469c3fa573a7ede1b05c05c421100ec53d68d325976df50180eb4cccd Feb 02 12:22:48 crc kubenswrapper[4846]: I0202 12:22:48.815955 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 12:22:48 crc kubenswrapper[4846]: I0202 12:22:48.845849 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-md5nm" event={"ID":"4686b9f2-68b9-4de6-a3ce-aa95d6730abe","Type":"ContainerStarted","Data":"169833a469c3fa573a7ede1b05c05c421100ec53d68d325976df50180eb4cccd"} Feb 02 12:22:50 crc kubenswrapper[4846]: I0202 12:22:50.858396 4846 generic.go:334] "Generic (PLEG): container finished" podID="4686b9f2-68b9-4de6-a3ce-aa95d6730abe" containerID="6ce66fb16deff0f3558660d1093b0c9c70cae937604eddce8f29d9ee7d328df9" exitCode=0 Feb 02 12:22:50 crc kubenswrapper[4846]: I0202 12:22:50.858447 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-md5nm" event={"ID":"4686b9f2-68b9-4de6-a3ce-aa95d6730abe","Type":"ContainerDied","Data":"6ce66fb16deff0f3558660d1093b0c9c70cae937604eddce8f29d9ee7d328df9"} Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.091021 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.223037 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hhf2\" (UniqueName: \"kubernetes.io/projected/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-kube-api-access-8hhf2\") pod \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.223087 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-node-mnt\") pod \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.223180 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-crc-storage\") pod \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\" (UID: \"4686b9f2-68b9-4de6-a3ce-aa95d6730abe\") " Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.223264 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "4686b9f2-68b9-4de6-a3ce-aa95d6730abe" (UID: "4686b9f2-68b9-4de6-a3ce-aa95d6730abe"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.223448 4846 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-node-mnt\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.229832 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-kube-api-access-8hhf2" (OuterVolumeSpecName: "kube-api-access-8hhf2") pod "4686b9f2-68b9-4de6-a3ce-aa95d6730abe" (UID: "4686b9f2-68b9-4de6-a3ce-aa95d6730abe"). InnerVolumeSpecName "kube-api-access-8hhf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.239771 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "4686b9f2-68b9-4de6-a3ce-aa95d6730abe" (UID: "4686b9f2-68b9-4de6-a3ce-aa95d6730abe"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.324867 4846 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-crc-storage\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.324913 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hhf2\" (UniqueName: \"kubernetes.io/projected/4686b9f2-68b9-4de6-a3ce-aa95d6730abe-kube-api-access-8hhf2\") on node \"crc\" DevicePath \"\"" Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.871885 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-md5nm" event={"ID":"4686b9f2-68b9-4de6-a3ce-aa95d6730abe","Type":"ContainerDied","Data":"169833a469c3fa573a7ede1b05c05c421100ec53d68d325976df50180eb4cccd"} Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.872092 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="169833a469c3fa573a7ede1b05c05c421100ec53d68d325976df50180eb4cccd" Feb 02 12:22:52 crc kubenswrapper[4846]: I0202 12:22:52.871956 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-md5nm" Feb 02 12:22:58 crc kubenswrapper[4846]: I0202 12:22:58.694156 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mvtvc" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.196394 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx"] Feb 02 12:22:59 crc kubenswrapper[4846]: E0202 12:22:59.196690 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4686b9f2-68b9-4de6-a3ce-aa95d6730abe" containerName="storage" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.196707 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4686b9f2-68b9-4de6-a3ce-aa95d6730abe" containerName="storage" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.196820 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4686b9f2-68b9-4de6-a3ce-aa95d6730abe" containerName="storage" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.197692 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.200569 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.210816 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx"] Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.218774 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49znz\" (UniqueName: \"kubernetes.io/projected/b7413188-295c-46c8-b396-cbfc04fc1178-kube-api-access-49znz\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.218820 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.218840 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.319935 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49znz\" (UniqueName: \"kubernetes.io/projected/b7413188-295c-46c8-b396-cbfc04fc1178-kube-api-access-49znz\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.319998 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.320021 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.320573 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.320664 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.340102 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49znz\" (UniqueName: \"kubernetes.io/projected/b7413188-295c-46c8-b396-cbfc04fc1178-kube-api-access-49znz\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.551352 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.746224 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx"] Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.908057 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" event={"ID":"b7413188-295c-46c8-b396-cbfc04fc1178","Type":"ContainerStarted","Data":"a49e195fb555d8771c42be84df63370b04c9b4896f0f8c986e522bf5842c563d"} Feb 02 12:22:59 crc kubenswrapper[4846]: I0202 12:22:59.908398 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" event={"ID":"b7413188-295c-46c8-b396-cbfc04fc1178","Type":"ContainerStarted","Data":"5a3d5ef85f1491661c81458f17cf6de4012f401113d18b0a9db79d65ccf29c3e"} Feb 02 12:23:00 crc kubenswrapper[4846]: I0202 12:23:00.914909 4846 generic.go:334] "Generic (PLEG): container finished" podID="b7413188-295c-46c8-b396-cbfc04fc1178" containerID="a49e195fb555d8771c42be84df63370b04c9b4896f0f8c986e522bf5842c563d" exitCode=0 Feb 02 12:23:00 crc kubenswrapper[4846]: I0202 12:23:00.914973 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" event={"ID":"b7413188-295c-46c8-b396-cbfc04fc1178","Type":"ContainerDied","Data":"a49e195fb555d8771c42be84df63370b04c9b4896f0f8c986e522bf5842c563d"} Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.537666 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zc7mx"] Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.539053 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.549786 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zc7mx"] Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.549938 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-catalog-content\") pod \"redhat-operators-zc7mx\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.550236 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2htl\" (UniqueName: \"kubernetes.io/projected/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-kube-api-access-v2htl\") pod \"redhat-operators-zc7mx\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.550422 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-utilities\") pod \"redhat-operators-zc7mx\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.652046 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-catalog-content\") pod \"redhat-operators-zc7mx\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.652096 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2htl\" (UniqueName: \"kubernetes.io/projected/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-kube-api-access-v2htl\") pod \"redhat-operators-zc7mx\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.652151 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-utilities\") pod \"redhat-operators-zc7mx\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.652692 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-utilities\") pod \"redhat-operators-zc7mx\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.652701 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-catalog-content\") pod \"redhat-operators-zc7mx\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.671125 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2htl\" (UniqueName: \"kubernetes.io/projected/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-kube-api-access-v2htl\") pod \"redhat-operators-zc7mx\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:01 crc kubenswrapper[4846]: I0202 12:23:01.867062 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:02 crc kubenswrapper[4846]: I0202 12:23:02.057834 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zc7mx"] Feb 02 12:23:02 crc kubenswrapper[4846]: W0202 12:23:02.068186 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a7e3682_8bdf_4e88_a3f4_8613b30beadc.slice/crio-f42ce97cc169c1b9801ffdfe5bd02796a1d5988465443234913fa57fb7a6b693 WatchSource:0}: Error finding container f42ce97cc169c1b9801ffdfe5bd02796a1d5988465443234913fa57fb7a6b693: Status 404 returned error can't find the container with id f42ce97cc169c1b9801ffdfe5bd02796a1d5988465443234913fa57fb7a6b693 Feb 02 12:23:02 crc kubenswrapper[4846]: I0202 12:23:02.926447 4846 generic.go:334] "Generic (PLEG): container finished" podID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerID="165a142323bdb7d49dc63f939e5b5cea8a556db3edf051f0d3d431367dcaa31e" exitCode=0 Feb 02 12:23:02 crc kubenswrapper[4846]: I0202 12:23:02.926531 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7mx" event={"ID":"0a7e3682-8bdf-4e88-a3f4-8613b30beadc","Type":"ContainerDied","Data":"165a142323bdb7d49dc63f939e5b5cea8a556db3edf051f0d3d431367dcaa31e"} Feb 02 12:23:02 crc kubenswrapper[4846]: I0202 12:23:02.926761 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7mx" event={"ID":"0a7e3682-8bdf-4e88-a3f4-8613b30beadc","Type":"ContainerStarted","Data":"f42ce97cc169c1b9801ffdfe5bd02796a1d5988465443234913fa57fb7a6b693"} Feb 02 12:23:02 crc kubenswrapper[4846]: I0202 12:23:02.928437 4846 generic.go:334] "Generic (PLEG): container finished" podID="b7413188-295c-46c8-b396-cbfc04fc1178" containerID="606ce584295b3594a43c131cbc10857182dba86ae970472fbff13a7f17e86213" exitCode=0 Feb 02 12:23:02 crc kubenswrapper[4846]: I0202 12:23:02.928494 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" event={"ID":"b7413188-295c-46c8-b396-cbfc04fc1178","Type":"ContainerDied","Data":"606ce584295b3594a43c131cbc10857182dba86ae970472fbff13a7f17e86213"} Feb 02 12:23:03 crc kubenswrapper[4846]: I0202 12:23:03.937472 4846 generic.go:334] "Generic (PLEG): container finished" podID="b7413188-295c-46c8-b396-cbfc04fc1178" containerID="c10689cbc53b91d64bfa9b4c43414d9a2e5ec16981bf5df7bb04c7d20971e6fb" exitCode=0 Feb 02 12:23:03 crc kubenswrapper[4846]: I0202 12:23:03.937523 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" event={"ID":"b7413188-295c-46c8-b396-cbfc04fc1178","Type":"ContainerDied","Data":"c10689cbc53b91d64bfa9b4c43414d9a2e5ec16981bf5df7bb04c7d20971e6fb"} Feb 02 12:23:04 crc kubenswrapper[4846]: I0202 12:23:04.947168 4846 generic.go:334] "Generic (PLEG): container finished" podID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerID="15ef0df99893c7ebfe3a48f5d052af89812819cb94f253137e21db29b548cc07" exitCode=0 Feb 02 12:23:04 crc kubenswrapper[4846]: I0202 12:23:04.947268 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7mx" event={"ID":"0a7e3682-8bdf-4e88-a3f4-8613b30beadc","Type":"ContainerDied","Data":"15ef0df99893c7ebfe3a48f5d052af89812819cb94f253137e21db29b548cc07"} Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.159076 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.294400 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-bundle\") pod \"b7413188-295c-46c8-b396-cbfc04fc1178\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.294478 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49znz\" (UniqueName: \"kubernetes.io/projected/b7413188-295c-46c8-b396-cbfc04fc1178-kube-api-access-49znz\") pod \"b7413188-295c-46c8-b396-cbfc04fc1178\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.294536 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-util\") pod \"b7413188-295c-46c8-b396-cbfc04fc1178\" (UID: \"b7413188-295c-46c8-b396-cbfc04fc1178\") " Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.295765 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-bundle" (OuterVolumeSpecName: "bundle") pod "b7413188-295c-46c8-b396-cbfc04fc1178" (UID: "b7413188-295c-46c8-b396-cbfc04fc1178"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.299984 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7413188-295c-46c8-b396-cbfc04fc1178-kube-api-access-49znz" (OuterVolumeSpecName: "kube-api-access-49znz") pod "b7413188-295c-46c8-b396-cbfc04fc1178" (UID: "b7413188-295c-46c8-b396-cbfc04fc1178"). InnerVolumeSpecName "kube-api-access-49znz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.396987 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.397240 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49znz\" (UniqueName: \"kubernetes.io/projected/b7413188-295c-46c8-b396-cbfc04fc1178-kube-api-access-49znz\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.402179 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-util" (OuterVolumeSpecName: "util") pod "b7413188-295c-46c8-b396-cbfc04fc1178" (UID: "b7413188-295c-46c8-b396-cbfc04fc1178"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.498959 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7413188-295c-46c8-b396-cbfc04fc1178-util\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.964890 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" event={"ID":"b7413188-295c-46c8-b396-cbfc04fc1178","Type":"ContainerDied","Data":"5a3d5ef85f1491661c81458f17cf6de4012f401113d18b0a9db79d65ccf29c3e"} Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.964935 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a3d5ef85f1491661c81458f17cf6de4012f401113d18b0a9db79d65ccf29c3e" Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.965002 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx" Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.968467 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7mx" event={"ID":"0a7e3682-8bdf-4e88-a3f4-8613b30beadc","Type":"ContainerStarted","Data":"8cee6bf8dd00e59f44bc1b935bee1623e34b732d2be07b821d8c689f0d040c39"} Feb 02 12:23:05 crc kubenswrapper[4846]: I0202 12:23:05.992999 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zc7mx" podStartSLOduration=2.505155706 podStartE2EDuration="4.992970833s" podCreationTimestamp="2026-02-02 12:23:01 +0000 UTC" firstStartedPulling="2026-02-02 12:23:02.930066068 +0000 UTC m=+814.158652931" lastFinishedPulling="2026-02-02 12:23:05.417881195 +0000 UTC m=+816.646468058" observedRunningTime="2026-02-02 12:23:05.987462459 +0000 UTC m=+817.216049332" watchObservedRunningTime="2026-02-02 12:23:05.992970833 +0000 UTC m=+817.221557706" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.588940 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-dgl5b"] Feb 02 12:23:10 crc kubenswrapper[4846]: E0202 12:23:10.589850 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7413188-295c-46c8-b396-cbfc04fc1178" containerName="extract" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.589868 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7413188-295c-46c8-b396-cbfc04fc1178" containerName="extract" Feb 02 12:23:10 crc kubenswrapper[4846]: E0202 12:23:10.589889 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7413188-295c-46c8-b396-cbfc04fc1178" containerName="pull" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.589896 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7413188-295c-46c8-b396-cbfc04fc1178" containerName="pull" Feb 02 12:23:10 crc kubenswrapper[4846]: E0202 12:23:10.589907 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7413188-295c-46c8-b396-cbfc04fc1178" containerName="util" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.589914 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7413188-295c-46c8-b396-cbfc04fc1178" containerName="util" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.590029 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7413188-295c-46c8-b396-cbfc04fc1178" containerName="extract" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.590510 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-dgl5b" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.594353 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.594835 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4sbxr" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.594951 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.604420 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-dgl5b"] Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.759813 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvwhm\" (UniqueName: \"kubernetes.io/projected/39214593-1488-4797-8b63-a77baba63417-kube-api-access-mvwhm\") pod \"nmstate-operator-646758c888-dgl5b\" (UID: \"39214593-1488-4797-8b63-a77baba63417\") " pod="openshift-nmstate/nmstate-operator-646758c888-dgl5b" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.860660 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvwhm\" (UniqueName: \"kubernetes.io/projected/39214593-1488-4797-8b63-a77baba63417-kube-api-access-mvwhm\") pod \"nmstate-operator-646758c888-dgl5b\" (UID: \"39214593-1488-4797-8b63-a77baba63417\") " pod="openshift-nmstate/nmstate-operator-646758c888-dgl5b" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.880265 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvwhm\" (UniqueName: \"kubernetes.io/projected/39214593-1488-4797-8b63-a77baba63417-kube-api-access-mvwhm\") pod \"nmstate-operator-646758c888-dgl5b\" (UID: \"39214593-1488-4797-8b63-a77baba63417\") " pod="openshift-nmstate/nmstate-operator-646758c888-dgl5b" Feb 02 12:23:10 crc kubenswrapper[4846]: I0202 12:23:10.905880 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-dgl5b" Feb 02 12:23:11 crc kubenswrapper[4846]: I0202 12:23:11.084396 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-dgl5b"] Feb 02 12:23:11 crc kubenswrapper[4846]: I0202 12:23:11.867715 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:11 crc kubenswrapper[4846]: I0202 12:23:11.867758 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:11 crc kubenswrapper[4846]: I0202 12:23:11.909084 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:12 crc kubenswrapper[4846]: I0202 12:23:12.008240 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-dgl5b" event={"ID":"39214593-1488-4797-8b63-a77baba63417","Type":"ContainerStarted","Data":"3459b9eafe1aa879289ae50003233cdb96d80cd1f70bf3d2f85a98e86012eb8e"} Feb 02 12:23:12 crc kubenswrapper[4846]: I0202 12:23:12.052504 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:14 crc kubenswrapper[4846]: I0202 12:23:14.018117 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-dgl5b" event={"ID":"39214593-1488-4797-8b63-a77baba63417","Type":"ContainerStarted","Data":"4693deb83b952b08b6b827805b9199c789f55e5dc4c19536654aa144e7bc91fd"} Feb 02 12:23:14 crc kubenswrapper[4846]: I0202 12:23:14.036391 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-dgl5b" podStartSLOduration=2.083856852 podStartE2EDuration="4.036370508s" podCreationTimestamp="2026-02-02 12:23:10 +0000 UTC" firstStartedPulling="2026-02-02 12:23:11.09123602 +0000 UTC m=+822.319822883" lastFinishedPulling="2026-02-02 12:23:13.043749676 +0000 UTC m=+824.272336539" observedRunningTime="2026-02-02 12:23:14.033364835 +0000 UTC m=+825.261951708" watchObservedRunningTime="2026-02-02 12:23:14.036370508 +0000 UTC m=+825.264957401" Feb 02 12:23:14 crc kubenswrapper[4846]: I0202 12:23:14.322072 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zc7mx"] Feb 02 12:23:14 crc kubenswrapper[4846]: I0202 12:23:14.322526 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zc7mx" podUID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerName="registry-server" containerID="cri-o://8cee6bf8dd00e59f44bc1b935bee1623e34b732d2be07b821d8c689f0d040c39" gracePeriod=2 Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.030064 4846 generic.go:334] "Generic (PLEG): container finished" podID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerID="8cee6bf8dd00e59f44bc1b935bee1623e34b732d2be07b821d8c689f0d040c39" exitCode=0 Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.030129 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7mx" event={"ID":"0a7e3682-8bdf-4e88-a3f4-8613b30beadc","Type":"ContainerDied","Data":"8cee6bf8dd00e59f44bc1b935bee1623e34b732d2be07b821d8c689f0d040c39"} Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.472383 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.632944 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-catalog-content\") pod \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.632985 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2htl\" (UniqueName: \"kubernetes.io/projected/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-kube-api-access-v2htl\") pod \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.633014 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-utilities\") pod \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\" (UID: \"0a7e3682-8bdf-4e88-a3f4-8613b30beadc\") " Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.633800 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-utilities" (OuterVolumeSpecName: "utilities") pod "0a7e3682-8bdf-4e88-a3f4-8613b30beadc" (UID: "0a7e3682-8bdf-4e88-a3f4-8613b30beadc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.640343 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-kube-api-access-v2htl" (OuterVolumeSpecName: "kube-api-access-v2htl") pod "0a7e3682-8bdf-4e88-a3f4-8613b30beadc" (UID: "0a7e3682-8bdf-4e88-a3f4-8613b30beadc"). InnerVolumeSpecName "kube-api-access-v2htl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.734263 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2htl\" (UniqueName: \"kubernetes.io/projected/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-kube-api-access-v2htl\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.734485 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.735966 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a7e3682-8bdf-4e88-a3f4-8613b30beadc" (UID: "0a7e3682-8bdf-4e88-a3f4-8613b30beadc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:23:16 crc kubenswrapper[4846]: I0202 12:23:16.835784 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a7e3682-8bdf-4e88-a3f4-8613b30beadc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:17 crc kubenswrapper[4846]: I0202 12:23:17.050468 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7mx" event={"ID":"0a7e3682-8bdf-4e88-a3f4-8613b30beadc","Type":"ContainerDied","Data":"f42ce97cc169c1b9801ffdfe5bd02796a1d5988465443234913fa57fb7a6b693"} Feb 02 12:23:17 crc kubenswrapper[4846]: I0202 12:23:17.050546 4846 scope.go:117] "RemoveContainer" containerID="8cee6bf8dd00e59f44bc1b935bee1623e34b732d2be07b821d8c689f0d040c39" Feb 02 12:23:17 crc kubenswrapper[4846]: I0202 12:23:17.050549 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc7mx" Feb 02 12:23:17 crc kubenswrapper[4846]: I0202 12:23:17.070462 4846 scope.go:117] "RemoveContainer" containerID="15ef0df99893c7ebfe3a48f5d052af89812819cb94f253137e21db29b548cc07" Feb 02 12:23:17 crc kubenswrapper[4846]: I0202 12:23:17.084304 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zc7mx"] Feb 02 12:23:17 crc kubenswrapper[4846]: I0202 12:23:17.089148 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zc7mx"] Feb 02 12:23:17 crc kubenswrapper[4846]: I0202 12:23:17.103493 4846 scope.go:117] "RemoveContainer" containerID="165a142323bdb7d49dc63f939e5b5cea8a556db3edf051f0d3d431367dcaa31e" Feb 02 12:23:17 crc kubenswrapper[4846]: I0202 12:23:17.429894 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" path="/var/lib/kubelet/pods/0a7e3682-8bdf-4e88-a3f4-8613b30beadc/volumes" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.371765 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-jzshh"] Feb 02 12:23:19 crc kubenswrapper[4846]: E0202 12:23:19.372951 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerName="registry-server" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.373030 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerName="registry-server" Feb 02 12:23:19 crc kubenswrapper[4846]: E0202 12:23:19.373114 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerName="extract-content" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.373168 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerName="extract-content" Feb 02 12:23:19 crc kubenswrapper[4846]: E0202 12:23:19.373238 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerName="extract-utilities" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.373308 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerName="extract-utilities" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.373531 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a7e3682-8bdf-4e88-a3f4-8613b30beadc" containerName="registry-server" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.374271 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-jzshh" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.376472 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-srhzd" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.390019 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-jzshh"] Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.395763 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg"] Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.396702 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.399182 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg"] Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.401279 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.414140 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-7t7k9"] Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.415347 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.468770 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk2x7\" (UniqueName: \"kubernetes.io/projected/395d0b79-6975-4843-9a3f-f53121bfc1bc-kube-api-access-dk2x7\") pod \"nmstate-metrics-54757c584b-jzshh\" (UID: \"395d0b79-6975-4843-9a3f-f53121bfc1bc\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-jzshh" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.502672 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r"] Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.503287 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.505526 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-fk6gh" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.505732 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.505879 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.522131 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r"] Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.570257 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f34abdcc-0d3c-4133-8437-7e354fb8cea7-dbus-socket\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.570588 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w89j2\" (UniqueName: \"kubernetes.io/projected/f34abdcc-0d3c-4133-8437-7e354fb8cea7-kube-api-access-w89j2\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.570852 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f34abdcc-0d3c-4133-8437-7e354fb8cea7-nmstate-lock\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.570903 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk2x7\" (UniqueName: \"kubernetes.io/projected/395d0b79-6975-4843-9a3f-f53121bfc1bc-kube-api-access-dk2x7\") pod \"nmstate-metrics-54757c584b-jzshh\" (UID: \"395d0b79-6975-4843-9a3f-f53121bfc1bc\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-jzshh" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.570942 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5f0ff130-2029-4436-8be0-a49dc96987b0-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-qhlrg\" (UID: \"5f0ff130-2029-4436-8be0-a49dc96987b0\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.570992 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f34abdcc-0d3c-4133-8437-7e354fb8cea7-ovs-socket\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.571032 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqzcw\" (UniqueName: \"kubernetes.io/projected/5f0ff130-2029-4436-8be0-a49dc96987b0-kube-api-access-gqzcw\") pod \"nmstate-webhook-8474b5b9d8-qhlrg\" (UID: \"5f0ff130-2029-4436-8be0-a49dc96987b0\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.590097 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk2x7\" (UniqueName: \"kubernetes.io/projected/395d0b79-6975-4843-9a3f-f53121bfc1bc-kube-api-access-dk2x7\") pod \"nmstate-metrics-54757c584b-jzshh\" (UID: \"395d0b79-6975-4843-9a3f-f53121bfc1bc\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-jzshh" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.672469 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w89j2\" (UniqueName: \"kubernetes.io/projected/f34abdcc-0d3c-4133-8437-7e354fb8cea7-kube-api-access-w89j2\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.672909 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f34abdcc-0d3c-4133-8437-7e354fb8cea7-nmstate-lock\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.672958 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5f0ff130-2029-4436-8be0-a49dc96987b0-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-qhlrg\" (UID: \"5f0ff130-2029-4436-8be0-a49dc96987b0\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.672985 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f34abdcc-0d3c-4133-8437-7e354fb8cea7-ovs-socket\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.673010 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-gpb7r\" (UID: \"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.673006 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f34abdcc-0d3c-4133-8437-7e354fb8cea7-nmstate-lock\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.673040 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqzcw\" (UniqueName: \"kubernetes.io/projected/5f0ff130-2029-4436-8be0-a49dc96987b0-kube-api-access-gqzcw\") pod \"nmstate-webhook-8474b5b9d8-qhlrg\" (UID: \"5f0ff130-2029-4436-8be0-a49dc96987b0\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.673067 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-gpb7r\" (UID: \"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.673072 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f34abdcc-0d3c-4133-8437-7e354fb8cea7-ovs-socket\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.673092 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f34abdcc-0d3c-4133-8437-7e354fb8cea7-dbus-socket\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.673132 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpzw2\" (UniqueName: \"kubernetes.io/projected/f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f-kube-api-access-dpzw2\") pod \"nmstate-console-plugin-7754f76f8b-gpb7r\" (UID: \"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.673540 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f34abdcc-0d3c-4133-8437-7e354fb8cea7-dbus-socket\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.686631 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5f0ff130-2029-4436-8be0-a49dc96987b0-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-qhlrg\" (UID: \"5f0ff130-2029-4436-8be0-a49dc96987b0\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.690160 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-jzshh" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.691527 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-75bdbb7d58-f8649"] Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.692334 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.698950 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqzcw\" (UniqueName: \"kubernetes.io/projected/5f0ff130-2029-4436-8be0-a49dc96987b0-kube-api-access-gqzcw\") pod \"nmstate-webhook-8474b5b9d8-qhlrg\" (UID: \"5f0ff130-2029-4436-8be0-a49dc96987b0\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.702973 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75bdbb7d58-f8649"] Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.704231 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w89j2\" (UniqueName: \"kubernetes.io/projected/f34abdcc-0d3c-4133-8437-7e354fb8cea7-kube-api-access-w89j2\") pod \"nmstate-handler-7t7k9\" (UID: \"f34abdcc-0d3c-4133-8437-7e354fb8cea7\") " pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.716165 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.739364 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.775573 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-oauth-serving-cert\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.775641 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5948465-e9a6-40bd-b594-faf70dc46a44-console-oauth-config\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.775681 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpzw2\" (UniqueName: \"kubernetes.io/projected/f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f-kube-api-access-dpzw2\") pod \"nmstate-console-plugin-7754f76f8b-gpb7r\" (UID: \"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.775714 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-service-ca\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.775746 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf9lc\" (UniqueName: \"kubernetes.io/projected/c5948465-e9a6-40bd-b594-faf70dc46a44-kube-api-access-xf9lc\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.775800 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5948465-e9a6-40bd-b594-faf70dc46a44-console-serving-cert\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.775851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-gpb7r\" (UID: \"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.775873 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-gpb7r\" (UID: \"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.775891 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-console-config\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.775905 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-trusted-ca-bundle\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.776787 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-gpb7r\" (UID: \"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.782307 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-gpb7r\" (UID: \"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.795376 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpzw2\" (UniqueName: \"kubernetes.io/projected/f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f-kube-api-access-dpzw2\") pod \"nmstate-console-plugin-7754f76f8b-gpb7r\" (UID: \"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.825205 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.877876 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf9lc\" (UniqueName: \"kubernetes.io/projected/c5948465-e9a6-40bd-b594-faf70dc46a44-kube-api-access-xf9lc\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.877926 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5948465-e9a6-40bd-b594-faf70dc46a44-console-serving-cert\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.877997 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-console-config\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.878012 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-trusted-ca-bundle\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.878056 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-oauth-serving-cert\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.878070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5948465-e9a6-40bd-b594-faf70dc46a44-console-oauth-config\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.878114 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-service-ca\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.879700 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-service-ca\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.880696 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-console-config\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.882000 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-trusted-ca-bundle\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.882438 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c5948465-e9a6-40bd-b594-faf70dc46a44-oauth-serving-cert\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.891075 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c5948465-e9a6-40bd-b594-faf70dc46a44-console-oauth-config\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.895405 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5948465-e9a6-40bd-b594-faf70dc46a44-console-serving-cert\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.896374 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf9lc\" (UniqueName: \"kubernetes.io/projected/c5948465-e9a6-40bd-b594-faf70dc46a44-kube-api-access-xf9lc\") pod \"console-75bdbb7d58-f8649\" (UID: \"c5948465-e9a6-40bd-b594-faf70dc46a44\") " pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.923685 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-jzshh"] Feb 02 12:23:19 crc kubenswrapper[4846]: I0202 12:23:19.965220 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg"] Feb 02 12:23:19 crc kubenswrapper[4846]: W0202 12:23:19.967771 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f0ff130_2029_4436_8be0_a49dc96987b0.slice/crio-8cdd6f8616bf9394e1999452cfa9dcb9e407d11b5d50e58cde92da0e379b5953 WatchSource:0}: Error finding container 8cdd6f8616bf9394e1999452cfa9dcb9e407d11b5d50e58cde92da0e379b5953: Status 404 returned error can't find the container with id 8cdd6f8616bf9394e1999452cfa9dcb9e407d11b5d50e58cde92da0e379b5953 Feb 02 12:23:20 crc kubenswrapper[4846]: I0202 12:23:20.019955 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r"] Feb 02 12:23:20 crc kubenswrapper[4846]: I0202 12:23:20.076365 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:20 crc kubenswrapper[4846]: I0202 12:23:20.078481 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7t7k9" event={"ID":"f34abdcc-0d3c-4133-8437-7e354fb8cea7","Type":"ContainerStarted","Data":"275efaef0300a985e28f79c778c207ae31edc9577a04c0e64a15ddb373c79d9b"} Feb 02 12:23:20 crc kubenswrapper[4846]: I0202 12:23:20.079841 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" event={"ID":"5f0ff130-2029-4436-8be0-a49dc96987b0","Type":"ContainerStarted","Data":"8cdd6f8616bf9394e1999452cfa9dcb9e407d11b5d50e58cde92da0e379b5953"} Feb 02 12:23:20 crc kubenswrapper[4846]: I0202 12:23:20.081023 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" event={"ID":"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f","Type":"ContainerStarted","Data":"f544a8d4813e12551d412497127e692df7f715fe84b34b534a71c2ac5e82bbdb"} Feb 02 12:23:20 crc kubenswrapper[4846]: I0202 12:23:20.082801 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-jzshh" event={"ID":"395d0b79-6975-4843-9a3f-f53121bfc1bc","Type":"ContainerStarted","Data":"31d04b7f082632f486998a63d7fa837b32d613e7fd7fb83b0c0e9cee0b150cdb"} Feb 02 12:23:20 crc kubenswrapper[4846]: I0202 12:23:20.257565 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75bdbb7d58-f8649"] Feb 02 12:23:20 crc kubenswrapper[4846]: W0202 12:23:20.264699 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5948465_e9a6_40bd_b594_faf70dc46a44.slice/crio-316a54c8f178102d3e848d35e89973c7cb32d327c725163652fbd4cf7b729583 WatchSource:0}: Error finding container 316a54c8f178102d3e848d35e89973c7cb32d327c725163652fbd4cf7b729583: Status 404 returned error can't find the container with id 316a54c8f178102d3e848d35e89973c7cb32d327c725163652fbd4cf7b729583 Feb 02 12:23:21 crc kubenswrapper[4846]: I0202 12:23:21.097406 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75bdbb7d58-f8649" event={"ID":"c5948465-e9a6-40bd-b594-faf70dc46a44","Type":"ContainerStarted","Data":"c09ad205275cbf6380263dfbf0320ada95f6d201fd363bb72b7154367147b8d1"} Feb 02 12:23:21 crc kubenswrapper[4846]: I0202 12:23:21.097862 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75bdbb7d58-f8649" event={"ID":"c5948465-e9a6-40bd-b594-faf70dc46a44","Type":"ContainerStarted","Data":"316a54c8f178102d3e848d35e89973c7cb32d327c725163652fbd4cf7b729583"} Feb 02 12:23:21 crc kubenswrapper[4846]: I0202 12:23:21.120467 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-75bdbb7d58-f8649" podStartSLOduration=2.120443213 podStartE2EDuration="2.120443213s" podCreationTimestamp="2026-02-02 12:23:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:23:21.113199957 +0000 UTC m=+832.341786820" watchObservedRunningTime="2026-02-02 12:23:21.120443213 +0000 UTC m=+832.349030076" Feb 02 12:23:23 crc kubenswrapper[4846]: I0202 12:23:23.151428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7t7k9" event={"ID":"f34abdcc-0d3c-4133-8437-7e354fb8cea7","Type":"ContainerStarted","Data":"3348d00a105a5325ad8d88c8eab1073980e34497a61a6e06c8b8aa963f6b33e9"} Feb 02 12:23:23 crc kubenswrapper[4846]: I0202 12:23:23.151971 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:23 crc kubenswrapper[4846]: I0202 12:23:23.153547 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" event={"ID":"5f0ff130-2029-4436-8be0-a49dc96987b0","Type":"ContainerStarted","Data":"56ddc9d69e76c7d2f750b859f6fa8ad1d2aef02c2c8e3ef84af334359b5fb9b9"} Feb 02 12:23:23 crc kubenswrapper[4846]: I0202 12:23:23.153655 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" Feb 02 12:23:23 crc kubenswrapper[4846]: I0202 12:23:23.155270 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" event={"ID":"f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f","Type":"ContainerStarted","Data":"513e68e5ed03416d6ce06391fc63f4d986c29c8a6d843c47f33730ea92a61b16"} Feb 02 12:23:23 crc kubenswrapper[4846]: I0202 12:23:23.156799 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-jzshh" event={"ID":"395d0b79-6975-4843-9a3f-f53121bfc1bc","Type":"ContainerStarted","Data":"91621dcaf351f27099d915d7e873fdf09f13681ffd65ddc6cfb4b92861e8ee55"} Feb 02 12:23:23 crc kubenswrapper[4846]: I0202 12:23:23.168594 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-7t7k9" podStartSLOduration=1.362587424 podStartE2EDuration="4.168571377s" podCreationTimestamp="2026-02-02 12:23:19 +0000 UTC" firstStartedPulling="2026-02-02 12:23:19.767201402 +0000 UTC m=+830.995788265" lastFinishedPulling="2026-02-02 12:23:22.573185345 +0000 UTC m=+833.801772218" observedRunningTime="2026-02-02 12:23:23.166020024 +0000 UTC m=+834.394606907" watchObservedRunningTime="2026-02-02 12:23:23.168571377 +0000 UTC m=+834.397158240" Feb 02 12:23:23 crc kubenswrapper[4846]: I0202 12:23:23.185378 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" podStartSLOduration=1.576329648 podStartE2EDuration="4.185362025s" podCreationTimestamp="2026-02-02 12:23:19 +0000 UTC" firstStartedPulling="2026-02-02 12:23:19.970934012 +0000 UTC m=+831.199520875" lastFinishedPulling="2026-02-02 12:23:22.579966379 +0000 UTC m=+833.808553252" observedRunningTime="2026-02-02 12:23:23.179949854 +0000 UTC m=+834.408536717" watchObservedRunningTime="2026-02-02 12:23:23.185362025 +0000 UTC m=+834.413948888" Feb 02 12:23:23 crc kubenswrapper[4846]: I0202 12:23:23.201009 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-gpb7r" podStartSLOduration=1.653818605 podStartE2EDuration="4.200987246s" podCreationTimestamp="2026-02-02 12:23:19 +0000 UTC" firstStartedPulling="2026-02-02 12:23:20.02590262 +0000 UTC m=+831.254489483" lastFinishedPulling="2026-02-02 12:23:22.573071261 +0000 UTC m=+833.801658124" observedRunningTime="2026-02-02 12:23:23.195143654 +0000 UTC m=+834.423730527" watchObservedRunningTime="2026-02-02 12:23:23.200987246 +0000 UTC m=+834.429574109" Feb 02 12:23:27 crc kubenswrapper[4846]: I0202 12:23:27.186873 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-jzshh" event={"ID":"395d0b79-6975-4843-9a3f-f53121bfc1bc","Type":"ContainerStarted","Data":"8730bf82da664997f75a2e28c22370459bf2cca1eadbbfa242366768afc88263"} Feb 02 12:23:27 crc kubenswrapper[4846]: I0202 12:23:27.205488 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-jzshh" podStartSLOduration=1.982409643 podStartE2EDuration="8.20545372s" podCreationTimestamp="2026-02-02 12:23:19 +0000 UTC" firstStartedPulling="2026-02-02 12:23:19.935798797 +0000 UTC m=+831.164385660" lastFinishedPulling="2026-02-02 12:23:26.158842873 +0000 UTC m=+837.387429737" observedRunningTime="2026-02-02 12:23:27.202366945 +0000 UTC m=+838.430953848" watchObservedRunningTime="2026-02-02 12:23:27.20545372 +0000 UTC m=+838.434040623" Feb 02 12:23:29 crc kubenswrapper[4846]: I0202 12:23:29.763532 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-7t7k9" Feb 02 12:23:30 crc kubenswrapper[4846]: I0202 12:23:30.076541 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:30 crc kubenswrapper[4846]: I0202 12:23:30.076664 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:30 crc kubenswrapper[4846]: I0202 12:23:30.082310 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:30 crc kubenswrapper[4846]: I0202 12:23:30.212532 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-75bdbb7d58-f8649" Feb 02 12:23:30 crc kubenswrapper[4846]: I0202 12:23:30.265492 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-79skc"] Feb 02 12:23:39 crc kubenswrapper[4846]: I0202 12:23:39.723508 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-qhlrg" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.594474 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc"] Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.596218 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.600085 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.614787 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc"] Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.685335 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwtx\" (UniqueName: \"kubernetes.io/projected/e2281eb7-94f9-4e30-85c7-6c7a22385c16-kube-api-access-djwtx\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.685481 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.685523 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.787157 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djwtx\" (UniqueName: \"kubernetes.io/projected/e2281eb7-94f9-4e30-85c7-6c7a22385c16-kube-api-access-djwtx\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.787890 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.788036 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.788932 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.788944 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.809291 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djwtx\" (UniqueName: \"kubernetes.io/projected/e2281eb7-94f9-4e30-85c7-6c7a22385c16-kube-api-access-djwtx\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:51 crc kubenswrapper[4846]: I0202 12:23:51.967979 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:52 crc kubenswrapper[4846]: I0202 12:23:52.169988 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc"] Feb 02 12:23:52 crc kubenswrapper[4846]: I0202 12:23:52.348092 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" event={"ID":"e2281eb7-94f9-4e30-85c7-6c7a22385c16","Type":"ContainerStarted","Data":"6ff362df77a47676b6f40116116121eabb710fda238eb9e364c23f3e80bd9b69"} Feb 02 12:23:52 crc kubenswrapper[4846]: I0202 12:23:52.348140 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" event={"ID":"e2281eb7-94f9-4e30-85c7-6c7a22385c16","Type":"ContainerStarted","Data":"d4d3132a4815badea18fc8a5522b609008d0c649afc0951cbd2025658ae127a7"} Feb 02 12:23:53 crc kubenswrapper[4846]: I0202 12:23:53.378043 4846 generic.go:334] "Generic (PLEG): container finished" podID="e2281eb7-94f9-4e30-85c7-6c7a22385c16" containerID="6ff362df77a47676b6f40116116121eabb710fda238eb9e364c23f3e80bd9b69" exitCode=0 Feb 02 12:23:53 crc kubenswrapper[4846]: I0202 12:23:53.378080 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" event={"ID":"e2281eb7-94f9-4e30-85c7-6c7a22385c16","Type":"ContainerDied","Data":"6ff362df77a47676b6f40116116121eabb710fda238eb9e364c23f3e80bd9b69"} Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.310246 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-79skc" podUID="af51b9e4-0ee3-4bf9-93e5-27b4039be121" containerName="console" containerID="cri-o://15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5" gracePeriod=15 Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.394076 4846 generic.go:334] "Generic (PLEG): container finished" podID="e2281eb7-94f9-4e30-85c7-6c7a22385c16" containerID="83188df45d55f118cdfccf1b3527309c48eadf0b5bf4b786826d115656cdd38f" exitCode=0 Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.394133 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" event={"ID":"e2281eb7-94f9-4e30-85c7-6c7a22385c16","Type":"ContainerDied","Data":"83188df45d55f118cdfccf1b3527309c48eadf0b5bf4b786826d115656cdd38f"} Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.698647 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-79skc_af51b9e4-0ee3-4bf9-93e5-27b4039be121/console/0.log" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.698714 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.836794 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-service-ca\") pod \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.836842 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-config\") pod \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.836927 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-oauth-config\") pod \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.836954 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-478m7\" (UniqueName: \"kubernetes.io/projected/af51b9e4-0ee3-4bf9-93e5-27b4039be121-kube-api-access-478m7\") pod \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.837059 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-oauth-serving-cert\") pod \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.837158 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert\") pod \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.837193 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-trusted-ca-bundle\") pod \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\" (UID: \"af51b9e4-0ee3-4bf9-93e5-27b4039be121\") " Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.838195 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-service-ca" (OuterVolumeSpecName: "service-ca") pod "af51b9e4-0ee3-4bf9-93e5-27b4039be121" (UID: "af51b9e4-0ee3-4bf9-93e5-27b4039be121"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.838185 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "af51b9e4-0ee3-4bf9-93e5-27b4039be121" (UID: "af51b9e4-0ee3-4bf9-93e5-27b4039be121"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.838245 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "af51b9e4-0ee3-4bf9-93e5-27b4039be121" (UID: "af51b9e4-0ee3-4bf9-93e5-27b4039be121"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.838419 4846 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.838437 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.838448 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.838504 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-config" (OuterVolumeSpecName: "console-config") pod "af51b9e4-0ee3-4bf9-93e5-27b4039be121" (UID: "af51b9e4-0ee3-4bf9-93e5-27b4039be121"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.843500 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "af51b9e4-0ee3-4bf9-93e5-27b4039be121" (UID: "af51b9e4-0ee3-4bf9-93e5-27b4039be121"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.844175 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "af51b9e4-0ee3-4bf9-93e5-27b4039be121" (UID: "af51b9e4-0ee3-4bf9-93e5-27b4039be121"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.844754 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af51b9e4-0ee3-4bf9-93e5-27b4039be121-kube-api-access-478m7" (OuterVolumeSpecName: "kube-api-access-478m7") pod "af51b9e4-0ee3-4bf9-93e5-27b4039be121" (UID: "af51b9e4-0ee3-4bf9-93e5-27b4039be121"). InnerVolumeSpecName "kube-api-access-478m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.939597 4846 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.939879 4846 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.939957 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-478m7\" (UniqueName: \"kubernetes.io/projected/af51b9e4-0ee3-4bf9-93e5-27b4039be121-kube-api-access-478m7\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:55 crc kubenswrapper[4846]: I0202 12:23:55.940030 4846 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af51b9e4-0ee3-4bf9-93e5-27b4039be121-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.401885 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-79skc_af51b9e4-0ee3-4bf9-93e5-27b4039be121/console/0.log" Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.403160 4846 generic.go:334] "Generic (PLEG): container finished" podID="af51b9e4-0ee3-4bf9-93e5-27b4039be121" containerID="15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5" exitCode=2 Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.403283 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-79skc" event={"ID":"af51b9e4-0ee3-4bf9-93e5-27b4039be121","Type":"ContainerDied","Data":"15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5"} Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.403384 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-79skc" event={"ID":"af51b9e4-0ee3-4bf9-93e5-27b4039be121","Type":"ContainerDied","Data":"3793dbe275be8eb1c21e9d5d00438ea825623b5a249f9e167b1c04f4a30a45cb"} Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.403267 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-79skc" Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.403445 4846 scope.go:117] "RemoveContainer" containerID="15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5" Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.406473 4846 generic.go:334] "Generic (PLEG): container finished" podID="e2281eb7-94f9-4e30-85c7-6c7a22385c16" containerID="e81e7429a8e7713d7f1cfcf4f6ee011fb7ff3ea1563baa163d5aa69bd2a5a8d7" exitCode=0 Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.406595 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" event={"ID":"e2281eb7-94f9-4e30-85c7-6c7a22385c16","Type":"ContainerDied","Data":"e81e7429a8e7713d7f1cfcf4f6ee011fb7ff3ea1563baa163d5aa69bd2a5a8d7"} Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.427788 4846 scope.go:117] "RemoveContainer" containerID="15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5" Feb 02 12:23:56 crc kubenswrapper[4846]: E0202 12:23:56.428370 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5\": container with ID starting with 15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5 not found: ID does not exist" containerID="15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5" Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.428398 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5"} err="failed to get container status \"15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5\": rpc error: code = NotFound desc = could not find container \"15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5\": container with ID starting with 15e7ebcdd327e6465ac07a6495e268a013623ebba4577b11fc0da06b49f074e5 not found: ID does not exist" Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.452546 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-79skc"] Feb 02 12:23:56 crc kubenswrapper[4846]: I0202 12:23:56.457319 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-79skc"] Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.430556 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af51b9e4-0ee3-4bf9-93e5-27b4039be121" path="/var/lib/kubelet/pods/af51b9e4-0ee3-4bf9-93e5-27b4039be121/volumes" Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.654812 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.762828 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-util\") pod \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.762884 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djwtx\" (UniqueName: \"kubernetes.io/projected/e2281eb7-94f9-4e30-85c7-6c7a22385c16-kube-api-access-djwtx\") pod \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.762927 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-bundle\") pod \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\" (UID: \"e2281eb7-94f9-4e30-85c7-6c7a22385c16\") " Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.764215 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-bundle" (OuterVolumeSpecName: "bundle") pod "e2281eb7-94f9-4e30-85c7-6c7a22385c16" (UID: "e2281eb7-94f9-4e30-85c7-6c7a22385c16"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.768110 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2281eb7-94f9-4e30-85c7-6c7a22385c16-kube-api-access-djwtx" (OuterVolumeSpecName: "kube-api-access-djwtx") pod "e2281eb7-94f9-4e30-85c7-6c7a22385c16" (UID: "e2281eb7-94f9-4e30-85c7-6c7a22385c16"). InnerVolumeSpecName "kube-api-access-djwtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.786161 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-util" (OuterVolumeSpecName: "util") pod "e2281eb7-94f9-4e30-85c7-6c7a22385c16" (UID: "e2281eb7-94f9-4e30-85c7-6c7a22385c16"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.864326 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-util\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.864370 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djwtx\" (UniqueName: \"kubernetes.io/projected/e2281eb7-94f9-4e30-85c7-6c7a22385c16-kube-api-access-djwtx\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:57 crc kubenswrapper[4846]: I0202 12:23:57.864392 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e2281eb7-94f9-4e30-85c7-6c7a22385c16-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:23:58 crc kubenswrapper[4846]: I0202 12:23:58.422772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" event={"ID":"e2281eb7-94f9-4e30-85c7-6c7a22385c16","Type":"ContainerDied","Data":"d4d3132a4815badea18fc8a5522b609008d0c649afc0951cbd2025658ae127a7"} Feb 02 12:23:58 crc kubenswrapper[4846]: I0202 12:23:58.422809 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4d3132a4815badea18fc8a5522b609008d0c649afc0951cbd2025658ae127a7" Feb 02 12:23:58 crc kubenswrapper[4846]: I0202 12:23:58.422836 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc" Feb 02 12:24:00 crc kubenswrapper[4846]: I0202 12:24:00.479421 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:24:00 crc kubenswrapper[4846]: I0202 12:24:00.481076 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.503870 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t"] Feb 02 12:24:06 crc kubenswrapper[4846]: E0202 12:24:06.504343 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2281eb7-94f9-4e30-85c7-6c7a22385c16" containerName="pull" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.504354 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2281eb7-94f9-4e30-85c7-6c7a22385c16" containerName="pull" Feb 02 12:24:06 crc kubenswrapper[4846]: E0202 12:24:06.504364 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2281eb7-94f9-4e30-85c7-6c7a22385c16" containerName="util" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.504370 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2281eb7-94f9-4e30-85c7-6c7a22385c16" containerName="util" Feb 02 12:24:06 crc kubenswrapper[4846]: E0202 12:24:06.504385 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2281eb7-94f9-4e30-85c7-6c7a22385c16" containerName="extract" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.504392 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2281eb7-94f9-4e30-85c7-6c7a22385c16" containerName="extract" Feb 02 12:24:06 crc kubenswrapper[4846]: E0202 12:24:06.504404 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af51b9e4-0ee3-4bf9-93e5-27b4039be121" containerName="console" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.504410 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="af51b9e4-0ee3-4bf9-93e5-27b4039be121" containerName="console" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.504516 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="af51b9e4-0ee3-4bf9-93e5-27b4039be121" containerName="console" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.504529 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2281eb7-94f9-4e30-85c7-6c7a22385c16" containerName="extract" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.504889 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.507703 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.507785 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-tzskg" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.507919 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.510532 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.512426 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.530989 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t"] Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.570525 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a520c385-fbf3-4941-968c-2163e685af23-webhook-cert\") pod \"metallb-operator-controller-manager-5b7785ffb5-mxf8t\" (UID: \"a520c385-fbf3-4941-968c-2163e685af23\") " pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.570580 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmvwz\" (UniqueName: \"kubernetes.io/projected/a520c385-fbf3-4941-968c-2163e685af23-kube-api-access-dmvwz\") pod \"metallb-operator-controller-manager-5b7785ffb5-mxf8t\" (UID: \"a520c385-fbf3-4941-968c-2163e685af23\") " pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.570611 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a520c385-fbf3-4941-968c-2163e685af23-apiservice-cert\") pod \"metallb-operator-controller-manager-5b7785ffb5-mxf8t\" (UID: \"a520c385-fbf3-4941-968c-2163e685af23\") " pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.672296 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a520c385-fbf3-4941-968c-2163e685af23-webhook-cert\") pod \"metallb-operator-controller-manager-5b7785ffb5-mxf8t\" (UID: \"a520c385-fbf3-4941-968c-2163e685af23\") " pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.672564 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmvwz\" (UniqueName: \"kubernetes.io/projected/a520c385-fbf3-4941-968c-2163e685af23-kube-api-access-dmvwz\") pod \"metallb-operator-controller-manager-5b7785ffb5-mxf8t\" (UID: \"a520c385-fbf3-4941-968c-2163e685af23\") " pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.672680 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a520c385-fbf3-4941-968c-2163e685af23-apiservice-cert\") pod \"metallb-operator-controller-manager-5b7785ffb5-mxf8t\" (UID: \"a520c385-fbf3-4941-968c-2163e685af23\") " pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.681035 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a520c385-fbf3-4941-968c-2163e685af23-webhook-cert\") pod \"metallb-operator-controller-manager-5b7785ffb5-mxf8t\" (UID: \"a520c385-fbf3-4941-968c-2163e685af23\") " pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.687241 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a520c385-fbf3-4941-968c-2163e685af23-apiservice-cert\") pod \"metallb-operator-controller-manager-5b7785ffb5-mxf8t\" (UID: \"a520c385-fbf3-4941-968c-2163e685af23\") " pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.690470 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmvwz\" (UniqueName: \"kubernetes.io/projected/a520c385-fbf3-4941-968c-2163e685af23-kube-api-access-dmvwz\") pod \"metallb-operator-controller-manager-5b7785ffb5-mxf8t\" (UID: \"a520c385-fbf3-4941-968c-2163e685af23\") " pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.822467 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.956324 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9"] Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.957427 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.963191 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.963299 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.963965 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-zcmw8" Feb 02 12:24:06 crc kubenswrapper[4846]: I0202 12:24:06.986675 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9"] Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.077073 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4efa0217-6d2e-4857-8585-559854a1df0f-apiservice-cert\") pod \"metallb-operator-webhook-server-76d4f8c956-swps9\" (UID: \"4efa0217-6d2e-4857-8585-559854a1df0f\") " pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.077125 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtcrk\" (UniqueName: \"kubernetes.io/projected/4efa0217-6d2e-4857-8585-559854a1df0f-kube-api-access-gtcrk\") pod \"metallb-operator-webhook-server-76d4f8c956-swps9\" (UID: \"4efa0217-6d2e-4857-8585-559854a1df0f\") " pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.077240 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4efa0217-6d2e-4857-8585-559854a1df0f-webhook-cert\") pod \"metallb-operator-webhook-server-76d4f8c956-swps9\" (UID: \"4efa0217-6d2e-4857-8585-559854a1df0f\") " pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.178454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4efa0217-6d2e-4857-8585-559854a1df0f-webhook-cert\") pod \"metallb-operator-webhook-server-76d4f8c956-swps9\" (UID: \"4efa0217-6d2e-4857-8585-559854a1df0f\") " pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.178544 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4efa0217-6d2e-4857-8585-559854a1df0f-apiservice-cert\") pod \"metallb-operator-webhook-server-76d4f8c956-swps9\" (UID: \"4efa0217-6d2e-4857-8585-559854a1df0f\") " pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.178576 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtcrk\" (UniqueName: \"kubernetes.io/projected/4efa0217-6d2e-4857-8585-559854a1df0f-kube-api-access-gtcrk\") pod \"metallb-operator-webhook-server-76d4f8c956-swps9\" (UID: \"4efa0217-6d2e-4857-8585-559854a1df0f\") " pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.184197 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4efa0217-6d2e-4857-8585-559854a1df0f-webhook-cert\") pod \"metallb-operator-webhook-server-76d4f8c956-swps9\" (UID: \"4efa0217-6d2e-4857-8585-559854a1df0f\") " pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.184231 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4efa0217-6d2e-4857-8585-559854a1df0f-apiservice-cert\") pod \"metallb-operator-webhook-server-76d4f8c956-swps9\" (UID: \"4efa0217-6d2e-4857-8585-559854a1df0f\") " pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.202337 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtcrk\" (UniqueName: \"kubernetes.io/projected/4efa0217-6d2e-4857-8585-559854a1df0f-kube-api-access-gtcrk\") pod \"metallb-operator-webhook-server-76d4f8c956-swps9\" (UID: \"4efa0217-6d2e-4857-8585-559854a1df0f\") " pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.307187 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.326559 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t"] Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.491338 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" event={"ID":"a520c385-fbf3-4941-968c-2163e685af23","Type":"ContainerStarted","Data":"07bc8183ff323f017e609d074f4ce5c93c0285f0659ba5132cca1b33c7262c2c"} Feb 02 12:24:07 crc kubenswrapper[4846]: I0202 12:24:07.574648 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9"] Feb 02 12:24:07 crc kubenswrapper[4846]: W0202 12:24:07.581299 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4efa0217_6d2e_4857_8585_559854a1df0f.slice/crio-19706dcdb3af94dcf558e441b2cee3f17ad24d26bc0ccb8bed42b3439315031b WatchSource:0}: Error finding container 19706dcdb3af94dcf558e441b2cee3f17ad24d26bc0ccb8bed42b3439315031b: Status 404 returned error can't find the container with id 19706dcdb3af94dcf558e441b2cee3f17ad24d26bc0ccb8bed42b3439315031b Feb 02 12:24:08 crc kubenswrapper[4846]: I0202 12:24:08.530825 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" event={"ID":"4efa0217-6d2e-4857-8585-559854a1df0f","Type":"ContainerStarted","Data":"19706dcdb3af94dcf558e441b2cee3f17ad24d26bc0ccb8bed42b3439315031b"} Feb 02 12:24:12 crc kubenswrapper[4846]: I0202 12:24:12.562012 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" event={"ID":"a520c385-fbf3-4941-968c-2163e685af23","Type":"ContainerStarted","Data":"24f9718efa24a8b4d1eb22f94abfb6efef6a0ed63e0cfb74ea6f2771ee6b48ff"} Feb 02 12:24:12 crc kubenswrapper[4846]: I0202 12:24:12.562643 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:12 crc kubenswrapper[4846]: I0202 12:24:12.563494 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" event={"ID":"4efa0217-6d2e-4857-8585-559854a1df0f","Type":"ContainerStarted","Data":"ac4a3d8b02cea97424b7b9ad79ab78f4e4241284faf2c685f86a1da78adfb789"} Feb 02 12:24:12 crc kubenswrapper[4846]: I0202 12:24:12.563741 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:12 crc kubenswrapper[4846]: I0202 12:24:12.590816 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" podStartSLOduration=2.000853554 podStartE2EDuration="6.590801585s" podCreationTimestamp="2026-02-02 12:24:06 +0000 UTC" firstStartedPulling="2026-02-02 12:24:07.338229133 +0000 UTC m=+878.566815996" lastFinishedPulling="2026-02-02 12:24:11.928177164 +0000 UTC m=+883.156764027" observedRunningTime="2026-02-02 12:24:12.588522569 +0000 UTC m=+883.817109442" watchObservedRunningTime="2026-02-02 12:24:12.590801585 +0000 UTC m=+883.819388448" Feb 02 12:24:12 crc kubenswrapper[4846]: I0202 12:24:12.610071 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" podStartSLOduration=2.244645069 podStartE2EDuration="6.610052636s" podCreationTimestamp="2026-02-02 12:24:06 +0000 UTC" firstStartedPulling="2026-02-02 12:24:07.584535149 +0000 UTC m=+878.813122052" lastFinishedPulling="2026-02-02 12:24:11.949942756 +0000 UTC m=+883.178529619" observedRunningTime="2026-02-02 12:24:12.608218811 +0000 UTC m=+883.836805684" watchObservedRunningTime="2026-02-02 12:24:12.610052636 +0000 UTC m=+883.838639509" Feb 02 12:24:27 crc kubenswrapper[4846]: I0202 12:24:27.313094 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-76d4f8c956-swps9" Feb 02 12:24:30 crc kubenswrapper[4846]: I0202 12:24:30.479593 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:24:30 crc kubenswrapper[4846]: I0202 12:24:30.480012 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:24:46 crc kubenswrapper[4846]: I0202 12:24:46.825353 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5b7785ffb5-mxf8t" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.458423 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-r692g"] Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.461311 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.463810 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.466153 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.466387 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-x5b7h" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.476188 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw"] Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.477122 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.478685 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.498707 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw"] Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.563274 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-5lpq4"] Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.564359 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.569260 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-p7fn9" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.569389 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.569643 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.569783 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.581819 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-lrmwh"] Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.582874 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.585760 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.586663 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx5hf\" (UniqueName: \"kubernetes.io/projected/162e27cd-61d6-4f44-801c-775077032ce7-kube-api-access-bx5hf\") pod \"frr-k8s-webhook-server-7df86c4f6c-2r6kw\" (UID: \"162e27cd-61d6-4f44-801c-775077032ce7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.587319 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7btmz\" (UniqueName: \"kubernetes.io/projected/ec7b6ba3-8411-4248-ad6e-d684293002c4-kube-api-access-7btmz\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.587357 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec7b6ba3-8411-4248-ad6e-d684293002c4-metrics-certs\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.587391 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-frr-conf\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.587422 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/162e27cd-61d6-4f44-801c-775077032ce7-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-2r6kw\" (UID: \"162e27cd-61d6-4f44-801c-775077032ce7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.587480 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-reloader\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.587613 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-metrics\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.587658 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-frr-sockets\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.587692 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ec7b6ba3-8411-4248-ad6e-d684293002c4-frr-startup\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.602499 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-lrmwh"] Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.688978 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec7b6ba3-8411-4248-ad6e-d684293002c4-metrics-certs\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689031 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7btmz\" (UniqueName: \"kubernetes.io/projected/ec7b6ba3-8411-4248-ad6e-d684293002c4-kube-api-access-7btmz\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689061 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-frr-conf\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689090 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/162e27cd-61d6-4f44-801c-775077032ce7-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-2r6kw\" (UID: \"162e27cd-61d6-4f44-801c-775077032ce7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689116 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-reloader\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689137 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-metrics\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689156 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-frr-sockets\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689186 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ec7b6ba3-8411-4248-ad6e-d684293002c4-frr-startup\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689219 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/77e7c4a3-0018-438a-9163-5a155ab480d1-metallb-excludel2\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689250 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2fxl\" (UniqueName: \"kubernetes.io/projected/77e7c4a3-0018-438a-9163-5a155ab480d1-kube-api-access-p2fxl\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689288 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx5hf\" (UniqueName: \"kubernetes.io/projected/162e27cd-61d6-4f44-801c-775077032ce7-kube-api-access-bx5hf\") pod \"frr-k8s-webhook-server-7df86c4f6c-2r6kw\" (UID: \"162e27cd-61d6-4f44-801c-775077032ce7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689311 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpl5s\" (UniqueName: \"kubernetes.io/projected/529cfe45-6776-47c3-8cc4-0c5cbcb6e130-kube-api-access-tpl5s\") pod \"controller-6968d8fdc4-lrmwh\" (UID: \"529cfe45-6776-47c3-8cc4-0c5cbcb6e130\") " pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689342 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/529cfe45-6776-47c3-8cc4-0c5cbcb6e130-cert\") pod \"controller-6968d8fdc4-lrmwh\" (UID: \"529cfe45-6776-47c3-8cc4-0c5cbcb6e130\") " pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689362 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/529cfe45-6776-47c3-8cc4-0c5cbcb6e130-metrics-certs\") pod \"controller-6968d8fdc4-lrmwh\" (UID: \"529cfe45-6776-47c3-8cc4-0c5cbcb6e130\") " pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689391 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-metrics-certs\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.689414 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-memberlist\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.690147 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ec7b6ba3-8411-4248-ad6e-d684293002c4-frr-startup\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.690476 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-reloader\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.690611 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-frr-sockets\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.690724 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-metrics\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.691032 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ec7b6ba3-8411-4248-ad6e-d684293002c4-frr-conf\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.695339 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/162e27cd-61d6-4f44-801c-775077032ce7-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-2r6kw\" (UID: \"162e27cd-61d6-4f44-801c-775077032ce7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.698833 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec7b6ba3-8411-4248-ad6e-d684293002c4-metrics-certs\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.706525 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx5hf\" (UniqueName: \"kubernetes.io/projected/162e27cd-61d6-4f44-801c-775077032ce7-kube-api-access-bx5hf\") pod \"frr-k8s-webhook-server-7df86c4f6c-2r6kw\" (UID: \"162e27cd-61d6-4f44-801c-775077032ce7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.709517 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7btmz\" (UniqueName: \"kubernetes.io/projected/ec7b6ba3-8411-4248-ad6e-d684293002c4-kube-api-access-7btmz\") pod \"frr-k8s-r692g\" (UID: \"ec7b6ba3-8411-4248-ad6e-d684293002c4\") " pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.780026 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.790612 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-metrics-certs\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.790693 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-memberlist\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.790767 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/77e7c4a3-0018-438a-9163-5a155ab480d1-metallb-excludel2\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.790797 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2fxl\" (UniqueName: \"kubernetes.io/projected/77e7c4a3-0018-438a-9163-5a155ab480d1-kube-api-access-p2fxl\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: E0202 12:24:47.790807 4846 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.790829 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpl5s\" (UniqueName: \"kubernetes.io/projected/529cfe45-6776-47c3-8cc4-0c5cbcb6e130-kube-api-access-tpl5s\") pod \"controller-6968d8fdc4-lrmwh\" (UID: \"529cfe45-6776-47c3-8cc4-0c5cbcb6e130\") " pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.790857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/529cfe45-6776-47c3-8cc4-0c5cbcb6e130-cert\") pod \"controller-6968d8fdc4-lrmwh\" (UID: \"529cfe45-6776-47c3-8cc4-0c5cbcb6e130\") " pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:47 crc kubenswrapper[4846]: E0202 12:24:47.790891 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-metrics-certs podName:77e7c4a3-0018-438a-9163-5a155ab480d1 nodeName:}" failed. No retries permitted until 2026-02-02 12:24:48.290865826 +0000 UTC m=+919.519452749 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-metrics-certs") pod "speaker-5lpq4" (UID: "77e7c4a3-0018-438a-9163-5a155ab480d1") : secret "speaker-certs-secret" not found Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.790924 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/529cfe45-6776-47c3-8cc4-0c5cbcb6e130-metrics-certs\") pod \"controller-6968d8fdc4-lrmwh\" (UID: \"529cfe45-6776-47c3-8cc4-0c5cbcb6e130\") " pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:47 crc kubenswrapper[4846]: E0202 12:24:47.791009 4846 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 02 12:24:47 crc kubenswrapper[4846]: E0202 12:24:47.791067 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-memberlist podName:77e7c4a3-0018-438a-9163-5a155ab480d1 nodeName:}" failed. No retries permitted until 2026-02-02 12:24:48.291047951 +0000 UTC m=+919.519634904 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-memberlist") pod "speaker-5lpq4" (UID: "77e7c4a3-0018-438a-9163-5a155ab480d1") : secret "metallb-memberlist" not found Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.791763 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/77e7c4a3-0018-438a-9163-5a155ab480d1-metallb-excludel2\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.792405 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.796446 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/529cfe45-6776-47c3-8cc4-0c5cbcb6e130-metrics-certs\") pod \"controller-6968d8fdc4-lrmwh\" (UID: \"529cfe45-6776-47c3-8cc4-0c5cbcb6e130\") " pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.797809 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.804567 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/529cfe45-6776-47c3-8cc4-0c5cbcb6e130-cert\") pod \"controller-6968d8fdc4-lrmwh\" (UID: \"529cfe45-6776-47c3-8cc4-0c5cbcb6e130\") " pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.812813 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpl5s\" (UniqueName: \"kubernetes.io/projected/529cfe45-6776-47c3-8cc4-0c5cbcb6e130-kube-api-access-tpl5s\") pod \"controller-6968d8fdc4-lrmwh\" (UID: \"529cfe45-6776-47c3-8cc4-0c5cbcb6e130\") " pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.813022 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2fxl\" (UniqueName: \"kubernetes.io/projected/77e7c4a3-0018-438a-9163-5a155ab480d1-kube-api-access-p2fxl\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:47 crc kubenswrapper[4846]: I0202 12:24:47.898540 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.085527 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw"] Feb 02 12:24:48 crc kubenswrapper[4846]: W0202 12:24:48.089107 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod162e27cd_61d6_4f44_801c_775077032ce7.slice/crio-ed004516da42be8c5684dc49ba5ccb99d71b59cc906122a6204b73d5919ccc27 WatchSource:0}: Error finding container ed004516da42be8c5684dc49ba5ccb99d71b59cc906122a6204b73d5919ccc27: Status 404 returned error can't find the container with id ed004516da42be8c5684dc49ba5ccb99d71b59cc906122a6204b73d5919ccc27 Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.194203 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-lrmwh"] Feb 02 12:24:48 crc kubenswrapper[4846]: W0202 12:24:48.197283 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod529cfe45_6776_47c3_8cc4_0c5cbcb6e130.slice/crio-51a2ba28f8b2522e330d7ed3ade1c01cc52b626fbb0fcc5f24318e45b81e6884 WatchSource:0}: Error finding container 51a2ba28f8b2522e330d7ed3ade1c01cc52b626fbb0fcc5f24318e45b81e6884: Status 404 returned error can't find the container with id 51a2ba28f8b2522e330d7ed3ade1c01cc52b626fbb0fcc5f24318e45b81e6884 Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.310324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-metrics-certs\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.310474 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-memberlist\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:48 crc kubenswrapper[4846]: E0202 12:24:48.311172 4846 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 02 12:24:48 crc kubenswrapper[4846]: E0202 12:24:48.311246 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-memberlist podName:77e7c4a3-0018-438a-9163-5a155ab480d1 nodeName:}" failed. No retries permitted until 2026-02-02 12:24:49.311226057 +0000 UTC m=+920.539812920 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-memberlist") pod "speaker-5lpq4" (UID: "77e7c4a3-0018-438a-9163-5a155ab480d1") : secret "metallb-memberlist" not found Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.322509 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-metrics-certs\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.758393 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-lrmwh" event={"ID":"529cfe45-6776-47c3-8cc4-0c5cbcb6e130","Type":"ContainerStarted","Data":"53a010f6e4332e4f7f62d2a711b86bcfab4f9a33dac2e9b595af91292fd27734"} Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.758432 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-lrmwh" event={"ID":"529cfe45-6776-47c3-8cc4-0c5cbcb6e130","Type":"ContainerStarted","Data":"38909c62b62a53b7c3ab1e95c566a9f35fed45bdda072e97b7552c5467ac6749"} Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.758441 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-lrmwh" event={"ID":"529cfe45-6776-47c3-8cc4-0c5cbcb6e130","Type":"ContainerStarted","Data":"51a2ba28f8b2522e330d7ed3ade1c01cc52b626fbb0fcc5f24318e45b81e6884"} Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.758482 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.760958 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r692g" event={"ID":"ec7b6ba3-8411-4248-ad6e-d684293002c4","Type":"ContainerStarted","Data":"5d954f9234965109d11afc06cedc1ec07e028a052f255783827da95e367ba910"} Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.761860 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" event={"ID":"162e27cd-61d6-4f44-801c-775077032ce7","Type":"ContainerStarted","Data":"ed004516da42be8c5684dc49ba5ccb99d71b59cc906122a6204b73d5919ccc27"} Feb 02 12:24:48 crc kubenswrapper[4846]: I0202 12:24:48.777305 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-lrmwh" podStartSLOduration=1.777286189 podStartE2EDuration="1.777286189s" podCreationTimestamp="2026-02-02 12:24:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:24:48.773609189 +0000 UTC m=+920.002196072" watchObservedRunningTime="2026-02-02 12:24:48.777286189 +0000 UTC m=+920.005873052" Feb 02 12:24:49 crc kubenswrapper[4846]: I0202 12:24:49.325014 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-memberlist\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:49 crc kubenswrapper[4846]: I0202 12:24:49.331412 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/77e7c4a3-0018-438a-9163-5a155ab480d1-memberlist\") pod \"speaker-5lpq4\" (UID: \"77e7c4a3-0018-438a-9163-5a155ab480d1\") " pod="metallb-system/speaker-5lpq4" Feb 02 12:24:49 crc kubenswrapper[4846]: I0202 12:24:49.380840 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5lpq4" Feb 02 12:24:49 crc kubenswrapper[4846]: W0202 12:24:49.406951 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77e7c4a3_0018_438a_9163_5a155ab480d1.slice/crio-39fce2da1931e1217fb81e54e792f3d6fc7cde93ecf9e803be95f06307bbe9cf WatchSource:0}: Error finding container 39fce2da1931e1217fb81e54e792f3d6fc7cde93ecf9e803be95f06307bbe9cf: Status 404 returned error can't find the container with id 39fce2da1931e1217fb81e54e792f3d6fc7cde93ecf9e803be95f06307bbe9cf Feb 02 12:24:49 crc kubenswrapper[4846]: I0202 12:24:49.771037 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5lpq4" event={"ID":"77e7c4a3-0018-438a-9163-5a155ab480d1","Type":"ContainerStarted","Data":"024efdf90a83f02ecb77b492d6e1a32f4d08967967fdeaa27c504c50cbc84adc"} Feb 02 12:24:49 crc kubenswrapper[4846]: I0202 12:24:49.771086 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5lpq4" event={"ID":"77e7c4a3-0018-438a-9163-5a155ab480d1","Type":"ContainerStarted","Data":"39fce2da1931e1217fb81e54e792f3d6fc7cde93ecf9e803be95f06307bbe9cf"} Feb 02 12:24:50 crc kubenswrapper[4846]: I0202 12:24:50.787417 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5lpq4" event={"ID":"77e7c4a3-0018-438a-9163-5a155ab480d1","Type":"ContainerStarted","Data":"99f178c4701bb91d216cb9d1fb62d0454f02452b8dc34c17b85897890f68ed59"} Feb 02 12:24:50 crc kubenswrapper[4846]: I0202 12:24:50.788658 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5lpq4" Feb 02 12:24:50 crc kubenswrapper[4846]: I0202 12:24:50.817076 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-5lpq4" podStartSLOduration=3.817054561 podStartE2EDuration="3.817054561s" podCreationTimestamp="2026-02-02 12:24:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:24:50.814719833 +0000 UTC m=+922.043306706" watchObservedRunningTime="2026-02-02 12:24:50.817054561 +0000 UTC m=+922.045641424" Feb 02 12:24:55 crc kubenswrapper[4846]: I0202 12:24:55.822277 4846 generic.go:334] "Generic (PLEG): container finished" podID="ec7b6ba3-8411-4248-ad6e-d684293002c4" containerID="4d48d950663da166fbfa34689c18e83ae7c914b86074182d1c9fe068ab1a8e59" exitCode=0 Feb 02 12:24:55 crc kubenswrapper[4846]: I0202 12:24:55.822494 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r692g" event={"ID":"ec7b6ba3-8411-4248-ad6e-d684293002c4","Type":"ContainerDied","Data":"4d48d950663da166fbfa34689c18e83ae7c914b86074182d1c9fe068ab1a8e59"} Feb 02 12:24:55 crc kubenswrapper[4846]: I0202 12:24:55.825992 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" event={"ID":"162e27cd-61d6-4f44-801c-775077032ce7","Type":"ContainerStarted","Data":"dffeca5aff10a02ce70d4c136df0c5cb2278e0b33b941dc84e140c153db24133"} Feb 02 12:24:55 crc kubenswrapper[4846]: I0202 12:24:55.826168 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" Feb 02 12:24:55 crc kubenswrapper[4846]: I0202 12:24:55.868646 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" podStartSLOduration=1.690614578 podStartE2EDuration="8.868613183s" podCreationTimestamp="2026-02-02 12:24:47 +0000 UTC" firstStartedPulling="2026-02-02 12:24:48.090882736 +0000 UTC m=+919.319469599" lastFinishedPulling="2026-02-02 12:24:55.268881321 +0000 UTC m=+926.497468204" observedRunningTime="2026-02-02 12:24:55.867099706 +0000 UTC m=+927.095686569" watchObservedRunningTime="2026-02-02 12:24:55.868613183 +0000 UTC m=+927.097200046" Feb 02 12:24:56 crc kubenswrapper[4846]: I0202 12:24:56.836812 4846 generic.go:334] "Generic (PLEG): container finished" podID="ec7b6ba3-8411-4248-ad6e-d684293002c4" containerID="7ae42f78f183d64af58506122c3c403b043bf246a22b053be15cc11d60a57347" exitCode=0 Feb 02 12:24:56 crc kubenswrapper[4846]: I0202 12:24:56.836942 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r692g" event={"ID":"ec7b6ba3-8411-4248-ad6e-d684293002c4","Type":"ContainerDied","Data":"7ae42f78f183d64af58506122c3c403b043bf246a22b053be15cc11d60a57347"} Feb 02 12:24:57 crc kubenswrapper[4846]: I0202 12:24:57.843687 4846 generic.go:334] "Generic (PLEG): container finished" podID="ec7b6ba3-8411-4248-ad6e-d684293002c4" containerID="11b30d6917a07f3b7f6ca64aadcc2cbef7b9bdcbac0809076d2b276cb65e18a3" exitCode=0 Feb 02 12:24:57 crc kubenswrapper[4846]: I0202 12:24:57.843727 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r692g" event={"ID":"ec7b6ba3-8411-4248-ad6e-d684293002c4","Type":"ContainerDied","Data":"11b30d6917a07f3b7f6ca64aadcc2cbef7b9bdcbac0809076d2b276cb65e18a3"} Feb 02 12:24:58 crc kubenswrapper[4846]: I0202 12:24:58.854439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r692g" event={"ID":"ec7b6ba3-8411-4248-ad6e-d684293002c4","Type":"ContainerStarted","Data":"7fe2945591e5408c02c3fb22b229b84aa04843148574275ae6add98bc5afd708"} Feb 02 12:24:58 crc kubenswrapper[4846]: I0202 12:24:58.854807 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r692g" event={"ID":"ec7b6ba3-8411-4248-ad6e-d684293002c4","Type":"ContainerStarted","Data":"57ca0eadfd59663eec3eff4cc8042a51a8d31e4698b71617ec678984b30ef61f"} Feb 02 12:24:58 crc kubenswrapper[4846]: I0202 12:24:58.854820 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r692g" event={"ID":"ec7b6ba3-8411-4248-ad6e-d684293002c4","Type":"ContainerStarted","Data":"17e98b28c691c82bb678f8590fd1ad9e509a4d4421abd89b4580edec1f5345da"} Feb 02 12:24:58 crc kubenswrapper[4846]: I0202 12:24:58.854829 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r692g" event={"ID":"ec7b6ba3-8411-4248-ad6e-d684293002c4","Type":"ContainerStarted","Data":"7eaa41f113d139aad0cf89fbf5589d9f9826f9b0299a580ad17e971531a8769f"} Feb 02 12:24:58 crc kubenswrapper[4846]: I0202 12:24:58.854837 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r692g" event={"ID":"ec7b6ba3-8411-4248-ad6e-d684293002c4","Type":"ContainerStarted","Data":"8a5a945f946d45e9d7298a3ad66a828531f019d34f0b1917d541cb30191f1368"} Feb 02 12:24:59 crc kubenswrapper[4846]: I0202 12:24:59.385111 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-5lpq4" Feb 02 12:24:59 crc kubenswrapper[4846]: I0202 12:24:59.870500 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r692g" event={"ID":"ec7b6ba3-8411-4248-ad6e-d684293002c4","Type":"ContainerStarted","Data":"e3a9e801294247507686601d22d8724c6440f3eaf8aefbb003ee42d00403f442"} Feb 02 12:24:59 crc kubenswrapper[4846]: I0202 12:24:59.871941 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-r692g" Feb 02 12:24:59 crc kubenswrapper[4846]: I0202 12:24:59.911147 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-r692g" podStartSLOduration=5.630816644 podStartE2EDuration="12.911118651s" podCreationTimestamp="2026-02-02 12:24:47 +0000 UTC" firstStartedPulling="2026-02-02 12:24:48.005785835 +0000 UTC m=+919.234372698" lastFinishedPulling="2026-02-02 12:24:55.286087842 +0000 UTC m=+926.514674705" observedRunningTime="2026-02-02 12:24:59.910908935 +0000 UTC m=+931.139495808" watchObservedRunningTime="2026-02-02 12:24:59.911118651 +0000 UTC m=+931.139705514" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.478823 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.478881 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.478922 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.479477 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06b5f4ddd20a5b1b0ef8c848baafd57c963bd52aab6eccb2c47d76fc10efc10c"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.479575 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://06b5f4ddd20a5b1b0ef8c848baafd57c963bd52aab6eccb2c47d76fc10efc10c" gracePeriod=600 Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.700814 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q"] Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.702313 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.704093 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.709504 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q"] Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.880195 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="06b5f4ddd20a5b1b0ef8c848baafd57c963bd52aab6eccb2c47d76fc10efc10c" exitCode=0 Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.880260 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"06b5f4ddd20a5b1b0ef8c848baafd57c963bd52aab6eccb2c47d76fc10efc10c"} Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.880302 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"d4f60d595c3f3559b0b02374c951ace4cd6f20670671fc67c4b4653981928977"} Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.880323 4846 scope.go:117] "RemoveContainer" containerID="025493c8107e9680bfbd487f264aead8432b3b590da7fb314c8c0774a547036f" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.886956 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49g5b\" (UniqueName: \"kubernetes.io/projected/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-kube-api-access-49g5b\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.887061 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.887118 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.988818 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49g5b\" (UniqueName: \"kubernetes.io/projected/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-kube-api-access-49g5b\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.988952 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.989006 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.989774 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:00 crc kubenswrapper[4846]: I0202 12:25:00.990306 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:01 crc kubenswrapper[4846]: I0202 12:25:01.012678 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49g5b\" (UniqueName: \"kubernetes.io/projected/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-kube-api-access-49g5b\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:01 crc kubenswrapper[4846]: I0202 12:25:01.028153 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:01 crc kubenswrapper[4846]: I0202 12:25:01.634035 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q"] Feb 02 12:25:01 crc kubenswrapper[4846]: W0202 12:25:01.641825 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9aa5a1ea_0fe6_4b98_adf6_a6af93581534.slice/crio-80889c062fa7f90985ad21cbb7d1708b362f159849116b0d724b2ad3b3764cd6 WatchSource:0}: Error finding container 80889c062fa7f90985ad21cbb7d1708b362f159849116b0d724b2ad3b3764cd6: Status 404 returned error can't find the container with id 80889c062fa7f90985ad21cbb7d1708b362f159849116b0d724b2ad3b3764cd6 Feb 02 12:25:01 crc kubenswrapper[4846]: I0202 12:25:01.887307 4846 generic.go:334] "Generic (PLEG): container finished" podID="9aa5a1ea-0fe6-4b98-adf6-a6af93581534" containerID="03af70fdaacd1330f8a0a3860fad3e353ab5bfb40d101166a052d01fec2ee1f1" exitCode=0 Feb 02 12:25:01 crc kubenswrapper[4846]: I0202 12:25:01.887369 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" event={"ID":"9aa5a1ea-0fe6-4b98-adf6-a6af93581534","Type":"ContainerDied","Data":"03af70fdaacd1330f8a0a3860fad3e353ab5bfb40d101166a052d01fec2ee1f1"} Feb 02 12:25:01 crc kubenswrapper[4846]: I0202 12:25:01.887909 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" event={"ID":"9aa5a1ea-0fe6-4b98-adf6-a6af93581534","Type":"ContainerStarted","Data":"80889c062fa7f90985ad21cbb7d1708b362f159849116b0d724b2ad3b3764cd6"} Feb 02 12:25:02 crc kubenswrapper[4846]: I0202 12:25:02.780423 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-r692g" Feb 02 12:25:02 crc kubenswrapper[4846]: I0202 12:25:02.856379 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-r692g" Feb 02 12:25:05 crc kubenswrapper[4846]: I0202 12:25:05.917180 4846 generic.go:334] "Generic (PLEG): container finished" podID="9aa5a1ea-0fe6-4b98-adf6-a6af93581534" containerID="50db4acdf74eb92399d6d1ed32c19a9a5f86b1c2befb622980fe6e4956fdc348" exitCode=0 Feb 02 12:25:05 crc kubenswrapper[4846]: I0202 12:25:05.917240 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" event={"ID":"9aa5a1ea-0fe6-4b98-adf6-a6af93581534","Type":"ContainerDied","Data":"50db4acdf74eb92399d6d1ed32c19a9a5f86b1c2befb622980fe6e4956fdc348"} Feb 02 12:25:06 crc kubenswrapper[4846]: I0202 12:25:06.927372 4846 generic.go:334] "Generic (PLEG): container finished" podID="9aa5a1ea-0fe6-4b98-adf6-a6af93581534" containerID="319cc97f1689327d29da722e22044fe4a1b972a63469f1ec8bfb9b03acb28eac" exitCode=0 Feb 02 12:25:06 crc kubenswrapper[4846]: I0202 12:25:06.927427 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" event={"ID":"9aa5a1ea-0fe6-4b98-adf6-a6af93581534","Type":"ContainerDied","Data":"319cc97f1689327d29da722e22044fe4a1b972a63469f1ec8bfb9b03acb28eac"} Feb 02 12:25:07 crc kubenswrapper[4846]: I0202 12:25:07.784530 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-r692g" Feb 02 12:25:07 crc kubenswrapper[4846]: I0202 12:25:07.796685 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-2r6kw" Feb 02 12:25:07 crc kubenswrapper[4846]: I0202 12:25:07.906764 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-lrmwh" Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.176009 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.288425 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-util\") pod \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.288523 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49g5b\" (UniqueName: \"kubernetes.io/projected/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-kube-api-access-49g5b\") pod \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.288783 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-bundle\") pod \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\" (UID: \"9aa5a1ea-0fe6-4b98-adf6-a6af93581534\") " Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.290157 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-bundle" (OuterVolumeSpecName: "bundle") pod "9aa5a1ea-0fe6-4b98-adf6-a6af93581534" (UID: "9aa5a1ea-0fe6-4b98-adf6-a6af93581534"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.295160 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-kube-api-access-49g5b" (OuterVolumeSpecName: "kube-api-access-49g5b") pod "9aa5a1ea-0fe6-4b98-adf6-a6af93581534" (UID: "9aa5a1ea-0fe6-4b98-adf6-a6af93581534"). InnerVolumeSpecName "kube-api-access-49g5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.300051 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-util" (OuterVolumeSpecName: "util") pod "9aa5a1ea-0fe6-4b98-adf6-a6af93581534" (UID: "9aa5a1ea-0fe6-4b98-adf6-a6af93581534"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.392411 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.392461 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-util\") on node \"crc\" DevicePath \"\"" Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.392477 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49g5b\" (UniqueName: \"kubernetes.io/projected/9aa5a1ea-0fe6-4b98-adf6-a6af93581534-kube-api-access-49g5b\") on node \"crc\" DevicePath \"\"" Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.943716 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" event={"ID":"9aa5a1ea-0fe6-4b98-adf6-a6af93581534","Type":"ContainerDied","Data":"80889c062fa7f90985ad21cbb7d1708b362f159849116b0d724b2ad3b3764cd6"} Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.943752 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80889c062fa7f90985ad21cbb7d1708b362f159849116b0d724b2ad3b3764cd6" Feb 02 12:25:08 crc kubenswrapper[4846]: I0202 12:25:08.943779 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.013908 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2"] Feb 02 12:25:13 crc kubenswrapper[4846]: E0202 12:25:13.014724 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa5a1ea-0fe6-4b98-adf6-a6af93581534" containerName="util" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.014740 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa5a1ea-0fe6-4b98-adf6-a6af93581534" containerName="util" Feb 02 12:25:13 crc kubenswrapper[4846]: E0202 12:25:13.014757 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa5a1ea-0fe6-4b98-adf6-a6af93581534" containerName="extract" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.014764 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa5a1ea-0fe6-4b98-adf6-a6af93581534" containerName="extract" Feb 02 12:25:13 crc kubenswrapper[4846]: E0202 12:25:13.014779 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa5a1ea-0fe6-4b98-adf6-a6af93581534" containerName="pull" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.014787 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa5a1ea-0fe6-4b98-adf6-a6af93581534" containerName="pull" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.015243 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa5a1ea-0fe6-4b98-adf6-a6af93581534" containerName="extract" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.015721 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.017519 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-4bhjm" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.018481 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.018722 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.033932 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2"] Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.152869 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/3cd38cdc-04b1-415b-8827-b5043658d489-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-89jb2\" (UID: \"3cd38cdc-04b1-415b-8827-b5043658d489\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.152934 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zb42\" (UniqueName: \"kubernetes.io/projected/3cd38cdc-04b1-415b-8827-b5043658d489-kube-api-access-9zb42\") pod \"cert-manager-operator-controller-manager-66c8bdd694-89jb2\" (UID: \"3cd38cdc-04b1-415b-8827-b5043658d489\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.254016 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/3cd38cdc-04b1-415b-8827-b5043658d489-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-89jb2\" (UID: \"3cd38cdc-04b1-415b-8827-b5043658d489\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.254121 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zb42\" (UniqueName: \"kubernetes.io/projected/3cd38cdc-04b1-415b-8827-b5043658d489-kube-api-access-9zb42\") pod \"cert-manager-operator-controller-manager-66c8bdd694-89jb2\" (UID: \"3cd38cdc-04b1-415b-8827-b5043658d489\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.254847 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/3cd38cdc-04b1-415b-8827-b5043658d489-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-89jb2\" (UID: \"3cd38cdc-04b1-415b-8827-b5043658d489\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.273514 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zb42\" (UniqueName: \"kubernetes.io/projected/3cd38cdc-04b1-415b-8827-b5043658d489-kube-api-access-9zb42\") pod \"cert-manager-operator-controller-manager-66c8bdd694-89jb2\" (UID: \"3cd38cdc-04b1-415b-8827-b5043658d489\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.345388 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.745028 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2"] Feb 02 12:25:13 crc kubenswrapper[4846]: I0202 12:25:13.968954 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" event={"ID":"3cd38cdc-04b1-415b-8827-b5043658d489","Type":"ContainerStarted","Data":"39d2a0ae8ef52761d34f0a2e37d75f82a430a9972fadb57467df106a74f7a79b"} Feb 02 12:25:17 crc kubenswrapper[4846]: I0202 12:25:17.000836 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" event={"ID":"3cd38cdc-04b1-415b-8827-b5043658d489","Type":"ContainerStarted","Data":"4238c42b6f64112ba3c99cb49e919689d14c607f2565a709848c6dea2abb2f90"} Feb 02 12:25:17 crc kubenswrapper[4846]: I0202 12:25:17.024121 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-89jb2" podStartSLOduration=2.145968797 podStartE2EDuration="5.024102059s" podCreationTimestamp="2026-02-02 12:25:12 +0000 UTC" firstStartedPulling="2026-02-02 12:25:13.756678133 +0000 UTC m=+944.985265006" lastFinishedPulling="2026-02-02 12:25:16.634811405 +0000 UTC m=+947.863398268" observedRunningTime="2026-02-02 12:25:17.023129535 +0000 UTC m=+948.251716408" watchObservedRunningTime="2026-02-02 12:25:17.024102059 +0000 UTC m=+948.252688932" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.653549 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-65bwf"] Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.655289 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.657084 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.657336 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.658283 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-9mhhm" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.676008 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-65bwf"] Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.764759 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-65bwf\" (UID: \"2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1\") " pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.764907 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96jlx\" (UniqueName: \"kubernetes.io/projected/2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1-kube-api-access-96jlx\") pod \"cert-manager-webhook-6888856db4-65bwf\" (UID: \"2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1\") " pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.866380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-65bwf\" (UID: \"2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1\") " pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.866458 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96jlx\" (UniqueName: \"kubernetes.io/projected/2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1-kube-api-access-96jlx\") pod \"cert-manager-webhook-6888856db4-65bwf\" (UID: \"2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1\") " pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.887394 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96jlx\" (UniqueName: \"kubernetes.io/projected/2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1-kube-api-access-96jlx\") pod \"cert-manager-webhook-6888856db4-65bwf\" (UID: \"2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1\") " pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.890087 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-65bwf\" (UID: \"2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1\") " pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" Feb 02 12:25:20 crc kubenswrapper[4846]: I0202 12:25:20.970226 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" Feb 02 12:25:21 crc kubenswrapper[4846]: I0202 12:25:21.405245 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-65bwf"] Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.029690 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" event={"ID":"2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1","Type":"ContainerStarted","Data":"3cc71528add71e62a1e6147c2ab8b86118e4b85afbec177c1bdeb1103a1e4ee6"} Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.652472 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2rw9t"] Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.653877 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.664683 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2rw9t"] Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.789236 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-utilities\") pod \"certified-operators-2rw9t\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.789284 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-catalog-content\") pod \"certified-operators-2rw9t\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.789306 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fwnm\" (UniqueName: \"kubernetes.io/projected/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-kube-api-access-7fwnm\") pod \"certified-operators-2rw9t\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.890325 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-utilities\") pod \"certified-operators-2rw9t\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.890371 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-catalog-content\") pod \"certified-operators-2rw9t\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.890397 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fwnm\" (UniqueName: \"kubernetes.io/projected/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-kube-api-access-7fwnm\") pod \"certified-operators-2rw9t\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.891164 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-utilities\") pod \"certified-operators-2rw9t\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.891284 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-catalog-content\") pod \"certified-operators-2rw9t\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.915896 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fwnm\" (UniqueName: \"kubernetes.io/projected/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-kube-api-access-7fwnm\") pod \"certified-operators-2rw9t\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:22 crc kubenswrapper[4846]: I0202 12:25:22.978266 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.552396 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-vghnr"] Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.553189 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.559319 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-wg6mk" Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.561643 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2rw9t"] Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.568754 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-vghnr"] Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.704984 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsbtx\" (UniqueName: \"kubernetes.io/projected/5ccfdad1-2b84-4146-8c7f-5bfbf8009c66-kube-api-access-jsbtx\") pod \"cert-manager-cainjector-5545bd876-vghnr\" (UID: \"5ccfdad1-2b84-4146-8c7f-5bfbf8009c66\") " pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.705448 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ccfdad1-2b84-4146-8c7f-5bfbf8009c66-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-vghnr\" (UID: \"5ccfdad1-2b84-4146-8c7f-5bfbf8009c66\") " pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.806905 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ccfdad1-2b84-4146-8c7f-5bfbf8009c66-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-vghnr\" (UID: \"5ccfdad1-2b84-4146-8c7f-5bfbf8009c66\") " pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.807005 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsbtx\" (UniqueName: \"kubernetes.io/projected/5ccfdad1-2b84-4146-8c7f-5bfbf8009c66-kube-api-access-jsbtx\") pod \"cert-manager-cainjector-5545bd876-vghnr\" (UID: \"5ccfdad1-2b84-4146-8c7f-5bfbf8009c66\") " pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.829004 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsbtx\" (UniqueName: \"kubernetes.io/projected/5ccfdad1-2b84-4146-8c7f-5bfbf8009c66-kube-api-access-jsbtx\") pod \"cert-manager-cainjector-5545bd876-vghnr\" (UID: \"5ccfdad1-2b84-4146-8c7f-5bfbf8009c66\") " pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.830341 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ccfdad1-2b84-4146-8c7f-5bfbf8009c66-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-vghnr\" (UID: \"5ccfdad1-2b84-4146-8c7f-5bfbf8009c66\") " pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" Feb 02 12:25:23 crc kubenswrapper[4846]: I0202 12:25:23.893603 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" Feb 02 12:25:24 crc kubenswrapper[4846]: I0202 12:25:24.051161 4846 generic.go:334] "Generic (PLEG): container finished" podID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerID="b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25" exitCode=0 Feb 02 12:25:24 crc kubenswrapper[4846]: I0202 12:25:24.051203 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rw9t" event={"ID":"884b18cb-a1be-4bdf-89c0-8a40cd6b016c","Type":"ContainerDied","Data":"b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25"} Feb 02 12:25:24 crc kubenswrapper[4846]: I0202 12:25:24.051229 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rw9t" event={"ID":"884b18cb-a1be-4bdf-89c0-8a40cd6b016c","Type":"ContainerStarted","Data":"79ebb549a063027c94b5ba8be430ec79980332877e09a235985478ac570f8c27"} Feb 02 12:25:24 crc kubenswrapper[4846]: I0202 12:25:24.354808 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-vghnr"] Feb 02 12:25:24 crc kubenswrapper[4846]: W0202 12:25:24.360705 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ccfdad1_2b84_4146_8c7f_5bfbf8009c66.slice/crio-080249d010c8556b3a21d633fc55057eae38397053d473816742b3f4267d2c06 WatchSource:0}: Error finding container 080249d010c8556b3a21d633fc55057eae38397053d473816742b3f4267d2c06: Status 404 returned error can't find the container with id 080249d010c8556b3a21d633fc55057eae38397053d473816742b3f4267d2c06 Feb 02 12:25:25 crc kubenswrapper[4846]: I0202 12:25:25.066271 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" event={"ID":"5ccfdad1-2b84-4146-8c7f-5bfbf8009c66","Type":"ContainerStarted","Data":"080249d010c8556b3a21d633fc55057eae38397053d473816742b3f4267d2c06"} Feb 02 12:25:27 crc kubenswrapper[4846]: I0202 12:25:27.078602 4846 generic.go:334] "Generic (PLEG): container finished" podID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerID="470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a" exitCode=0 Feb 02 12:25:27 crc kubenswrapper[4846]: I0202 12:25:27.078665 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rw9t" event={"ID":"884b18cb-a1be-4bdf-89c0-8a40cd6b016c","Type":"ContainerDied","Data":"470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a"} Feb 02 12:25:27 crc kubenswrapper[4846]: I0202 12:25:27.091704 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" event={"ID":"2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1","Type":"ContainerStarted","Data":"5e995dde7ea7fcf953963944a14de199467e82684c31a331fc671f9823f797c4"} Feb 02 12:25:27 crc kubenswrapper[4846]: I0202 12:25:27.091767 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" Feb 02 12:25:27 crc kubenswrapper[4846]: I0202 12:25:27.093203 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" event={"ID":"5ccfdad1-2b84-4146-8c7f-5bfbf8009c66","Type":"ContainerStarted","Data":"cd02d53f66caa9b38e5497b84e15648bd2700aac05439d447f134e134050c795"} Feb 02 12:25:27 crc kubenswrapper[4846]: I0202 12:25:27.110479 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" podStartSLOduration=2.265233652 podStartE2EDuration="7.110459037s" podCreationTimestamp="2026-02-02 12:25:20 +0000 UTC" firstStartedPulling="2026-02-02 12:25:21.416034245 +0000 UTC m=+952.644621098" lastFinishedPulling="2026-02-02 12:25:26.26125962 +0000 UTC m=+957.489846483" observedRunningTime="2026-02-02 12:25:27.107793211 +0000 UTC m=+958.336380074" watchObservedRunningTime="2026-02-02 12:25:27.110459037 +0000 UTC m=+958.339045900" Feb 02 12:25:27 crc kubenswrapper[4846]: I0202 12:25:27.128447 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-vghnr" podStartSLOduration=2.234274126 podStartE2EDuration="4.128430116s" podCreationTimestamp="2026-02-02 12:25:23 +0000 UTC" firstStartedPulling="2026-02-02 12:25:24.363533754 +0000 UTC m=+955.592120617" lastFinishedPulling="2026-02-02 12:25:26.257689744 +0000 UTC m=+957.486276607" observedRunningTime="2026-02-02 12:25:27.128171749 +0000 UTC m=+958.356758622" watchObservedRunningTime="2026-02-02 12:25:27.128430116 +0000 UTC m=+958.357016979" Feb 02 12:25:29 crc kubenswrapper[4846]: I0202 12:25:29.106483 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rw9t" event={"ID":"884b18cb-a1be-4bdf-89c0-8a40cd6b016c","Type":"ContainerStarted","Data":"217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03"} Feb 02 12:25:29 crc kubenswrapper[4846]: I0202 12:25:29.133462 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2rw9t" podStartSLOduration=2.577617442 podStartE2EDuration="7.133447219s" podCreationTimestamp="2026-02-02 12:25:22 +0000 UTC" firstStartedPulling="2026-02-02 12:25:24.05477093 +0000 UTC m=+955.283357823" lastFinishedPulling="2026-02-02 12:25:28.610600747 +0000 UTC m=+959.839187600" observedRunningTime="2026-02-02 12:25:29.131071011 +0000 UTC m=+960.359657884" watchObservedRunningTime="2026-02-02 12:25:29.133447219 +0000 UTC m=+960.362034082" Feb 02 12:25:32 crc kubenswrapper[4846]: I0202 12:25:32.978706 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:32 crc kubenswrapper[4846]: I0202 12:25:32.979288 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:33 crc kubenswrapper[4846]: I0202 12:25:33.019207 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:33 crc kubenswrapper[4846]: I0202 12:25:33.161380 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:33 crc kubenswrapper[4846]: I0202 12:25:33.245043 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2rw9t"] Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.137418 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2rw9t" podUID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerName="registry-server" containerID="cri-o://217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03" gracePeriod=2 Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.459551 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.577091 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-catalog-content\") pod \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.577191 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fwnm\" (UniqueName: \"kubernetes.io/projected/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-kube-api-access-7fwnm\") pod \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.577253 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-utilities\") pod \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\" (UID: \"884b18cb-a1be-4bdf-89c0-8a40cd6b016c\") " Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.578395 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-utilities" (OuterVolumeSpecName: "utilities") pod "884b18cb-a1be-4bdf-89c0-8a40cd6b016c" (UID: "884b18cb-a1be-4bdf-89c0-8a40cd6b016c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.585842 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-kube-api-access-7fwnm" (OuterVolumeSpecName: "kube-api-access-7fwnm") pod "884b18cb-a1be-4bdf-89c0-8a40cd6b016c" (UID: "884b18cb-a1be-4bdf-89c0-8a40cd6b016c"). InnerVolumeSpecName "kube-api-access-7fwnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.629370 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "884b18cb-a1be-4bdf-89c0-8a40cd6b016c" (UID: "884b18cb-a1be-4bdf-89c0-8a40cd6b016c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.678700 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fwnm\" (UniqueName: \"kubernetes.io/projected/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-kube-api-access-7fwnm\") on node \"crc\" DevicePath \"\"" Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.678743 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.678770 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/884b18cb-a1be-4bdf-89c0-8a40cd6b016c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:25:35 crc kubenswrapper[4846]: I0202 12:25:35.973518 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-65bwf" Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.145178 4846 generic.go:334] "Generic (PLEG): container finished" podID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerID="217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03" exitCode=0 Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.145215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rw9t" event={"ID":"884b18cb-a1be-4bdf-89c0-8a40cd6b016c","Type":"ContainerDied","Data":"217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03"} Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.145242 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2rw9t" event={"ID":"884b18cb-a1be-4bdf-89c0-8a40cd6b016c","Type":"ContainerDied","Data":"79ebb549a063027c94b5ba8be430ec79980332877e09a235985478ac570f8c27"} Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.145268 4846 scope.go:117] "RemoveContainer" containerID="217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03" Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.145281 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2rw9t" Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.164567 4846 scope.go:117] "RemoveContainer" containerID="470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a" Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.172885 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2rw9t"] Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.177101 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2rw9t"] Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.198853 4846 scope.go:117] "RemoveContainer" containerID="b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25" Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.214634 4846 scope.go:117] "RemoveContainer" containerID="217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03" Feb 02 12:25:36 crc kubenswrapper[4846]: E0202 12:25:36.215031 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03\": container with ID starting with 217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03 not found: ID does not exist" containerID="217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03" Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.215116 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03"} err="failed to get container status \"217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03\": rpc error: code = NotFound desc = could not find container \"217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03\": container with ID starting with 217b9c97f708b14e39e18052beed753d28079d884ccaaec2be1d7cef1b573d03 not found: ID does not exist" Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.215180 4846 scope.go:117] "RemoveContainer" containerID="470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a" Feb 02 12:25:36 crc kubenswrapper[4846]: E0202 12:25:36.215638 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a\": container with ID starting with 470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a not found: ID does not exist" containerID="470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a" Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.215748 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a"} err="failed to get container status \"470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a\": rpc error: code = NotFound desc = could not find container \"470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a\": container with ID starting with 470b257353a84e9f5257be548d8c5c90ab212c9c2bde9443254ee1124dc46e6a not found: ID does not exist" Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.215832 4846 scope.go:117] "RemoveContainer" containerID="b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25" Feb 02 12:25:36 crc kubenswrapper[4846]: E0202 12:25:36.216208 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25\": container with ID starting with b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25 not found: ID does not exist" containerID="b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25" Feb 02 12:25:36 crc kubenswrapper[4846]: I0202 12:25:36.216337 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25"} err="failed to get container status \"b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25\": rpc error: code = NotFound desc = could not find container \"b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25\": container with ID starting with b1b602030862604f81acf2ce2e708f31a2ba184923f05e1d04dd0f928c226f25 not found: ID does not exist" Feb 02 12:25:37 crc kubenswrapper[4846]: I0202 12:25:37.641336 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" path="/var/lib/kubelet/pods/884b18cb-a1be-4bdf-89c0-8a40cd6b016c/volumes" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.618047 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-nb4jl"] Feb 02 12:25:38 crc kubenswrapper[4846]: E0202 12:25:38.618332 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerName="registry-server" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.618349 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerName="registry-server" Feb 02 12:25:38 crc kubenswrapper[4846]: E0202 12:25:38.618362 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerName="extract-content" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.618370 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerName="extract-content" Feb 02 12:25:38 crc kubenswrapper[4846]: E0202 12:25:38.618390 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerName="extract-utilities" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.618398 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerName="extract-utilities" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.618518 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="884b18cb-a1be-4bdf-89c0-8a40cd6b016c" containerName="registry-server" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.619017 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-nb4jl" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.621011 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-nvhhv" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.626505 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-nb4jl"] Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.716708 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhr94\" (UniqueName: \"kubernetes.io/projected/f6b1cf52-1897-4864-8335-257ac2391d17-kube-api-access-zhr94\") pod \"cert-manager-545d4d4674-nb4jl\" (UID: \"f6b1cf52-1897-4864-8335-257ac2391d17\") " pod="cert-manager/cert-manager-545d4d4674-nb4jl" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.716780 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f6b1cf52-1897-4864-8335-257ac2391d17-bound-sa-token\") pod \"cert-manager-545d4d4674-nb4jl\" (UID: \"f6b1cf52-1897-4864-8335-257ac2391d17\") " pod="cert-manager/cert-manager-545d4d4674-nb4jl" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.818481 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhr94\" (UniqueName: \"kubernetes.io/projected/f6b1cf52-1897-4864-8335-257ac2391d17-kube-api-access-zhr94\") pod \"cert-manager-545d4d4674-nb4jl\" (UID: \"f6b1cf52-1897-4864-8335-257ac2391d17\") " pod="cert-manager/cert-manager-545d4d4674-nb4jl" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.818532 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f6b1cf52-1897-4864-8335-257ac2391d17-bound-sa-token\") pod \"cert-manager-545d4d4674-nb4jl\" (UID: \"f6b1cf52-1897-4864-8335-257ac2391d17\") " pod="cert-manager/cert-manager-545d4d4674-nb4jl" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.841185 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f6b1cf52-1897-4864-8335-257ac2391d17-bound-sa-token\") pod \"cert-manager-545d4d4674-nb4jl\" (UID: \"f6b1cf52-1897-4864-8335-257ac2391d17\") " pod="cert-manager/cert-manager-545d4d4674-nb4jl" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.842680 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhr94\" (UniqueName: \"kubernetes.io/projected/f6b1cf52-1897-4864-8335-257ac2391d17-kube-api-access-zhr94\") pod \"cert-manager-545d4d4674-nb4jl\" (UID: \"f6b1cf52-1897-4864-8335-257ac2391d17\") " pod="cert-manager/cert-manager-545d4d4674-nb4jl" Feb 02 12:25:38 crc kubenswrapper[4846]: I0202 12:25:38.938425 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-nb4jl" Feb 02 12:25:39 crc kubenswrapper[4846]: I0202 12:25:39.347701 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-nb4jl"] Feb 02 12:25:39 crc kubenswrapper[4846]: I0202 12:25:39.650579 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-nb4jl" event={"ID":"f6b1cf52-1897-4864-8335-257ac2391d17","Type":"ContainerStarted","Data":"38f94f031a3dd8019af76218173faf967487c852c29988e1c4d0c2de2ebbdde0"} Feb 02 12:25:39 crc kubenswrapper[4846]: I0202 12:25:39.650985 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-nb4jl" event={"ID":"f6b1cf52-1897-4864-8335-257ac2391d17","Type":"ContainerStarted","Data":"b646249faf39af5f5d73747c6470e47610bb47cf007cf5c936816719020bfe73"} Feb 02 12:25:39 crc kubenswrapper[4846]: I0202 12:25:39.667606 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-nb4jl" podStartSLOduration=1.667590972 podStartE2EDuration="1.667590972s" podCreationTimestamp="2026-02-02 12:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:25:39.66588027 +0000 UTC m=+970.894467143" watchObservedRunningTime="2026-02-02 12:25:39.667590972 +0000 UTC m=+970.896177835" Feb 02 12:25:39 crc kubenswrapper[4846]: I0202 12:25:39.993915 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c7snf"] Feb 02 12:25:39 crc kubenswrapper[4846]: I0202 12:25:39.995285 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.002935 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7snf"] Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.139035 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-utilities\") pod \"redhat-marketplace-c7snf\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.139085 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-catalog-content\") pod \"redhat-marketplace-c7snf\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.139123 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwvc4\" (UniqueName: \"kubernetes.io/projected/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-kube-api-access-kwvc4\") pod \"redhat-marketplace-c7snf\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.240777 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-utilities\") pod \"redhat-marketplace-c7snf\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.241090 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-catalog-content\") pod \"redhat-marketplace-c7snf\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.241150 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwvc4\" (UniqueName: \"kubernetes.io/projected/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-kube-api-access-kwvc4\") pod \"redhat-marketplace-c7snf\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.241224 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-utilities\") pod \"redhat-marketplace-c7snf\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.241488 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-catalog-content\") pod \"redhat-marketplace-c7snf\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.263644 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwvc4\" (UniqueName: \"kubernetes.io/projected/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-kube-api-access-kwvc4\") pod \"redhat-marketplace-c7snf\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.320054 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:40 crc kubenswrapper[4846]: I0202 12:25:40.767800 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7snf"] Feb 02 12:25:41 crc kubenswrapper[4846]: I0202 12:25:41.662606 4846 generic.go:334] "Generic (PLEG): container finished" podID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerID="d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d" exitCode=0 Feb 02 12:25:41 crc kubenswrapper[4846]: I0202 12:25:41.662687 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7snf" event={"ID":"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8","Type":"ContainerDied","Data":"d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d"} Feb 02 12:25:41 crc kubenswrapper[4846]: I0202 12:25:41.662952 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7snf" event={"ID":"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8","Type":"ContainerStarted","Data":"2e8f227269b6801c4d31afe40028b380b686a04f7873ff7f5e172d587e0c4c01"} Feb 02 12:25:42 crc kubenswrapper[4846]: I0202 12:25:42.670399 4846 generic.go:334] "Generic (PLEG): container finished" podID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerID="a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e" exitCode=0 Feb 02 12:25:42 crc kubenswrapper[4846]: I0202 12:25:42.670446 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7snf" event={"ID":"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8","Type":"ContainerDied","Data":"a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e"} Feb 02 12:25:43 crc kubenswrapper[4846]: I0202 12:25:43.679161 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7snf" event={"ID":"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8","Type":"ContainerStarted","Data":"0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc"} Feb 02 12:25:43 crc kubenswrapper[4846]: I0202 12:25:43.703374 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c7snf" podStartSLOduration=3.303139196 podStartE2EDuration="4.703356332s" podCreationTimestamp="2026-02-02 12:25:39 +0000 UTC" firstStartedPulling="2026-02-02 12:25:41.665814337 +0000 UTC m=+972.894401220" lastFinishedPulling="2026-02-02 12:25:43.066031483 +0000 UTC m=+974.294618356" observedRunningTime="2026-02-02 12:25:43.698929733 +0000 UTC m=+974.927516626" watchObservedRunningTime="2026-02-02 12:25:43.703356332 +0000 UTC m=+974.931943195" Feb 02 12:25:48 crc kubenswrapper[4846]: I0202 12:25:48.954818 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vfv69"] Feb 02 12:25:48 crc kubenswrapper[4846]: I0202 12:25:48.955845 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vfv69" Feb 02 12:25:48 crc kubenswrapper[4846]: I0202 12:25:48.959112 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 02 12:25:48 crc kubenswrapper[4846]: I0202 12:25:48.961048 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 02 12:25:48 crc kubenswrapper[4846]: I0202 12:25:48.961681 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-j4wtc" Feb 02 12:25:48 crc kubenswrapper[4846]: I0202 12:25:48.978599 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vfv69"] Feb 02 12:25:49 crc kubenswrapper[4846]: I0202 12:25:49.063262 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ppms\" (UniqueName: \"kubernetes.io/projected/50c4c682-276a-4a69-b42d-d8b244b49341-kube-api-access-4ppms\") pod \"openstack-operator-index-vfv69\" (UID: \"50c4c682-276a-4a69-b42d-d8b244b49341\") " pod="openstack-operators/openstack-operator-index-vfv69" Feb 02 12:25:49 crc kubenswrapper[4846]: I0202 12:25:49.164989 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ppms\" (UniqueName: \"kubernetes.io/projected/50c4c682-276a-4a69-b42d-d8b244b49341-kube-api-access-4ppms\") pod \"openstack-operator-index-vfv69\" (UID: \"50c4c682-276a-4a69-b42d-d8b244b49341\") " pod="openstack-operators/openstack-operator-index-vfv69" Feb 02 12:25:49 crc kubenswrapper[4846]: I0202 12:25:49.183764 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ppms\" (UniqueName: \"kubernetes.io/projected/50c4c682-276a-4a69-b42d-d8b244b49341-kube-api-access-4ppms\") pod \"openstack-operator-index-vfv69\" (UID: \"50c4c682-276a-4a69-b42d-d8b244b49341\") " pod="openstack-operators/openstack-operator-index-vfv69" Feb 02 12:25:49 crc kubenswrapper[4846]: I0202 12:25:49.277507 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vfv69" Feb 02 12:25:49 crc kubenswrapper[4846]: I0202 12:25:49.668963 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vfv69"] Feb 02 12:25:49 crc kubenswrapper[4846]: I0202 12:25:49.716213 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vfv69" event={"ID":"50c4c682-276a-4a69-b42d-d8b244b49341","Type":"ContainerStarted","Data":"3e3b766f253f6ef9b3bba9b371b383f67edc5e3d1d5d5ab998f9565f01efa51f"} Feb 02 12:25:50 crc kubenswrapper[4846]: I0202 12:25:50.320198 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:50 crc kubenswrapper[4846]: I0202 12:25:50.320491 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:50 crc kubenswrapper[4846]: I0202 12:25:50.368441 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:50 crc kubenswrapper[4846]: I0202 12:25:50.731798 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vfv69" event={"ID":"50c4c682-276a-4a69-b42d-d8b244b49341","Type":"ContainerStarted","Data":"a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0"} Feb 02 12:25:50 crc kubenswrapper[4846]: I0202 12:25:50.744869 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vfv69" podStartSLOduration=2.042817402 podStartE2EDuration="2.744848642s" podCreationTimestamp="2026-02-02 12:25:48 +0000 UTC" firstStartedPulling="2026-02-02 12:25:49.675583389 +0000 UTC m=+980.904170282" lastFinishedPulling="2026-02-02 12:25:50.377614659 +0000 UTC m=+981.606201522" observedRunningTime="2026-02-02 12:25:50.743800357 +0000 UTC m=+981.972387240" watchObservedRunningTime="2026-02-02 12:25:50.744848642 +0000 UTC m=+981.973435505" Feb 02 12:25:50 crc kubenswrapper[4846]: I0202 12:25:50.779943 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:52 crc kubenswrapper[4846]: I0202 12:25:52.940384 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-vfv69"] Feb 02 12:25:52 crc kubenswrapper[4846]: I0202 12:25:52.940971 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-vfv69" podUID="50c4c682-276a-4a69-b42d-d8b244b49341" containerName="registry-server" containerID="cri-o://a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0" gracePeriod=2 Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.308412 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vfv69" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.419936 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ppms\" (UniqueName: \"kubernetes.io/projected/50c4c682-276a-4a69-b42d-d8b244b49341-kube-api-access-4ppms\") pod \"50c4c682-276a-4a69-b42d-d8b244b49341\" (UID: \"50c4c682-276a-4a69-b42d-d8b244b49341\") " Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.425283 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50c4c682-276a-4a69-b42d-d8b244b49341-kube-api-access-4ppms" (OuterVolumeSpecName: "kube-api-access-4ppms") pod "50c4c682-276a-4a69-b42d-d8b244b49341" (UID: "50c4c682-276a-4a69-b42d-d8b244b49341"). InnerVolumeSpecName "kube-api-access-4ppms". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.521932 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ppms\" (UniqueName: \"kubernetes.io/projected/50c4c682-276a-4a69-b42d-d8b244b49341-kube-api-access-4ppms\") on node \"crc\" DevicePath \"\"" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.750216 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xl544"] Feb 02 12:25:53 crc kubenswrapper[4846]: E0202 12:25:53.750450 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c4c682-276a-4a69-b42d-d8b244b49341" containerName="registry-server" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.750460 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c4c682-276a-4a69-b42d-d8b244b49341" containerName="registry-server" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.750566 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="50c4c682-276a-4a69-b42d-d8b244b49341" containerName="registry-server" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.751010 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xl544" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.758129 4846 generic.go:334] "Generic (PLEG): container finished" podID="50c4c682-276a-4a69-b42d-d8b244b49341" containerID="a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0" exitCode=0 Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.758348 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vfv69" event={"ID":"50c4c682-276a-4a69-b42d-d8b244b49341","Type":"ContainerDied","Data":"a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0"} Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.758542 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vfv69" event={"ID":"50c4c682-276a-4a69-b42d-d8b244b49341","Type":"ContainerDied","Data":"3e3b766f253f6ef9b3bba9b371b383f67edc5e3d1d5d5ab998f9565f01efa51f"} Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.758907 4846 scope.go:117] "RemoveContainer" containerID="a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.759327 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vfv69" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.770407 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xl544"] Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.791858 4846 scope.go:117] "RemoveContainer" containerID="a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0" Feb 02 12:25:53 crc kubenswrapper[4846]: E0202 12:25:53.792855 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0\": container with ID starting with a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0 not found: ID does not exist" containerID="a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.792992 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0"} err="failed to get container status \"a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0\": rpc error: code = NotFound desc = could not find container \"a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0\": container with ID starting with a83ec40df9a57c83319a22c3178cf1d013edeeaa2d21e19682080809ecd97fe0 not found: ID does not exist" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.811115 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-vfv69"] Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.819522 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-vfv69"] Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.826128 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gplf2\" (UniqueName: \"kubernetes.io/projected/122f29f2-1a54-4f6f-a913-24fb8948af5d-kube-api-access-gplf2\") pod \"openstack-operator-index-xl544\" (UID: \"122f29f2-1a54-4f6f-a913-24fb8948af5d\") " pod="openstack-operators/openstack-operator-index-xl544" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.927282 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gplf2\" (UniqueName: \"kubernetes.io/projected/122f29f2-1a54-4f6f-a913-24fb8948af5d-kube-api-access-gplf2\") pod \"openstack-operator-index-xl544\" (UID: \"122f29f2-1a54-4f6f-a913-24fb8948af5d\") " pod="openstack-operators/openstack-operator-index-xl544" Feb 02 12:25:53 crc kubenswrapper[4846]: I0202 12:25:53.950837 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gplf2\" (UniqueName: \"kubernetes.io/projected/122f29f2-1a54-4f6f-a913-24fb8948af5d-kube-api-access-gplf2\") pod \"openstack-operator-index-xl544\" (UID: \"122f29f2-1a54-4f6f-a913-24fb8948af5d\") " pod="openstack-operators/openstack-operator-index-xl544" Feb 02 12:25:54 crc kubenswrapper[4846]: I0202 12:25:54.081715 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xl544" Feb 02 12:25:54 crc kubenswrapper[4846]: I0202 12:25:54.273309 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xl544"] Feb 02 12:25:54 crc kubenswrapper[4846]: W0202 12:25:54.278228 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod122f29f2_1a54_4f6f_a913_24fb8948af5d.slice/crio-7f691dbed7458a5bfcd944d8272037f05a34e15cd42698ad89bbc94ab5bf9257 WatchSource:0}: Error finding container 7f691dbed7458a5bfcd944d8272037f05a34e15cd42698ad89bbc94ab5bf9257: Status 404 returned error can't find the container with id 7f691dbed7458a5bfcd944d8272037f05a34e15cd42698ad89bbc94ab5bf9257 Feb 02 12:25:54 crc kubenswrapper[4846]: I0202 12:25:54.775616 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xl544" event={"ID":"122f29f2-1a54-4f6f-a913-24fb8948af5d","Type":"ContainerStarted","Data":"7f691dbed7458a5bfcd944d8272037f05a34e15cd42698ad89bbc94ab5bf9257"} Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.142410 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7snf"] Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.143213 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c7snf" podUID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerName="registry-server" containerID="cri-o://0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc" gracePeriod=2 Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.442800 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50c4c682-276a-4a69-b42d-d8b244b49341" path="/var/lib/kubelet/pods/50c4c682-276a-4a69-b42d-d8b244b49341/volumes" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.494778 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.648712 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-catalog-content\") pod \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.648793 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwvc4\" (UniqueName: \"kubernetes.io/projected/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-kube-api-access-kwvc4\") pod \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.648839 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-utilities\") pod \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\" (UID: \"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8\") " Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.649725 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-utilities" (OuterVolumeSpecName: "utilities") pod "5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" (UID: "5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.653316 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-kube-api-access-kwvc4" (OuterVolumeSpecName: "kube-api-access-kwvc4") pod "5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" (UID: "5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8"). InnerVolumeSpecName "kube-api-access-kwvc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.670385 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" (UID: "5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.750357 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.750584 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwvc4\" (UniqueName: \"kubernetes.io/projected/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-kube-api-access-kwvc4\") on node \"crc\" DevicePath \"\"" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.750713 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.787188 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xl544" event={"ID":"122f29f2-1a54-4f6f-a913-24fb8948af5d","Type":"ContainerStarted","Data":"096946dc8bf1701217777818e3e8edfa17bb75be5aea3c2b249401d5fb8c74ca"} Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.790951 4846 generic.go:334] "Generic (PLEG): container finished" podID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerID="0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc" exitCode=0 Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.790981 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7snf" event={"ID":"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8","Type":"ContainerDied","Data":"0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc"} Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.791003 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7snf" event={"ID":"5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8","Type":"ContainerDied","Data":"2e8f227269b6801c4d31afe40028b380b686a04f7873ff7f5e172d587e0c4c01"} Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.791032 4846 scope.go:117] "RemoveContainer" containerID="0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.791341 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7snf" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.821652 4846 scope.go:117] "RemoveContainer" containerID="a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.822297 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xl544" podStartSLOduration=2.353750551 podStartE2EDuration="2.822286092s" podCreationTimestamp="2026-02-02 12:25:53 +0000 UTC" firstStartedPulling="2026-02-02 12:25:54.282272644 +0000 UTC m=+985.510859507" lastFinishedPulling="2026-02-02 12:25:54.750808135 +0000 UTC m=+985.979395048" observedRunningTime="2026-02-02 12:25:55.806088867 +0000 UTC m=+987.034675730" watchObservedRunningTime="2026-02-02 12:25:55.822286092 +0000 UTC m=+987.050872965" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.826652 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7snf"] Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.830336 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7snf"] Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.856812 4846 scope.go:117] "RemoveContainer" containerID="d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.871432 4846 scope.go:117] "RemoveContainer" containerID="0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc" Feb 02 12:25:55 crc kubenswrapper[4846]: E0202 12:25:55.871876 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc\": container with ID starting with 0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc not found: ID does not exist" containerID="0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.871914 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc"} err="failed to get container status \"0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc\": rpc error: code = NotFound desc = could not find container \"0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc\": container with ID starting with 0706f826000752b5d115e0a6d0e078521fcf5802eda2f9ed9005162e5cdc80cc not found: ID does not exist" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.871945 4846 scope.go:117] "RemoveContainer" containerID="a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e" Feb 02 12:25:55 crc kubenswrapper[4846]: E0202 12:25:55.880817 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e\": container with ID starting with a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e not found: ID does not exist" containerID="a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.880862 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e"} err="failed to get container status \"a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e\": rpc error: code = NotFound desc = could not find container \"a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e\": container with ID starting with a2485551fc8df46467321eff3779a72a2aeb4244b838743805ae251feb7be10e not found: ID does not exist" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.880893 4846 scope.go:117] "RemoveContainer" containerID="d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d" Feb 02 12:25:55 crc kubenswrapper[4846]: E0202 12:25:55.881354 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d\": container with ID starting with d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d not found: ID does not exist" containerID="d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d" Feb 02 12:25:55 crc kubenswrapper[4846]: I0202 12:25:55.881408 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d"} err="failed to get container status \"d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d\": rpc error: code = NotFound desc = could not find container \"d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d\": container with ID starting with d52efa147eee41d2c3405f5abad49e0b0f1ba9c9c6b227f181c2ff9879f4d68d not found: ID does not exist" Feb 02 12:25:57 crc kubenswrapper[4846]: I0202 12:25:57.430934 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" path="/var/lib/kubelet/pods/5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8/volumes" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.082193 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-xl544" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.083077 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-xl544" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.129997 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-xl544" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.874889 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-xl544" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.952664 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dkwgn"] Feb 02 12:26:04 crc kubenswrapper[4846]: E0202 12:26:04.952889 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerName="extract-utilities" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.952900 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerName="extract-utilities" Feb 02 12:26:04 crc kubenswrapper[4846]: E0202 12:26:04.952918 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerName="extract-content" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.952924 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerName="extract-content" Feb 02 12:26:04 crc kubenswrapper[4846]: E0202 12:26:04.952938 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerName="registry-server" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.952944 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerName="registry-server" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.953038 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d2723a6-fd8f-4552-bc6c-9136d7e0e9c8" containerName="registry-server" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.953841 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:04 crc kubenswrapper[4846]: I0202 12:26:04.964797 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dkwgn"] Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.078067 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-utilities\") pod \"community-operators-dkwgn\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.078321 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-catalog-content\") pod \"community-operators-dkwgn\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.078378 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sh7m\" (UniqueName: \"kubernetes.io/projected/30e3636e-eb78-430f-82eb-957e38c9a2c7-kube-api-access-6sh7m\") pod \"community-operators-dkwgn\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.180170 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-catalog-content\") pod \"community-operators-dkwgn\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.180247 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sh7m\" (UniqueName: \"kubernetes.io/projected/30e3636e-eb78-430f-82eb-957e38c9a2c7-kube-api-access-6sh7m\") pod \"community-operators-dkwgn\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.180324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-utilities\") pod \"community-operators-dkwgn\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.181007 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-utilities\") pod \"community-operators-dkwgn\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.181060 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-catalog-content\") pod \"community-operators-dkwgn\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.200535 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sh7m\" (UniqueName: \"kubernetes.io/projected/30e3636e-eb78-430f-82eb-957e38c9a2c7-kube-api-access-6sh7m\") pod \"community-operators-dkwgn\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.277426 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.571746 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dkwgn"] Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.849556 4846 generic.go:334] "Generic (PLEG): container finished" podID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerID="6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3" exitCode=0 Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.849641 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkwgn" event={"ID":"30e3636e-eb78-430f-82eb-957e38c9a2c7","Type":"ContainerDied","Data":"6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3"} Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.849925 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkwgn" event={"ID":"30e3636e-eb78-430f-82eb-957e38c9a2c7","Type":"ContainerStarted","Data":"7a1645e335d859b9256c778daee1b120c2226bbc58ea6bf6b54dfe9697bbf589"} Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.979055 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw"] Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.980157 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.981736 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-btfbc" Feb 02 12:26:05 crc kubenswrapper[4846]: I0202 12:26:05.988255 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw"] Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.093379 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-util\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.093675 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpxl8\" (UniqueName: \"kubernetes.io/projected/0841d2fe-5012-47c0-a591-849d4f8cc0f5-kube-api-access-qpxl8\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.093749 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-bundle\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.195048 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpxl8\" (UniqueName: \"kubernetes.io/projected/0841d2fe-5012-47c0-a591-849d4f8cc0f5-kube-api-access-qpxl8\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.196032 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-bundle\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.196188 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-util\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.196573 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-util\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.196559 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-bundle\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.212516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpxl8\" (UniqueName: \"kubernetes.io/projected/0841d2fe-5012-47c0-a591-849d4f8cc0f5-kube-api-access-qpxl8\") pod \"805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.302315 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.519877 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw"] Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.865243 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkwgn" event={"ID":"30e3636e-eb78-430f-82eb-957e38c9a2c7","Type":"ContainerStarted","Data":"ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154"} Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.867372 4846 generic.go:334] "Generic (PLEG): container finished" podID="0841d2fe-5012-47c0-a591-849d4f8cc0f5" containerID="6252a4946f28e98fbaaa64ff5eb85c46161a5e3a0eb46ca32c3ff4f8318cf384" exitCode=0 Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.867453 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" event={"ID":"0841d2fe-5012-47c0-a591-849d4f8cc0f5","Type":"ContainerDied","Data":"6252a4946f28e98fbaaa64ff5eb85c46161a5e3a0eb46ca32c3ff4f8318cf384"} Feb 02 12:26:06 crc kubenswrapper[4846]: I0202 12:26:06.867493 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" event={"ID":"0841d2fe-5012-47c0-a591-849d4f8cc0f5","Type":"ContainerStarted","Data":"2bdd1fc284ee1415506d2cc895b0f619823e99d7c6375fdaa69ea4321a494ec4"} Feb 02 12:26:07 crc kubenswrapper[4846]: I0202 12:26:07.874523 4846 generic.go:334] "Generic (PLEG): container finished" podID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerID="ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154" exitCode=0 Feb 02 12:26:07 crc kubenswrapper[4846]: I0202 12:26:07.874649 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkwgn" event={"ID":"30e3636e-eb78-430f-82eb-957e38c9a2c7","Type":"ContainerDied","Data":"ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154"} Feb 02 12:26:07 crc kubenswrapper[4846]: I0202 12:26:07.878579 4846 generic.go:334] "Generic (PLEG): container finished" podID="0841d2fe-5012-47c0-a591-849d4f8cc0f5" containerID="f48086ad82a649213d5babdf6115d005d821fe413d5cd41de384cf79660a3deb" exitCode=0 Feb 02 12:26:07 crc kubenswrapper[4846]: I0202 12:26:07.878611 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" event={"ID":"0841d2fe-5012-47c0-a591-849d4f8cc0f5","Type":"ContainerDied","Data":"f48086ad82a649213d5babdf6115d005d821fe413d5cd41de384cf79660a3deb"} Feb 02 12:26:08 crc kubenswrapper[4846]: I0202 12:26:08.887991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkwgn" event={"ID":"30e3636e-eb78-430f-82eb-957e38c9a2c7","Type":"ContainerStarted","Data":"bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42"} Feb 02 12:26:08 crc kubenswrapper[4846]: I0202 12:26:08.894195 4846 generic.go:334] "Generic (PLEG): container finished" podID="0841d2fe-5012-47c0-a591-849d4f8cc0f5" containerID="135bc41d3906dc4ea47d4a851e3a0a5844c076425afc63ca4ecfd737a6339cc8" exitCode=0 Feb 02 12:26:08 crc kubenswrapper[4846]: I0202 12:26:08.894254 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" event={"ID":"0841d2fe-5012-47c0-a591-849d4f8cc0f5","Type":"ContainerDied","Data":"135bc41d3906dc4ea47d4a851e3a0a5844c076425afc63ca4ecfd737a6339cc8"} Feb 02 12:26:08 crc kubenswrapper[4846]: I0202 12:26:08.909467 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dkwgn" podStartSLOduration=2.47094417 podStartE2EDuration="4.909439495s" podCreationTimestamp="2026-02-02 12:26:04 +0000 UTC" firstStartedPulling="2026-02-02 12:26:05.8515429 +0000 UTC m=+997.080129763" lastFinishedPulling="2026-02-02 12:26:08.290038225 +0000 UTC m=+999.518625088" observedRunningTime="2026-02-02 12:26:08.903876589 +0000 UTC m=+1000.132463472" watchObservedRunningTime="2026-02-02 12:26:08.909439495 +0000 UTC m=+1000.138026358" Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.118852 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.150070 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-bundle\") pod \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.150126 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-util\") pod \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.150226 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpxl8\" (UniqueName: \"kubernetes.io/projected/0841d2fe-5012-47c0-a591-849d4f8cc0f5-kube-api-access-qpxl8\") pod \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\" (UID: \"0841d2fe-5012-47c0-a591-849d4f8cc0f5\") " Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.150944 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-bundle" (OuterVolumeSpecName: "bundle") pod "0841d2fe-5012-47c0-a591-849d4f8cc0f5" (UID: "0841d2fe-5012-47c0-a591-849d4f8cc0f5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.155005 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0841d2fe-5012-47c0-a591-849d4f8cc0f5-kube-api-access-qpxl8" (OuterVolumeSpecName: "kube-api-access-qpxl8") pod "0841d2fe-5012-47c0-a591-849d4f8cc0f5" (UID: "0841d2fe-5012-47c0-a591-849d4f8cc0f5"). InnerVolumeSpecName "kube-api-access-qpxl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.163906 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-util" (OuterVolumeSpecName: "util") pod "0841d2fe-5012-47c0-a591-849d4f8cc0f5" (UID: "0841d2fe-5012-47c0-a591-849d4f8cc0f5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.251599 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.251667 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0841d2fe-5012-47c0-a591-849d4f8cc0f5-util\") on node \"crc\" DevicePath \"\"" Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.251679 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpxl8\" (UniqueName: \"kubernetes.io/projected/0841d2fe-5012-47c0-a591-849d4f8cc0f5-kube-api-access-qpxl8\") on node \"crc\" DevicePath \"\"" Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.974491 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" event={"ID":"0841d2fe-5012-47c0-a591-849d4f8cc0f5","Type":"ContainerDied","Data":"2bdd1fc284ee1415506d2cc895b0f619823e99d7c6375fdaa69ea4321a494ec4"} Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.974521 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw" Feb 02 12:26:10 crc kubenswrapper[4846]: I0202 12:26:10.974530 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bdd1fc284ee1415506d2cc895b0f619823e99d7c6375fdaa69ea4321a494ec4" Feb 02 12:26:11 crc kubenswrapper[4846]: E0202 12:26:11.099006 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0841d2fe_5012_47c0_a591_849d4f8cc0f5.slice/crio-2bdd1fc284ee1415506d2cc895b0f619823e99d7c6375fdaa69ea4321a494ec4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0841d2fe_5012_47c0_a591_849d4f8cc0f5.slice\": RecentStats: unable to find data in memory cache]" Feb 02 12:26:15 crc kubenswrapper[4846]: I0202 12:26:15.278398 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:15 crc kubenswrapper[4846]: I0202 12:26:15.278808 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:15 crc kubenswrapper[4846]: I0202 12:26:15.320236 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:16 crc kubenswrapper[4846]: I0202 12:26:16.052891 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.134955 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dkwgn"] Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.349899 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf"] Feb 02 12:26:17 crc kubenswrapper[4846]: E0202 12:26:17.350144 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0841d2fe-5012-47c0-a591-849d4f8cc0f5" containerName="pull" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.350167 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0841d2fe-5012-47c0-a591-849d4f8cc0f5" containerName="pull" Feb 02 12:26:17 crc kubenswrapper[4846]: E0202 12:26:17.350180 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0841d2fe-5012-47c0-a591-849d4f8cc0f5" containerName="extract" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.350186 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0841d2fe-5012-47c0-a591-849d4f8cc0f5" containerName="extract" Feb 02 12:26:17 crc kubenswrapper[4846]: E0202 12:26:17.350210 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0841d2fe-5012-47c0-a591-849d4f8cc0f5" containerName="util" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.350219 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0841d2fe-5012-47c0-a591-849d4f8cc0f5" containerName="util" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.350365 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0841d2fe-5012-47c0-a591-849d4f8cc0f5" containerName="extract" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.350774 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.355895 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-ddgfp" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.400039 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf"] Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.542908 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dr58\" (UniqueName: \"kubernetes.io/projected/80b73f30-8c98-477e-9f49-f7a39a5cf559-kube-api-access-2dr58\") pod \"openstack-operator-controller-init-6bf6665fd-bzplf\" (UID: \"80b73f30-8c98-477e-9f49-f7a39a5cf559\") " pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.644289 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dr58\" (UniqueName: \"kubernetes.io/projected/80b73f30-8c98-477e-9f49-f7a39a5cf559-kube-api-access-2dr58\") pod \"openstack-operator-controller-init-6bf6665fd-bzplf\" (UID: \"80b73f30-8c98-477e-9f49-f7a39a5cf559\") " pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.662129 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dr58\" (UniqueName: \"kubernetes.io/projected/80b73f30-8c98-477e-9f49-f7a39a5cf559-kube-api-access-2dr58\") pod \"openstack-operator-controller-init-6bf6665fd-bzplf\" (UID: \"80b73f30-8c98-477e-9f49-f7a39a5cf559\") " pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf" Feb 02 12:26:17 crc kubenswrapper[4846]: I0202 12:26:17.668923 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf" Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.020065 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dkwgn" podUID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerName="registry-server" containerID="cri-o://bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42" gracePeriod=2 Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.141066 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf"] Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.332418 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.454303 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-utilities\") pod \"30e3636e-eb78-430f-82eb-957e38c9a2c7\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.454448 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-catalog-content\") pod \"30e3636e-eb78-430f-82eb-957e38c9a2c7\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.454949 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sh7m\" (UniqueName: \"kubernetes.io/projected/30e3636e-eb78-430f-82eb-957e38c9a2c7-kube-api-access-6sh7m\") pod \"30e3636e-eb78-430f-82eb-957e38c9a2c7\" (UID: \"30e3636e-eb78-430f-82eb-957e38c9a2c7\") " Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.455337 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-utilities" (OuterVolumeSpecName: "utilities") pod "30e3636e-eb78-430f-82eb-957e38c9a2c7" (UID: "30e3636e-eb78-430f-82eb-957e38c9a2c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.460828 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30e3636e-eb78-430f-82eb-957e38c9a2c7-kube-api-access-6sh7m" (OuterVolumeSpecName: "kube-api-access-6sh7m") pod "30e3636e-eb78-430f-82eb-957e38c9a2c7" (UID: "30e3636e-eb78-430f-82eb-957e38c9a2c7"). InnerVolumeSpecName "kube-api-access-6sh7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.503877 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30e3636e-eb78-430f-82eb-957e38c9a2c7" (UID: "30e3636e-eb78-430f-82eb-957e38c9a2c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.556155 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sh7m\" (UniqueName: \"kubernetes.io/projected/30e3636e-eb78-430f-82eb-957e38c9a2c7-kube-api-access-6sh7m\") on node \"crc\" DevicePath \"\"" Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.556189 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:26:18 crc kubenswrapper[4846]: I0202 12:26:18.556198 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30e3636e-eb78-430f-82eb-957e38c9a2c7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.036931 4846 generic.go:334] "Generic (PLEG): container finished" podID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerID="bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42" exitCode=0 Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.037007 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkwgn" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.036997 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkwgn" event={"ID":"30e3636e-eb78-430f-82eb-957e38c9a2c7","Type":"ContainerDied","Data":"bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42"} Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.037420 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkwgn" event={"ID":"30e3636e-eb78-430f-82eb-957e38c9a2c7","Type":"ContainerDied","Data":"7a1645e335d859b9256c778daee1b120c2226bbc58ea6bf6b54dfe9697bbf589"} Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.037448 4846 scope.go:117] "RemoveContainer" containerID="bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.039242 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf" event={"ID":"80b73f30-8c98-477e-9f49-f7a39a5cf559","Type":"ContainerStarted","Data":"cee62858d5215b8dec7cf21e9d58610d574c5331396c8379b5ad83192b4f48f6"} Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.074141 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dkwgn"] Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.078700 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dkwgn"] Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.079045 4846 scope.go:117] "RemoveContainer" containerID="ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.095260 4846 scope.go:117] "RemoveContainer" containerID="6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.126289 4846 scope.go:117] "RemoveContainer" containerID="bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42" Feb 02 12:26:19 crc kubenswrapper[4846]: E0202 12:26:19.126744 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42\": container with ID starting with bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42 not found: ID does not exist" containerID="bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.126776 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42"} err="failed to get container status \"bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42\": rpc error: code = NotFound desc = could not find container \"bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42\": container with ID starting with bfa28d864026e7e4b3620a54d841070dc299b0b16d19600d3b999fd217aeac42 not found: ID does not exist" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.126800 4846 scope.go:117] "RemoveContainer" containerID="ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154" Feb 02 12:26:19 crc kubenswrapper[4846]: E0202 12:26:19.127085 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154\": container with ID starting with ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154 not found: ID does not exist" containerID="ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.127106 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154"} err="failed to get container status \"ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154\": rpc error: code = NotFound desc = could not find container \"ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154\": container with ID starting with ac8b27de5d7de3b02b692d838df08d5fb306b46f982c1b96646c2c0796f49154 not found: ID does not exist" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.127119 4846 scope.go:117] "RemoveContainer" containerID="6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3" Feb 02 12:26:19 crc kubenswrapper[4846]: E0202 12:26:19.127322 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3\": container with ID starting with 6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3 not found: ID does not exist" containerID="6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.127337 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3"} err="failed to get container status \"6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3\": rpc error: code = NotFound desc = could not find container \"6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3\": container with ID starting with 6e3b834e874fae0531cfc4bf80fc6a1740f7f36b2dcc8456818efbc270cefba3 not found: ID does not exist" Feb 02 12:26:19 crc kubenswrapper[4846]: I0202 12:26:19.454971 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30e3636e-eb78-430f-82eb-957e38c9a2c7" path="/var/lib/kubelet/pods/30e3636e-eb78-430f-82eb-957e38c9a2c7/volumes" Feb 02 12:26:23 crc kubenswrapper[4846]: I0202 12:26:23.085407 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf" event={"ID":"80b73f30-8c98-477e-9f49-f7a39a5cf559","Type":"ContainerStarted","Data":"50947a7ebe586d5a5f8f2c6c2b441c37d2c2fe135ffa1172933c91acc100a167"} Feb 02 12:26:23 crc kubenswrapper[4846]: I0202 12:26:23.085866 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf" Feb 02 12:26:23 crc kubenswrapper[4846]: I0202 12:26:23.122069 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf" podStartSLOduration=1.8525515769999998 podStartE2EDuration="6.122045115s" podCreationTimestamp="2026-02-02 12:26:17 +0000 UTC" firstStartedPulling="2026-02-02 12:26:18.15275685 +0000 UTC m=+1009.381343713" lastFinishedPulling="2026-02-02 12:26:22.422250398 +0000 UTC m=+1013.650837251" observedRunningTime="2026-02-02 12:26:23.118934379 +0000 UTC m=+1014.347521242" watchObservedRunningTime="2026-02-02 12:26:23.122045115 +0000 UTC m=+1014.350631988" Feb 02 12:26:27 crc kubenswrapper[4846]: I0202 12:26:27.671772 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-6bf6665fd-bzplf" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.814765 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8"] Feb 02 12:26:46 crc kubenswrapper[4846]: E0202 12:26:46.815375 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerName="extract-content" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.815385 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerName="extract-content" Feb 02 12:26:46 crc kubenswrapper[4846]: E0202 12:26:46.815399 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerName="extract-utilities" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.815405 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerName="extract-utilities" Feb 02 12:26:46 crc kubenswrapper[4846]: E0202 12:26:46.815418 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerName="registry-server" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.815425 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerName="registry-server" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.820920 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="30e3636e-eb78-430f-82eb-957e38c9a2c7" containerName="registry-server" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.821657 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.840790 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-6f79t" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.845545 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8"] Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.850609 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz"] Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.851406 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.853266 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-vgdkq" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.855432 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td"] Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.856225 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.859044 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-cnz8v" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.874218 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw"] Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.874965 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.876489 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-2fhzp" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.890368 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td"] Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.912684 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6"] Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.913472 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.915522 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-dt5vn" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.932268 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw"] Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.939172 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6"] Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.941827 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g28lk\" (UniqueName: \"kubernetes.io/projected/c0f1cd1a-da1b-4292-8622-87119d40ec03-kube-api-access-g28lk\") pod \"barbican-operator-controller-manager-7b6c4d8c5f-8lrz8\" (UID: \"c0f1cd1a-da1b-4292-8622-87119d40ec03\") " pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.955710 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f"] Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.956519 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.959146 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-62krx" Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.964835 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz"] Feb 02 12:26:46 crc kubenswrapper[4846]: I0202 12:26:46.978719 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.001874 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.002797 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.004652 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dwslt" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.004801 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.024239 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.026387 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.027354 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.029363 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-xv9qf" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.033279 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.036388 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.040749 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4jqn6" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.042542 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgtm6\" (UniqueName: \"kubernetes.io/projected/478eaaaa-5884-4b31-a87e-8655bedb96f1-kube-api-access-lgtm6\") pod \"heat-operator-controller-manager-69d6db494d-jscz6\" (UID: \"478eaaaa-5884-4b31-a87e-8655bedb96f1\") " pod="openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.042591 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdxcf\" (UniqueName: \"kubernetes.io/projected/6ba95849-98ff-4d50-8c41-605aa99c3b2f-kube-api-access-fdxcf\") pod \"cinder-operator-controller-manager-8d874c8fc-hc6td\" (UID: \"6ba95849-98ff-4d50-8c41-605aa99c3b2f\") " pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.042657 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m9ht\" (UniqueName: \"kubernetes.io/projected/b9c7230d-0e38-4540-9074-09a47500dd40-kube-api-access-5m9ht\") pod \"glance-operator-controller-manager-8886f4c47-bklmw\" (UID: \"b9c7230d-0e38-4540-9074-09a47500dd40\") " pod="openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.042690 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g28lk\" (UniqueName: \"kubernetes.io/projected/c0f1cd1a-da1b-4292-8622-87119d40ec03-kube-api-access-g28lk\") pod \"barbican-operator-controller-manager-7b6c4d8c5f-8lrz8\" (UID: \"c0f1cd1a-da1b-4292-8622-87119d40ec03\") " pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.042724 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxjkk\" (UniqueName: \"kubernetes.io/projected/fced078c-1f83-492e-9f21-e9e82c1f9275-kube-api-access-sxjkk\") pod \"designate-operator-controller-manager-6d9697b7f4-88lrz\" (UID: \"fced078c-1f83-492e-9f21-e9e82c1f9275\") " pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.068425 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.069194 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.089317 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.090138 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.099013 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.099126 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-wlfq7" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.099522 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7mc72" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.102637 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.104532 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g28lk\" (UniqueName: \"kubernetes.io/projected/c0f1cd1a-da1b-4292-8622-87119d40ec03-kube-api-access-g28lk\") pod \"barbican-operator-controller-manager-7b6c4d8c5f-8lrz8\" (UID: \"c0f1cd1a-da1b-4292-8622-87119d40ec03\") " pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.118314 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.118350 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.131396 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.132367 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.136865 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-rvzmw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.144200 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m9ht\" (UniqueName: \"kubernetes.io/projected/b9c7230d-0e38-4540-9074-09a47500dd40-kube-api-access-5m9ht\") pod \"glance-operator-controller-manager-8886f4c47-bklmw\" (UID: \"b9c7230d-0e38-4540-9074-09a47500dd40\") " pod="openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.144256 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82klx\" (UniqueName: \"kubernetes.io/projected/69f3343d-ca8a-4ff2-bc0c-14121a6938a8-kube-api-access-82klx\") pod \"ironic-operator-controller-manager-5f4b8bd54d-7lr2n\" (UID: \"69f3343d-ca8a-4ff2-bc0c-14121a6938a8\") " pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.144306 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkjm4\" (UniqueName: \"kubernetes.io/projected/46d6dfa1-87b3-4e53-bcfd-27c709ad7911-kube-api-access-qkjm4\") pod \"horizon-operator-controller-manager-5fb775575f-cz76f\" (UID: \"46d6dfa1-87b3-4e53-bcfd-27c709ad7911\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.144345 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hs5t\" (UniqueName: \"kubernetes.io/projected/a70f0094-0775-49ea-9002-9f3da3ff87d1-kube-api-access-5hs5t\") pod \"keystone-operator-controller-manager-84f48565d4-hfvt6\" (UID: \"a70f0094-0775-49ea-9002-9f3da3ff87d1\") " pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.144372 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxjkk\" (UniqueName: \"kubernetes.io/projected/fced078c-1f83-492e-9f21-e9e82c1f9275-kube-api-access-sxjkk\") pod \"designate-operator-controller-manager-6d9697b7f4-88lrz\" (UID: \"fced078c-1f83-492e-9f21-e9e82c1f9275\") " pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.144419 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgtm6\" (UniqueName: \"kubernetes.io/projected/478eaaaa-5884-4b31-a87e-8655bedb96f1-kube-api-access-lgtm6\") pod \"heat-operator-controller-manager-69d6db494d-jscz6\" (UID: \"478eaaaa-5884-4b31-a87e-8655bedb96f1\") " pod="openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.144461 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdxcf\" (UniqueName: \"kubernetes.io/projected/6ba95849-98ff-4d50-8c41-605aa99c3b2f-kube-api-access-fdxcf\") pod \"cinder-operator-controller-manager-8d874c8fc-hc6td\" (UID: \"6ba95849-98ff-4d50-8c41-605aa99c3b2f\") " pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.144491 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wpp5\" (UniqueName: \"kubernetes.io/projected/f75d8c59-636e-4496-bd81-e4e53f37a713-kube-api-access-5wpp5\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.144538 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.156521 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.157236 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.161060 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.162006 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.163937 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-g2dvx" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.169147 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxjkk\" (UniqueName: \"kubernetes.io/projected/fced078c-1f83-492e-9f21-e9e82c1f9275-kube-api-access-sxjkk\") pod \"designate-operator-controller-manager-6d9697b7f4-88lrz\" (UID: \"fced078c-1f83-492e-9f21-e9e82c1f9275\") " pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.169421 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdxcf\" (UniqueName: \"kubernetes.io/projected/6ba95849-98ff-4d50-8c41-605aa99c3b2f-kube-api-access-fdxcf\") pod \"cinder-operator-controller-manager-8d874c8fc-hc6td\" (UID: \"6ba95849-98ff-4d50-8c41-605aa99c3b2f\") " pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.171703 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgtm6\" (UniqueName: \"kubernetes.io/projected/478eaaaa-5884-4b31-a87e-8655bedb96f1-kube-api-access-lgtm6\") pod \"heat-operator-controller-manager-69d6db494d-jscz6\" (UID: \"478eaaaa-5884-4b31-a87e-8655bedb96f1\") " pod="openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.172658 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.182712 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.183502 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.184983 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m9ht\" (UniqueName: \"kubernetes.io/projected/b9c7230d-0e38-4540-9074-09a47500dd40-kube-api-access-5m9ht\") pod \"glance-operator-controller-manager-8886f4c47-bklmw\" (UID: \"b9c7230d-0e38-4540-9074-09a47500dd40\") " pod="openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.186025 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.197831 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-997tz" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.200154 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.219412 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.230600 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.240083 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.243545 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-2lcx4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.245726 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.258465 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.258546 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdw7t\" (UniqueName: \"kubernetes.io/projected/2e0bfdb6-f6e5-41c4-ab53-f13863d687b8-kube-api-access-bdw7t\") pod \"mariadb-operator-controller-manager-67bf948998-nvnpj\" (UID: \"2e0bfdb6-f6e5-41c4-ab53-f13863d687b8\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.258758 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82klx\" (UniqueName: \"kubernetes.io/projected/69f3343d-ca8a-4ff2-bc0c-14121a6938a8-kube-api-access-82klx\") pod \"ironic-operator-controller-manager-5f4b8bd54d-7lr2n\" (UID: \"69f3343d-ca8a-4ff2-bc0c-14121a6938a8\") " pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.258856 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkjm4\" (UniqueName: \"kubernetes.io/projected/46d6dfa1-87b3-4e53-bcfd-27c709ad7911-kube-api-access-qkjm4\") pod \"horizon-operator-controller-manager-5fb775575f-cz76f\" (UID: \"46d6dfa1-87b3-4e53-bcfd-27c709ad7911\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.258901 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hs5t\" (UniqueName: \"kubernetes.io/projected/a70f0094-0775-49ea-9002-9f3da3ff87d1-kube-api-access-5hs5t\") pod \"keystone-operator-controller-manager-84f48565d4-hfvt6\" (UID: \"a70f0094-0775-49ea-9002-9f3da3ff87d1\") " pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.258934 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrrd7\" (UniqueName: \"kubernetes.io/projected/d6f0f90b-675e-40e9-b541-7ff04333a885-kube-api-access-jrrd7\") pod \"manila-operator-controller-manager-7dd968899f-5vzvg\" (UID: \"d6f0f90b-675e-40e9-b541-7ff04333a885\") " pod="openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg" Feb 02 12:26:47 crc kubenswrapper[4846]: E0202 12:26:47.258977 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 12:26:47 crc kubenswrapper[4846]: E0202 12:26:47.259024 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert podName:f75d8c59-636e-4496-bd81-e4e53f37a713 nodeName:}" failed. No retries permitted until 2026-02-02 12:26:47.759008527 +0000 UTC m=+1038.987595390 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert") pod "infra-operator-controller-manager-79955696d6-c6sk4" (UID: "f75d8c59-636e-4496-bd81-e4e53f37a713") : secret "infra-operator-webhook-server-cert" not found Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.258979 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qmmm\" (UniqueName: \"kubernetes.io/projected/f8ebd280-c5fa-4cf3-890b-93a018bcb2d2-kube-api-access-7qmmm\") pod \"neutron-operator-controller-manager-585dbc889-hrnrk\" (UID: \"f8ebd280-c5fa-4cf3-890b-93a018bcb2d2\") " pod="openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.259469 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wpp5\" (UniqueName: \"kubernetes.io/projected/f75d8c59-636e-4496-bd81-e4e53f37a713-kube-api-access-5wpp5\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.262805 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.263860 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.265378 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.266386 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-qzw99" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.269308 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.274971 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.276871 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wpp5\" (UniqueName: \"kubernetes.io/projected/f75d8c59-636e-4496-bd81-e4e53f37a713-kube-api-access-5wpp5\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.279510 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkjm4\" (UniqueName: \"kubernetes.io/projected/46d6dfa1-87b3-4e53-bcfd-27c709ad7911-kube-api-access-qkjm4\") pod \"horizon-operator-controller-manager-5fb775575f-cz76f\" (UID: \"46d6dfa1-87b3-4e53-bcfd-27c709ad7911\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.280382 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.281220 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.283132 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-7j7xq" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.283808 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.286373 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82klx\" (UniqueName: \"kubernetes.io/projected/69f3343d-ca8a-4ff2-bc0c-14121a6938a8-kube-api-access-82klx\") pod \"ironic-operator-controller-manager-5f4b8bd54d-7lr2n\" (UID: \"69f3343d-ca8a-4ff2-bc0c-14121a6938a8\") " pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.286975 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hs5t\" (UniqueName: \"kubernetes.io/projected/a70f0094-0775-49ea-9002-9f3da3ff87d1-kube-api-access-5hs5t\") pod \"keystone-operator-controller-manager-84f48565d4-hfvt6\" (UID: \"a70f0094-0775-49ea-9002-9f3da3ff87d1\") " pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.297909 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.306330 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.316151 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.324487 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.327719 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-pljx4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.334268 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.346140 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.347320 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.351013 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.360196 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-sw8lv" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.363251 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chs5q\" (UniqueName: \"kubernetes.io/projected/2497b300-a910-41c8-b70d-1c9acd871a90-kube-api-access-chs5q\") pod \"ovn-operator-controller-manager-788c46999f-tw44b\" (UID: \"2497b300-a910-41c8-b70d-1c9acd871a90\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.363415 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdw7t\" (UniqueName: \"kubernetes.io/projected/2e0bfdb6-f6e5-41c4-ab53-f13863d687b8-kube-api-access-bdw7t\") pod \"mariadb-operator-controller-manager-67bf948998-nvnpj\" (UID: \"2e0bfdb6-f6e5-41c4-ab53-f13863d687b8\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.363480 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfzjm\" (UniqueName: \"kubernetes.io/projected/8d9a7d64-a757-48b8-bdb9-4ec3e2921321-kube-api-access-wfzjm\") pod \"octavia-operator-controller-manager-6687f8d877-vfmzp\" (UID: \"8d9a7d64-a757-48b8-bdb9-4ec3e2921321\") " pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.363556 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj46z\" (UniqueName: \"kubernetes.io/projected/f4ec43b3-d763-4478-8ef9-0453a2fb730a-kube-api-access-dj46z\") pod \"nova-operator-controller-manager-55bff696bd-89dd7\" (UID: \"f4ec43b3-d763-4478-8ef9-0453a2fb730a\") " pod="openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.363585 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrrd7\" (UniqueName: \"kubernetes.io/projected/d6f0f90b-675e-40e9-b541-7ff04333a885-kube-api-access-jrrd7\") pod \"manila-operator-controller-manager-7dd968899f-5vzvg\" (UID: \"d6f0f90b-675e-40e9-b541-7ff04333a885\") " pod="openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.364536 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qmmm\" (UniqueName: \"kubernetes.io/projected/f8ebd280-c5fa-4cf3-890b-93a018bcb2d2-kube-api-access-7qmmm\") pod \"neutron-operator-controller-manager-585dbc889-hrnrk\" (UID: \"f8ebd280-c5fa-4cf3-890b-93a018bcb2d2\") " pod="openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.372486 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.376962 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.384363 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.384539 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.397294 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-cgqpt" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.402336 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qmmm\" (UniqueName: \"kubernetes.io/projected/f8ebd280-c5fa-4cf3-890b-93a018bcb2d2-kube-api-access-7qmmm\") pod \"neutron-operator-controller-manager-585dbc889-hrnrk\" (UID: \"f8ebd280-c5fa-4cf3-890b-93a018bcb2d2\") " pod="openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.408285 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdw7t\" (UniqueName: \"kubernetes.io/projected/2e0bfdb6-f6e5-41c4-ab53-f13863d687b8-kube-api-access-bdw7t\") pod \"mariadb-operator-controller-manager-67bf948998-nvnpj\" (UID: \"2e0bfdb6-f6e5-41c4-ab53-f13863d687b8\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.408553 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrrd7\" (UniqueName: \"kubernetes.io/projected/d6f0f90b-675e-40e9-b541-7ff04333a885-kube-api-access-jrrd7\") pod \"manila-operator-controller-manager-7dd968899f-5vzvg\" (UID: \"d6f0f90b-675e-40e9-b541-7ff04333a885\") " pod="openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.420467 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.433995 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.458153 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.467111 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chs5q\" (UniqueName: \"kubernetes.io/projected/2497b300-a910-41c8-b70d-1c9acd871a90-kube-api-access-chs5q\") pod \"ovn-operator-controller-manager-788c46999f-tw44b\" (UID: \"2497b300-a910-41c8-b70d-1c9acd871a90\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.467838 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brmmw\" (UniqueName: \"kubernetes.io/projected/204fa43c-5360-4282-b002-ae22e0f0595d-kube-api-access-brmmw\") pod \"test-operator-controller-manager-56f8bfcd9f-6g958\" (UID: \"204fa43c-5360-4282-b002-ae22e0f0595d\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.467932 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ghwx\" (UniqueName: \"kubernetes.io/projected/9c5ac153-1fd9-491b-942d-2b0b3e3e9da5-kube-api-access-5ghwx\") pod \"swift-operator-controller-manager-68fc8c869-kkbb8\" (UID: \"9c5ac153-1fd9-491b-942d-2b0b3e3e9da5\") " pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.467957 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdmp7\" (UniqueName: \"kubernetes.io/projected/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-kube-api-access-cdmp7\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.467987 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5xc2\" (UniqueName: \"kubernetes.io/projected/cb8f2427-7089-4ecf-8ae9-4ece48067de8-kube-api-access-w5xc2\") pod \"telemetry-operator-controller-manager-64b5b76f97-xk9n4\" (UID: \"cb8f2427-7089-4ecf-8ae9-4ece48067de8\") " pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.468036 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfzjm\" (UniqueName: \"kubernetes.io/projected/8d9a7d64-a757-48b8-bdb9-4ec3e2921321-kube-api-access-wfzjm\") pod \"octavia-operator-controller-manager-6687f8d877-vfmzp\" (UID: \"8d9a7d64-a757-48b8-bdb9-4ec3e2921321\") " pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.468064 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj46z\" (UniqueName: \"kubernetes.io/projected/f4ec43b3-d763-4478-8ef9-0453a2fb730a-kube-api-access-dj46z\") pod \"nova-operator-controller-manager-55bff696bd-89dd7\" (UID: \"f4ec43b3-d763-4478-8ef9-0453a2fb730a\") " pod="openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.468136 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdsrf\" (UniqueName: \"kubernetes.io/projected/124b64f2-5773-4d35-ab14-f4ad0fe4f318-kube-api-access-gdsrf\") pod \"placement-operator-controller-manager-5b964cf4cd-pqnqm\" (UID: \"124b64f2-5773-4d35-ab14-f4ad0fe4f318\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.468168 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.471741 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-8gkkw"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.472533 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-564965969-8gkkw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.483956 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-8gkkw"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.497099 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj46z\" (UniqueName: \"kubernetes.io/projected/f4ec43b3-d763-4478-8ef9-0453a2fb730a-kube-api-access-dj46z\") pod \"nova-operator-controller-manager-55bff696bd-89dd7\" (UID: \"f4ec43b3-d763-4478-8ef9-0453a2fb730a\") " pod="openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.503174 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-hlsnh" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.505635 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfzjm\" (UniqueName: \"kubernetes.io/projected/8d9a7d64-a757-48b8-bdb9-4ec3e2921321-kube-api-access-wfzjm\") pod \"octavia-operator-controller-manager-6687f8d877-vfmzp\" (UID: \"8d9a7d64-a757-48b8-bdb9-4ec3e2921321\") " pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.516422 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chs5q\" (UniqueName: \"kubernetes.io/projected/2497b300-a910-41c8-b70d-1c9acd871a90-kube-api-access-chs5q\") pod \"ovn-operator-controller-manager-788c46999f-tw44b\" (UID: \"2497b300-a910-41c8-b70d-1c9acd871a90\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.569690 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdsrf\" (UniqueName: \"kubernetes.io/projected/124b64f2-5773-4d35-ab14-f4ad0fe4f318-kube-api-access-gdsrf\") pod \"placement-operator-controller-manager-5b964cf4cd-pqnqm\" (UID: \"124b64f2-5773-4d35-ab14-f4ad0fe4f318\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.569772 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.569832 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brmmw\" (UniqueName: \"kubernetes.io/projected/204fa43c-5360-4282-b002-ae22e0f0595d-kube-api-access-brmmw\") pod \"test-operator-controller-manager-56f8bfcd9f-6g958\" (UID: \"204fa43c-5360-4282-b002-ae22e0f0595d\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" Feb 02 12:26:47 crc kubenswrapper[4846]: E0202 12:26:47.572863 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:26:47 crc kubenswrapper[4846]: E0202 12:26:47.572956 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert podName:b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1 nodeName:}" failed. No retries permitted until 2026-02-02 12:26:48.072931236 +0000 UTC m=+1039.301518099 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" (UID: "b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.573086 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ghwx\" (UniqueName: \"kubernetes.io/projected/9c5ac153-1fd9-491b-942d-2b0b3e3e9da5-kube-api-access-5ghwx\") pod \"swift-operator-controller-manager-68fc8c869-kkbb8\" (UID: \"9c5ac153-1fd9-491b-942d-2b0b3e3e9da5\") " pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.573159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdmp7\" (UniqueName: \"kubernetes.io/projected/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-kube-api-access-cdmp7\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.573202 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5xc2\" (UniqueName: \"kubernetes.io/projected/cb8f2427-7089-4ecf-8ae9-4ece48067de8-kube-api-access-w5xc2\") pod \"telemetry-operator-controller-manager-64b5b76f97-xk9n4\" (UID: \"cb8f2427-7089-4ecf-8ae9-4ece48067de8\") " pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.589566 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.600181 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ghwx\" (UniqueName: \"kubernetes.io/projected/9c5ac153-1fd9-491b-942d-2b0b3e3e9da5-kube-api-access-5ghwx\") pod \"swift-operator-controller-manager-68fc8c869-kkbb8\" (UID: \"9c5ac153-1fd9-491b-942d-2b0b3e3e9da5\") " pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.601158 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.602259 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.608504 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5xc2\" (UniqueName: \"kubernetes.io/projected/cb8f2427-7089-4ecf-8ae9-4ece48067de8-kube-api-access-w5xc2\") pod \"telemetry-operator-controller-manager-64b5b76f97-xk9n4\" (UID: \"cb8f2427-7089-4ecf-8ae9-4ece48067de8\") " pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.608509 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-dgrk9" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.609809 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.611910 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.613714 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdmp7\" (UniqueName: \"kubernetes.io/projected/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-kube-api-access-cdmp7\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.614092 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.624142 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brmmw\" (UniqueName: \"kubernetes.io/projected/204fa43c-5360-4282-b002-ae22e0f0595d-kube-api-access-brmmw\") pod \"test-operator-controller-manager-56f8bfcd9f-6g958\" (UID: \"204fa43c-5360-4282-b002-ae22e0f0595d\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.624340 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.634756 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdsrf\" (UniqueName: \"kubernetes.io/projected/124b64f2-5773-4d35-ab14-f4ad0fe4f318-kube-api-access-gdsrf\") pod \"placement-operator-controller-manager-5b964cf4cd-pqnqm\" (UID: \"124b64f2-5773-4d35-ab14-f4ad0fe4f318\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.644005 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.676869 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.678073 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.678150 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf9jm\" (UniqueName: \"kubernetes.io/projected/8ee9f8c5-8078-4ad1-b7c8-4a8e2dfd4e41-kube-api-access-tf9jm\") pod \"watcher-operator-controller-manager-564965969-8gkkw\" (UID: \"8ee9f8c5-8078-4ad1-b7c8-4a8e2dfd4e41\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-8gkkw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.678206 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4pbb\" (UniqueName: \"kubernetes.io/projected/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-kube-api-access-b4pbb\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.678240 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.703369 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.704923 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.708824 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-cxffd" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.719971 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.778895 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.788461 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.788523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.788581 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw4df\" (UniqueName: \"kubernetes.io/projected/b45d49d2-cad0-49fa-addf-766dba1544a4-kube-api-access-pw4df\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9tnhp\" (UID: \"b45d49d2-cad0-49fa-addf-766dba1544a4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.788634 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.788703 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf9jm\" (UniqueName: \"kubernetes.io/projected/8ee9f8c5-8078-4ad1-b7c8-4a8e2dfd4e41-kube-api-access-tf9jm\") pod \"watcher-operator-controller-manager-564965969-8gkkw\" (UID: \"8ee9f8c5-8078-4ad1-b7c8-4a8e2dfd4e41\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-8gkkw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.788751 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4pbb\" (UniqueName: \"kubernetes.io/projected/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-kube-api-access-b4pbb\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:47 crc kubenswrapper[4846]: E0202 12:26:47.789147 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 12:26:47 crc kubenswrapper[4846]: E0202 12:26:47.789199 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:26:48.289181475 +0000 UTC m=+1039.517768328 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "metrics-server-cert" not found Feb 02 12:26:47 crc kubenswrapper[4846]: E0202 12:26:47.789253 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 12:26:47 crc kubenswrapper[4846]: E0202 12:26:47.789279 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:26:48.289271088 +0000 UTC m=+1039.517858051 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "webhook-server-cert" not found Feb 02 12:26:47 crc kubenswrapper[4846]: E0202 12:26:47.789299 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 12:26:47 crc kubenswrapper[4846]: E0202 12:26:47.789397 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert podName:f75d8c59-636e-4496-bd81-e4e53f37a713 nodeName:}" failed. No retries permitted until 2026-02-02 12:26:48.78936784 +0000 UTC m=+1040.017954773 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert") pod "infra-operator-controller-manager-79955696d6-c6sk4" (UID: "f75d8c59-636e-4496-bd81-e4e53f37a713") : secret "infra-operator-webhook-server-cert" not found Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.818219 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf9jm\" (UniqueName: \"kubernetes.io/projected/8ee9f8c5-8078-4ad1-b7c8-4a8e2dfd4e41-kube-api-access-tf9jm\") pod \"watcher-operator-controller-manager-564965969-8gkkw\" (UID: \"8ee9f8c5-8078-4ad1-b7c8-4a8e2dfd4e41\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-8gkkw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.821122 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4pbb\" (UniqueName: \"kubernetes.io/projected/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-kube-api-access-b4pbb\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.821685 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.854927 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.877109 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8"] Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.887072 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.890612 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw4df\" (UniqueName: \"kubernetes.io/projected/b45d49d2-cad0-49fa-addf-766dba1544a4-kube-api-access-pw4df\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9tnhp\" (UID: \"b45d49d2-cad0-49fa-addf-766dba1544a4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" Feb 02 12:26:47 crc kubenswrapper[4846]: W0202 12:26:47.891579 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0f1cd1a_da1b_4292_8622_87119d40ec03.slice/crio-9af15bc93da3db72d50c32fac712d84b6aead34d8332566500baaa18244e5f43 WatchSource:0}: Error finding container 9af15bc93da3db72d50c32fac712d84b6aead34d8332566500baaa18244e5f43: Status 404 returned error can't find the container with id 9af15bc93da3db72d50c32fac712d84b6aead34d8332566500baaa18244e5f43 Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.912473 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-564965969-8gkkw" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.918817 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw4df\" (UniqueName: \"kubernetes.io/projected/b45d49d2-cad0-49fa-addf-766dba1544a4-kube-api-access-pw4df\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9tnhp\" (UID: \"b45d49d2-cad0-49fa-addf-766dba1544a4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" Feb 02 12:26:47 crc kubenswrapper[4846]: I0202 12:26:47.957537 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td"] Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.078070 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.099318 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:26:48 crc kubenswrapper[4846]: E0202 12:26:48.099518 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:26:48 crc kubenswrapper[4846]: E0202 12:26:48.099565 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert podName:b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1 nodeName:}" failed. No retries permitted until 2026-02-02 12:26:49.099551426 +0000 UTC m=+1040.328138289 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" (UID: "b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.303090 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8" event={"ID":"c0f1cd1a-da1b-4292-8622-87119d40ec03","Type":"ContainerStarted","Data":"9af15bc93da3db72d50c32fac712d84b6aead34d8332566500baaa18244e5f43"} Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.303834 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.303918 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:48 crc kubenswrapper[4846]: E0202 12:26:48.304147 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 12:26:48 crc kubenswrapper[4846]: E0202 12:26:48.304198 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 12:26:48 crc kubenswrapper[4846]: E0202 12:26:48.304201 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:26:49.304183311 +0000 UTC m=+1040.532770184 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "webhook-server-cert" not found Feb 02 12:26:48 crc kubenswrapper[4846]: E0202 12:26:48.304242 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:26:49.304232082 +0000 UTC m=+1040.532818945 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "metrics-server-cert" not found Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.310983 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td" event={"ID":"6ba95849-98ff-4d50-8c41-605aa99c3b2f","Type":"ContainerStarted","Data":"dc831e7b650a33aacf87341865566cedc70fc132fa723f7247a3103538956115"} Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.331702 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n"] Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.338178 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz"] Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.344031 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw"] Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.360881 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6"] Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.602493 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f"] Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.610350 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6"] Feb 02 12:26:48 crc kubenswrapper[4846]: W0202 12:26:48.619720 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda70f0094_0775_49ea_9002_9f3da3ff87d1.slice/crio-0ef5dc37fac802ebd72fc9f5be9b50de93236340473956a55222ccd616cea94c WatchSource:0}: Error finding container 0ef5dc37fac802ebd72fc9f5be9b50de93236340473956a55222ccd616cea94c: Status 404 returned error can't find the container with id 0ef5dc37fac802ebd72fc9f5be9b50de93236340473956a55222ccd616cea94c Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.689661 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg"] Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.700708 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b"] Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.712085 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp"] Feb 02 12:26:48 crc kubenswrapper[4846]: W0202 12:26:48.713294 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d9a7d64_a757_48b8_bdb9_4ec3e2921321.slice/crio-da6a028a89c29c24802924e52478caf8ac403a130dcc184ed3cda69cd699c59e WatchSource:0}: Error finding container da6a028a89c29c24802924e52478caf8ac403a130dcc184ed3cda69cd699c59e: Status 404 returned error can't find the container with id da6a028a89c29c24802924e52478caf8ac403a130dcc184ed3cda69cd699c59e Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.811095 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:26:48 crc kubenswrapper[4846]: E0202 12:26:48.811384 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 12:26:48 crc kubenswrapper[4846]: E0202 12:26:48.811461 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert podName:f75d8c59-636e-4496-bd81-e4e53f37a713 nodeName:}" failed. No retries permitted until 2026-02-02 12:26:50.811423879 +0000 UTC m=+1042.040010742 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert") pod "infra-operator-controller-manager-79955696d6-c6sk4" (UID: "f75d8c59-636e-4496-bd81-e4e53f37a713") : secret "infra-operator-webhook-server-cert" not found Feb 02 12:26:48 crc kubenswrapper[4846]: I0202 12:26:48.832544 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj"] Feb 02 12:26:48 crc kubenswrapper[4846]: W0202 12:26:48.839240 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e0bfdb6_f6e5_41c4_ab53_f13863d687b8.slice/crio-1de7b508fa97de55948ef7f1ca73b76fbdf8cf431b91a0ace012df4058e6e5a0 WatchSource:0}: Error finding container 1de7b508fa97de55948ef7f1ca73b76fbdf8cf431b91a0ace012df4058e6e5a0: Status 404 returned error can't find the container with id 1de7b508fa97de55948ef7f1ca73b76fbdf8cf431b91a0ace012df4058e6e5a0 Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.025832 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-8gkkw"] Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.030742 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958"] Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.051608 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk"] Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.059667 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7"] Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.065071 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4"] Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.074991 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm"] Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.083725 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8"] Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.101219 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp"] Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.115494 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:3e01e99d3ca1b6c20b1bb015b00cfcbffc584f22a93dc6fe4019d63b813c0241,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-brmmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-56f8bfcd9f-6g958_openstack-operators(204fa43c-5360-4282-b002-ae22e0f0595d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.115600 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gdsrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b964cf4cd-pqnqm_openstack-operators(124b64f2-5773-4d35-ab14-f4ad0fe4f318): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.116825 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" podUID="124b64f2-5773-4d35-ab14-f4ad0fe4f318" Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.116901 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" podUID="204fa43c-5360-4282-b002-ae22e0f0595d" Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.117702 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.117848 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.117893 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert podName:b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1 nodeName:}" failed. No retries permitted until 2026-02-02 12:26:51.117877104 +0000 UTC m=+1042.346463967 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" (UID: "b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.118472 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:42ad717de1b82267d244b016e5491a5b66a5c3deb6b8c2906a379e1296a2c382,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5ghwx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68fc8c869-kkbb8_openstack-operators(9c5ac153-1fd9-491b-942d-2b0b3e3e9da5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.120461 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" podUID="9c5ac153-1fd9-491b-942d-2b0b3e3e9da5" Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.129826 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f9bf288cd0c13912404027a58ea3b90d4092b641e8265adc5c88644ea7fe901a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w5xc2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-64b5b76f97-xk9n4_openstack-operators(cb8f2427-7089-4ecf-8ae9-4ece48067de8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.131795 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" podUID="cb8f2427-7089-4ecf-8ae9-4ece48067de8" Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.147173 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pw4df,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-9tnhp_openstack-operators(b45d49d2-cad0-49fa-addf-766dba1544a4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.149436 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" podUID="b45d49d2-cad0-49fa-addf-766dba1544a4" Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.318731 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b" event={"ID":"2497b300-a910-41c8-b70d-1c9acd871a90","Type":"ContainerStarted","Data":"6672ee2b6feb01ca50ba1ebc74c4613611a1d131a226cbbe95e876ca1e81ec06"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.319924 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" event={"ID":"204fa43c-5360-4282-b002-ae22e0f0595d","Type":"ContainerStarted","Data":"30b2a2480ab4ca9fc5a1c9bcc2c6099cb6fe8115d67b2b61a39596f8f51397c5"} Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.321740 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:3e01e99d3ca1b6c20b1bb015b00cfcbffc584f22a93dc6fe4019d63b813c0241\\\"\"" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" podUID="204fa43c-5360-4282-b002-ae22e0f0595d" Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.323910 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" event={"ID":"b45d49d2-cad0-49fa-addf-766dba1544a4","Type":"ContainerStarted","Data":"d540f34e276eb5ebd48dd197ce8441674ede2b54030e29fb9bad949df85fd5af"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.326307 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.326392 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.326561 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.326614 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:26:51.326598999 +0000 UTC m=+1042.555185872 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "webhook-server-cert" not found Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.326686 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.326716 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:26:51.326707242 +0000 UTC m=+1042.555294105 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "metrics-server-cert" not found Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.327367 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7" event={"ID":"f4ec43b3-d763-4478-8ef9-0453a2fb730a","Type":"ContainerStarted","Data":"fb24de614a263401f9bde97a58e3128006f4c5d28f7f4e959d4c923d9b5cd397"} Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.332002 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" podUID="b45d49d2-cad0-49fa-addf-766dba1544a4" Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.332258 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6" event={"ID":"478eaaaa-5884-4b31-a87e-8655bedb96f1","Type":"ContainerStarted","Data":"aced1d32d344fb826dcd393495fa5ab3a7cd26317ed61b2fb547c13defab62fc"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.337471 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp" event={"ID":"8d9a7d64-a757-48b8-bdb9-4ec3e2921321","Type":"ContainerStarted","Data":"da6a028a89c29c24802924e52478caf8ac403a130dcc184ed3cda69cd699c59e"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.339599 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-564965969-8gkkw" event={"ID":"8ee9f8c5-8078-4ad1-b7c8-4a8e2dfd4e41","Type":"ContainerStarted","Data":"25367e8a8b8a7b3a02d3258c5afdff165339661afb571607492c3212dd76cbdf"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.342370 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6" event={"ID":"a70f0094-0775-49ea-9002-9f3da3ff87d1","Type":"ContainerStarted","Data":"0ef5dc37fac802ebd72fc9f5be9b50de93236340473956a55222ccd616cea94c"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.343492 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" event={"ID":"9c5ac153-1fd9-491b-942d-2b0b3e3e9da5","Type":"ContainerStarted","Data":"2be3b2ea61a7cc1f0b390ac7ac1e04dd0a0a180722d106459573cd8ba0958d1b"} Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.345660 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:42ad717de1b82267d244b016e5491a5b66a5c3deb6b8c2906a379e1296a2c382\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" podUID="9c5ac153-1fd9-491b-942d-2b0b3e3e9da5" Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.346874 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" event={"ID":"cb8f2427-7089-4ecf-8ae9-4ece48067de8","Type":"ContainerStarted","Data":"81c50eda80d02b3f2a80d4bff7d9a23aa0997b83f64a07c27f58f86824ae7a05"} Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.355340 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f9bf288cd0c13912404027a58ea3b90d4092b641e8265adc5c88644ea7fe901a\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" podUID="cb8f2427-7089-4ecf-8ae9-4ece48067de8" Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.357192 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj" event={"ID":"2e0bfdb6-f6e5-41c4-ab53-f13863d687b8","Type":"ContainerStarted","Data":"1de7b508fa97de55948ef7f1ca73b76fbdf8cf431b91a0ace012df4058e6e5a0"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.366311 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" event={"ID":"124b64f2-5773-4d35-ab14-f4ad0fe4f318","Type":"ContainerStarted","Data":"38595d26c08802f52c6e023eb6fbd39ec054af31380565eb2985df382c78fa87"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.373229 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg" event={"ID":"d6f0f90b-675e-40e9-b541-7ff04333a885","Type":"ContainerStarted","Data":"a096b8f38b9daff5828ae8d197ce62c1f0051a7d2c7e887460aff1b978f3d729"} Feb 02 12:26:49 crc kubenswrapper[4846]: E0202 12:26:49.373374 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" podUID="124b64f2-5773-4d35-ab14-f4ad0fe4f318" Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.381786 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw" event={"ID":"b9c7230d-0e38-4540-9074-09a47500dd40","Type":"ContainerStarted","Data":"a7ae3c50129f697b0bdb3e22ea3d6fc4c409ba8d12067fcb1a0542938a041fa0"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.390681 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk" event={"ID":"f8ebd280-c5fa-4cf3-890b-93a018bcb2d2","Type":"ContainerStarted","Data":"96f38574dee0769c9903cdfb05d9b30c5a60b4569ab1e40e06e2e8fdb3ef6009"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.395056 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n" event={"ID":"69f3343d-ca8a-4ff2-bc0c-14121a6938a8","Type":"ContainerStarted","Data":"64b21c8b369305a830089c3395ca4547c7aea9054783f736d725e07d0b493bf7"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.397395 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f" event={"ID":"46d6dfa1-87b3-4e53-bcfd-27c709ad7911","Type":"ContainerStarted","Data":"aaaf29a4048337ca90a243819121a1a6117085c6ead17951ee6ae688b02fc426"} Feb 02 12:26:49 crc kubenswrapper[4846]: I0202 12:26:49.403307 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz" event={"ID":"fced078c-1f83-492e-9f21-e9e82c1f9275","Type":"ContainerStarted","Data":"8e0bd1b56fec150d5124df76ed2a6f4a4658a14426e77a50b7df055a1b747156"} Feb 02 12:26:50 crc kubenswrapper[4846]: E0202 12:26:50.421595 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" podUID="b45d49d2-cad0-49fa-addf-766dba1544a4" Feb 02 12:26:50 crc kubenswrapper[4846]: E0202 12:26:50.421602 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f9bf288cd0c13912404027a58ea3b90d4092b641e8265adc5c88644ea7fe901a\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" podUID="cb8f2427-7089-4ecf-8ae9-4ece48067de8" Feb 02 12:26:50 crc kubenswrapper[4846]: E0202 12:26:50.422860 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:3e01e99d3ca1b6c20b1bb015b00cfcbffc584f22a93dc6fe4019d63b813c0241\\\"\"" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" podUID="204fa43c-5360-4282-b002-ae22e0f0595d" Feb 02 12:26:50 crc kubenswrapper[4846]: E0202 12:26:50.422927 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:42ad717de1b82267d244b016e5491a5b66a5c3deb6b8c2906a379e1296a2c382\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" podUID="9c5ac153-1fd9-491b-942d-2b0b3e3e9da5" Feb 02 12:26:50 crc kubenswrapper[4846]: E0202 12:26:50.423575 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" podUID="124b64f2-5773-4d35-ab14-f4ad0fe4f318" Feb 02 12:26:50 crc kubenswrapper[4846]: I0202 12:26:50.847088 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:26:50 crc kubenswrapper[4846]: E0202 12:26:50.847265 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 12:26:50 crc kubenswrapper[4846]: E0202 12:26:50.847336 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert podName:f75d8c59-636e-4496-bd81-e4e53f37a713 nodeName:}" failed. No retries permitted until 2026-02-02 12:26:54.847315706 +0000 UTC m=+1046.075902569 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert") pod "infra-operator-controller-manager-79955696d6-c6sk4" (UID: "f75d8c59-636e-4496-bd81-e4e53f37a713") : secret "infra-operator-webhook-server-cert" not found Feb 02 12:26:51 crc kubenswrapper[4846]: I0202 12:26:51.150324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:26:51 crc kubenswrapper[4846]: E0202 12:26:51.150532 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:26:51 crc kubenswrapper[4846]: E0202 12:26:51.150754 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert podName:b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1 nodeName:}" failed. No retries permitted until 2026-02-02 12:26:55.150730786 +0000 UTC m=+1046.379317649 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" (UID: "b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:26:51 crc kubenswrapper[4846]: I0202 12:26:51.354182 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:51 crc kubenswrapper[4846]: I0202 12:26:51.354265 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:51 crc kubenswrapper[4846]: E0202 12:26:51.354396 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 12:26:51 crc kubenswrapper[4846]: E0202 12:26:51.354441 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:26:55.354427379 +0000 UTC m=+1046.583014242 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "webhook-server-cert" not found Feb 02 12:26:51 crc kubenswrapper[4846]: E0202 12:26:51.354761 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 12:26:51 crc kubenswrapper[4846]: E0202 12:26:51.354784 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:26:55.354777378 +0000 UTC m=+1046.583364241 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "metrics-server-cert" not found Feb 02 12:26:54 crc kubenswrapper[4846]: I0202 12:26:54.903137 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:26:54 crc kubenswrapper[4846]: E0202 12:26:54.903808 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 12:26:54 crc kubenswrapper[4846]: E0202 12:26:54.903853 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert podName:f75d8c59-636e-4496-bd81-e4e53f37a713 nodeName:}" failed. No retries permitted until 2026-02-02 12:27:02.903840036 +0000 UTC m=+1054.132426899 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert") pod "infra-operator-controller-manager-79955696d6-c6sk4" (UID: "f75d8c59-636e-4496-bd81-e4e53f37a713") : secret "infra-operator-webhook-server-cert" not found Feb 02 12:26:55 crc kubenswrapper[4846]: I0202 12:26:55.215009 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:26:55 crc kubenswrapper[4846]: E0202 12:26:55.215213 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:26:55 crc kubenswrapper[4846]: E0202 12:26:55.215300 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert podName:b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1 nodeName:}" failed. No retries permitted until 2026-02-02 12:27:03.215276603 +0000 UTC m=+1054.443863566 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" (UID: "b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:26:55 crc kubenswrapper[4846]: I0202 12:26:55.418096 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:26:55 crc kubenswrapper[4846]: E0202 12:26:55.418363 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 12:26:55 crc kubenswrapper[4846]: E0202 12:26:55.418414 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 12:26:55 crc kubenswrapper[4846]: E0202 12:26:55.418497 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:27:03.418465784 +0000 UTC m=+1054.647052647 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "metrics-server-cert" not found Feb 02 12:26:55 crc kubenswrapper[4846]: E0202 12:26:55.418520 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:27:03.418513035 +0000 UTC m=+1054.647099898 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "webhook-server-cert" not found Feb 02 12:26:55 crc kubenswrapper[4846]: I0202 12:26:55.418859 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:27:00 crc kubenswrapper[4846]: I0202 12:27:00.478586 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:27:00 crc kubenswrapper[4846]: I0202 12:27:00.478956 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.540868 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8" event={"ID":"c0f1cd1a-da1b-4292-8622-87119d40ec03","Type":"ContainerStarted","Data":"27f3b35acf5b397df94f66f4dd8f45f552d60a93c5ff70db16ba81cec3ae7ef1"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.541799 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.564982 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp" event={"ID":"8d9a7d64-a757-48b8-bdb9-4ec3e2921321","Type":"ContainerStarted","Data":"93bd75a6cb4cf12a308d9088f3c580ce0abfbf5fba73d02eeffbb170a72bc6e5"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.565661 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.583594 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n" event={"ID":"69f3343d-ca8a-4ff2-bc0c-14121a6938a8","Type":"ContainerStarted","Data":"d161d4632be166f90bbab34c36f8ab9c68343ca648bda8afc44736c6fa39a495"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.583841 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.601302 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj" event={"ID":"2e0bfdb6-f6e5-41c4-ab53-f13863d687b8","Type":"ContainerStarted","Data":"9bc2364aa8f114277d8aefb6fa73f250ec60433990f493dcd6070b50461bdde4"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.601929 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.606265 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8" podStartSLOduration=3.421954431 podStartE2EDuration="15.606250661s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:47.89689701 +0000 UTC m=+1039.125483873" lastFinishedPulling="2026-02-02 12:27:00.08119324 +0000 UTC m=+1051.309780103" observedRunningTime="2026-02-02 12:27:01.574027773 +0000 UTC m=+1052.802614636" watchObservedRunningTime="2026-02-02 12:27:01.606250661 +0000 UTC m=+1052.834837524" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.611192 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp" podStartSLOduration=3.195990402 podStartE2EDuration="14.611177811s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.716996678 +0000 UTC m=+1039.945583541" lastFinishedPulling="2026-02-02 12:27:00.132184087 +0000 UTC m=+1051.360770950" observedRunningTime="2026-02-02 12:27:01.600970252 +0000 UTC m=+1052.829557115" watchObservedRunningTime="2026-02-02 12:27:01.611177811 +0000 UTC m=+1052.839764674" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.623856 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk" event={"ID":"f8ebd280-c5fa-4cf3-890b-93a018bcb2d2","Type":"ContainerStarted","Data":"3e6ede663ba81b4df33d6b49179884f345498accf090e83262408dff1533a1a5"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.624459 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.646680 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz" event={"ID":"fced078c-1f83-492e-9f21-e9e82c1f9275","Type":"ContainerStarted","Data":"8742398f51d2da4d89202e1dbb19eca6116e8b2e13f4e2a124d48cef349c0972"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.647317 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.659175 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw" event={"ID":"b9c7230d-0e38-4540-9074-09a47500dd40","Type":"ContainerStarted","Data":"5e4f17fdb8a1b082457377bb3ffa891e48a8265515bb33a61df6a0e585c1c5b3"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.659791 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.662107 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n" podStartSLOduration=3.909327303 podStartE2EDuration="15.662096748s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.40635251 +0000 UTC m=+1039.634939373" lastFinishedPulling="2026-02-02 12:27:00.159121955 +0000 UTC m=+1051.387708818" observedRunningTime="2026-02-02 12:27:01.660005046 +0000 UTC m=+1052.888591909" watchObservedRunningTime="2026-02-02 12:27:01.662096748 +0000 UTC m=+1052.890683611" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.686781 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td" event={"ID":"6ba95849-98ff-4d50-8c41-605aa99c3b2f","Type":"ContainerStarted","Data":"07835cb5551f89cefbac8fa74e5fd39ff11b468a8e0ecac3b6860c69c3f13b80"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.687376 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.689250 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj" podStartSLOduration=4.44278622 podStartE2EDuration="15.689240251s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.842345485 +0000 UTC m=+1040.070932348" lastFinishedPulling="2026-02-02 12:27:00.088799516 +0000 UTC m=+1051.317386379" observedRunningTime="2026-02-02 12:27:01.679960165 +0000 UTC m=+1052.908547028" watchObservedRunningTime="2026-02-02 12:27:01.689240251 +0000 UTC m=+1052.917827114" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.703023 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6" event={"ID":"478eaaaa-5884-4b31-a87e-8655bedb96f1","Type":"ContainerStarted","Data":"3c31a0e48d677b34d17a2eca56f2dabdd5eefaa0b88d4a47f3b024a2b8086742"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.703576 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.704771 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz" podStartSLOduration=4.003972957 podStartE2EDuration="15.704760541s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.408568715 +0000 UTC m=+1039.637155578" lastFinishedPulling="2026-02-02 12:27:00.109356299 +0000 UTC m=+1051.337943162" observedRunningTime="2026-02-02 12:27:01.702797073 +0000 UTC m=+1052.931383936" watchObservedRunningTime="2026-02-02 12:27:01.704760541 +0000 UTC m=+1052.933347404" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.729494 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6" event={"ID":"a70f0094-0775-49ea-9002-9f3da3ff87d1","Type":"ContainerStarted","Data":"e4c9ed8ac0dc8fca1236489c8a06675292ff550ccd014232db4e282c0e7cd40c"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.730126 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.747450 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw" podStartSLOduration=4.01429424 podStartE2EDuration="15.747433985s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.399213356 +0000 UTC m=+1039.627800219" lastFinishedPulling="2026-02-02 12:27:00.132353101 +0000 UTC m=+1051.360939964" observedRunningTime="2026-02-02 12:27:01.744079262 +0000 UTC m=+1052.972666125" watchObservedRunningTime="2026-02-02 12:27:01.747433985 +0000 UTC m=+1052.976020838" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.750799 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg" event={"ID":"d6f0f90b-675e-40e9-b541-7ff04333a885","Type":"ContainerStarted","Data":"dec5a1aa87df8d7f2f39ea2e575c40846573e7d371afacf3a8095a662dfff085"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.751386 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.769458 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7" event={"ID":"f4ec43b3-d763-4478-8ef9-0453a2fb730a","Type":"ContainerStarted","Data":"47cf125696e08a1397bc3f5f76ca56fa88e927bb591c3689f9b68580729e6ccc"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.769504 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.804729 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b" event={"ID":"2497b300-a910-41c8-b70d-1c9acd871a90","Type":"ContainerStarted","Data":"353f9f54ee80e332eb327e194eba945cd65c13559a3129641abd8adfff873f69"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.804879 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.814801 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f" event={"ID":"46d6dfa1-87b3-4e53-bcfd-27c709ad7911","Type":"ContainerStarted","Data":"daf83e20e4dd2d22950bb52deb39625f4b2d464fc3e51133e89bc0ffaef82fed"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.815221 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.816526 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-564965969-8gkkw" event={"ID":"8ee9f8c5-8078-4ad1-b7c8-4a8e2dfd4e41","Type":"ContainerStarted","Data":"a0c596701a794bd6f3c676e21a83bbc59085edeec74fed086707775ef7976ed6"} Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.816779 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-564965969-8gkkw" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.829414 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk" podStartSLOduration=4.779168698 podStartE2EDuration="15.829391339s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:49.080643013 +0000 UTC m=+1040.309229876" lastFinishedPulling="2026-02-02 12:27:00.130865654 +0000 UTC m=+1051.359452517" observedRunningTime="2026-02-02 12:27:01.801054306 +0000 UTC m=+1053.029641169" watchObservedRunningTime="2026-02-02 12:27:01.829391339 +0000 UTC m=+1053.057978202" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.831537 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6" podStartSLOduration=4.345171243 podStartE2EDuration="15.831526902s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.622910457 +0000 UTC m=+1039.851497320" lastFinishedPulling="2026-02-02 12:27:00.109266116 +0000 UTC m=+1051.337852979" observedRunningTime="2026-02-02 12:27:01.830880405 +0000 UTC m=+1053.059467268" watchObservedRunningTime="2026-02-02 12:27:01.831526902 +0000 UTC m=+1053.060113765" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.894572 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg" podStartSLOduration=4.492782233 podStartE2EDuration="15.894549783s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.712608811 +0000 UTC m=+1039.941195674" lastFinishedPulling="2026-02-02 12:27:00.114376361 +0000 UTC m=+1051.342963224" observedRunningTime="2026-02-02 12:27:01.871259113 +0000 UTC m=+1053.099845966" watchObservedRunningTime="2026-02-02 12:27:01.894549783 +0000 UTC m=+1053.123136646" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.916408 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6" podStartSLOduration=4.216592698 podStartE2EDuration="15.916388587s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.431937497 +0000 UTC m=+1039.660524360" lastFinishedPulling="2026-02-02 12:27:00.131733386 +0000 UTC m=+1051.360320249" observedRunningTime="2026-02-02 12:27:01.902988579 +0000 UTC m=+1053.131575442" watchObservedRunningTime="2026-02-02 12:27:01.916388587 +0000 UTC m=+1053.144975450" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.942309 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td" podStartSLOduration=3.853331503 podStartE2EDuration="15.942288271s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.041116188 +0000 UTC m=+1039.269703051" lastFinishedPulling="2026-02-02 12:27:00.130072956 +0000 UTC m=+1051.358659819" observedRunningTime="2026-02-02 12:27:01.937892023 +0000 UTC m=+1053.166478886" watchObservedRunningTime="2026-02-02 12:27:01.942288271 +0000 UTC m=+1053.170875134" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.965778 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7" podStartSLOduration=3.900417764 podStartE2EDuration="14.965758915s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="2026-02-02 12:26:49.096171243 +0000 UTC m=+1040.324758096" lastFinishedPulling="2026-02-02 12:27:00.161512384 +0000 UTC m=+1051.390099247" observedRunningTime="2026-02-02 12:27:01.965102018 +0000 UTC m=+1053.193688891" watchObservedRunningTime="2026-02-02 12:27:01.965758915 +0000 UTC m=+1053.194345778" Feb 02 12:27:01 crc kubenswrapper[4846]: I0202 12:27:01.987123 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-564965969-8gkkw" podStartSLOduration=3.948782857 podStartE2EDuration="14.987103487s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="2026-02-02 12:26:49.093464627 +0000 UTC m=+1040.322051490" lastFinishedPulling="2026-02-02 12:27:00.131785257 +0000 UTC m=+1051.360372120" observedRunningTime="2026-02-02 12:27:01.980843714 +0000 UTC m=+1053.209430567" watchObservedRunningTime="2026-02-02 12:27:01.987103487 +0000 UTC m=+1053.215690350" Feb 02 12:27:02 crc kubenswrapper[4846]: I0202 12:27:02.007212 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b" podStartSLOduration=3.6178159 podStartE2EDuration="15.007150747s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.714085627 +0000 UTC m=+1039.942672490" lastFinishedPulling="2026-02-02 12:27:00.103420474 +0000 UTC m=+1051.332007337" observedRunningTime="2026-02-02 12:27:01.999861009 +0000 UTC m=+1053.228447892" watchObservedRunningTime="2026-02-02 12:27:02.007150747 +0000 UTC m=+1053.235737610" Feb 02 12:27:02 crc kubenswrapper[4846]: I0202 12:27:02.025789 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f" podStartSLOduration=4.515663584 podStartE2EDuration="16.025773273s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:26:48.613934378 +0000 UTC m=+1039.842521241" lastFinishedPulling="2026-02-02 12:27:00.124044067 +0000 UTC m=+1051.352630930" observedRunningTime="2026-02-02 12:27:02.023238971 +0000 UTC m=+1053.251825824" watchObservedRunningTime="2026-02-02 12:27:02.025773273 +0000 UTC m=+1053.254360136" Feb 02 12:27:02 crc kubenswrapper[4846]: I0202 12:27:02.931054 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:27:02 crc kubenswrapper[4846]: E0202 12:27:02.931345 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 02 12:27:02 crc kubenswrapper[4846]: E0202 12:27:02.931392 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert podName:f75d8c59-636e-4496-bd81-e4e53f37a713 nodeName:}" failed. No retries permitted until 2026-02-02 12:27:18.931375623 +0000 UTC m=+1070.159962616 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert") pod "infra-operator-controller-manager-79955696d6-c6sk4" (UID: "f75d8c59-636e-4496-bd81-e4e53f37a713") : secret "infra-operator-webhook-server-cert" not found Feb 02 12:27:03 crc kubenswrapper[4846]: I0202 12:27:03.234748 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:27:03 crc kubenswrapper[4846]: E0202 12:27:03.234963 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:27:03 crc kubenswrapper[4846]: E0202 12:27:03.235034 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert podName:b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1 nodeName:}" failed. No retries permitted until 2026-02-02 12:27:19.23501726 +0000 UTC m=+1070.463604123 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert") pod "openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" (UID: "b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 02 12:27:03 crc kubenswrapper[4846]: I0202 12:27:03.438619 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:27:03 crc kubenswrapper[4846]: I0202 12:27:03.438729 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:27:03 crc kubenswrapper[4846]: E0202 12:27:03.438892 4846 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 02 12:27:03 crc kubenswrapper[4846]: E0202 12:27:03.438909 4846 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 02 12:27:03 crc kubenswrapper[4846]: E0202 12:27:03.438958 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:27:19.438940828 +0000 UTC m=+1070.667527701 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "webhook-server-cert" not found Feb 02 12:27:03 crc kubenswrapper[4846]: E0202 12:27:03.438979 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs podName:b08a9394-cfb5-4249-8f6c-a3fbe49bbeea nodeName:}" failed. No retries permitted until 2026-02-02 12:27:19.438969088 +0000 UTC m=+1070.667555951 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs") pod "openstack-operator-controller-manager-646f757d77-rlzgs" (UID: "b08a9394-cfb5-4249-8f6c-a3fbe49bbeea") : secret "metrics-server-cert" not found Feb 02 12:27:06 crc kubenswrapper[4846]: I0202 12:27:06.872158 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" event={"ID":"b45d49d2-cad0-49fa-addf-766dba1544a4","Type":"ContainerStarted","Data":"961782ee553f7f745cbf2c3908bfbe06565aa4b789f295933275a5b44f8fcc25"} Feb 02 12:27:06 crc kubenswrapper[4846]: I0202 12:27:06.895701 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9tnhp" podStartSLOduration=2.491666766 podStartE2EDuration="19.895669487s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="2026-02-02 12:26:49.147073108 +0000 UTC m=+1040.375659971" lastFinishedPulling="2026-02-02 12:27:06.551075829 +0000 UTC m=+1057.779662692" observedRunningTime="2026-02-02 12:27:06.893213777 +0000 UTC m=+1058.121800650" watchObservedRunningTime="2026-02-02 12:27:06.895669487 +0000 UTC m=+1058.124256350" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.159808 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b6c4d8c5f-8lrz8" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.202529 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d9697b7f4-88lrz" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.207173 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d874c8fc-hc6td" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.230964 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8886f4c47-bklmw" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.266069 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69d6db494d-jscz6" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.291144 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-cz76f" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.375762 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5f4b8bd54d-7lr2n" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.444358 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-84f48565d4-hfvt6" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.444409 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7dd968899f-5vzvg" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.466170 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-nvnpj" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.594104 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-585dbc889-hrnrk" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.628706 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6687f8d877-vfmzp" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.647163 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-55bff696bd-89dd7" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.689422 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-tw44b" Feb 02 12:27:07 crc kubenswrapper[4846]: I0202 12:27:07.915866 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-564965969-8gkkw" Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.903589 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" event={"ID":"204fa43c-5360-4282-b002-ae22e0f0595d","Type":"ContainerStarted","Data":"9f8970d436fefbc3b867b1acbab648e5afdc2cbe497ee876be301457a45ac02b"} Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.904099 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.906050 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" event={"ID":"9c5ac153-1fd9-491b-942d-2b0b3e3e9da5","Type":"ContainerStarted","Data":"6b22c40f86ce05b6b478e6ac20ce7e5ec77eef7444347d9d0c99dc69043b67a6"} Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.906223 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.919130 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" event={"ID":"cb8f2427-7089-4ecf-8ae9-4ece48067de8","Type":"ContainerStarted","Data":"7e707f3a60b0593c169033be5ce9145bc1afe346820f1d2146ae64142b15a2b2"} Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.920055 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.923736 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" event={"ID":"124b64f2-5773-4d35-ab14-f4ad0fe4f318","Type":"ContainerStarted","Data":"46a541afa4275d16e13f3bb906666eb9d41d00c87e8bec067794ee2b26210af4"} Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.924574 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.948545 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" podStartSLOduration=2.784405516 podStartE2EDuration="23.948530916s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="2026-02-02 12:26:49.115349082 +0000 UTC m=+1040.343935945" lastFinishedPulling="2026-02-02 12:27:10.279474482 +0000 UTC m=+1061.508061345" observedRunningTime="2026-02-02 12:27:10.946128808 +0000 UTC m=+1062.174715671" watchObservedRunningTime="2026-02-02 12:27:10.948530916 +0000 UTC m=+1062.177117779" Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.973398 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" podStartSLOduration=2.818797277 podStartE2EDuration="23.973382654s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="2026-02-02 12:26:49.129667842 +0000 UTC m=+1040.358254705" lastFinishedPulling="2026-02-02 12:27:10.284253229 +0000 UTC m=+1061.512840082" observedRunningTime="2026-02-02 12:27:10.971063327 +0000 UTC m=+1062.199650190" watchObservedRunningTime="2026-02-02 12:27:10.973382654 +0000 UTC m=+1062.201969517" Feb 02 12:27:10 crc kubenswrapper[4846]: I0202 12:27:10.986853 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" podStartSLOduration=2.806447167 podStartE2EDuration="23.986836344s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="2026-02-02 12:26:49.115543307 +0000 UTC m=+1040.344130160" lastFinishedPulling="2026-02-02 12:27:10.295932474 +0000 UTC m=+1061.524519337" observedRunningTime="2026-02-02 12:27:10.985975832 +0000 UTC m=+1062.214562695" watchObservedRunningTime="2026-02-02 12:27:10.986836344 +0000 UTC m=+1062.215423207" Feb 02 12:27:11 crc kubenswrapper[4846]: I0202 12:27:11.008264 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" podStartSLOduration=2.831357745 podStartE2EDuration="24.008239767s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="2026-02-02 12:26:49.118397497 +0000 UTC m=+1040.346984360" lastFinishedPulling="2026-02-02 12:27:10.295279529 +0000 UTC m=+1061.523866382" observedRunningTime="2026-02-02 12:27:10.998597471 +0000 UTC m=+1062.227184344" watchObservedRunningTime="2026-02-02 12:27:11.008239767 +0000 UTC m=+1062.236826630" Feb 02 12:27:17 crc kubenswrapper[4846]: I0202 12:27:17.782568 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-pqnqm" Feb 02 12:27:17 crc kubenswrapper[4846]: I0202 12:27:17.824560 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" Feb 02 12:27:17 crc kubenswrapper[4846]: I0202 12:27:17.870478 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" Feb 02 12:27:17 crc kubenswrapper[4846]: I0202 12:27:17.890707 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-6g958" Feb 02 12:27:18 crc kubenswrapper[4846]: I0202 12:27:18.987637 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:27:18 crc kubenswrapper[4846]: I0202 12:27:18.993931 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f75d8c59-636e-4496-bd81-e4e53f37a713-cert\") pod \"infra-operator-controller-manager-79955696d6-c6sk4\" (UID: \"f75d8c59-636e-4496-bd81-e4e53f37a713\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.141710 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.291559 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.309375 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1-cert\") pod \"openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw\" (UID: \"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.497904 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.498001 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.502047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-metrics-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.502965 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b08a9394-cfb5-4249-8f6c-a3fbe49bbeea-webhook-certs\") pod \"openstack-operator-controller-manager-646f757d77-rlzgs\" (UID: \"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea\") " pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.506464 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.578134 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4"] Feb 02 12:27:19 crc kubenswrapper[4846]: W0202 12:27:19.586800 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf75d8c59_636e_4496_bd81_e4e53f37a713.slice/crio-43687cd67eb8f2d278bb73801ff4beb578974eb8668d44939085e07f7f96b358 WatchSource:0}: Error finding container 43687cd67eb8f2d278bb73801ff4beb578974eb8668d44939085e07f7f96b358: Status 404 returned error can't find the container with id 43687cd67eb8f2d278bb73801ff4beb578974eb8668d44939085e07f7f96b358 Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.712824 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw"] Feb 02 12:27:19 crc kubenswrapper[4846]: W0202 12:27:19.718921 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb75ce3f5_f42f_4c8f_aa7f_6e8bc0a08ee1.slice/crio-98032ea7d7e8f5be4c8669386b75437e65fec4381bfc35ceccbdfab0341f5398 WatchSource:0}: Error finding container 98032ea7d7e8f5be4c8669386b75437e65fec4381bfc35ceccbdfab0341f5398: Status 404 returned error can't find the container with id 98032ea7d7e8f5be4c8669386b75437e65fec4381bfc35ceccbdfab0341f5398 Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.774838 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.988497 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" event={"ID":"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1","Type":"ContainerStarted","Data":"98032ea7d7e8f5be4c8669386b75437e65fec4381bfc35ceccbdfab0341f5398"} Feb 02 12:27:19 crc kubenswrapper[4846]: I0202 12:27:19.989751 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" event={"ID":"f75d8c59-636e-4496-bd81-e4e53f37a713","Type":"ContainerStarted","Data":"43687cd67eb8f2d278bb73801ff4beb578974eb8668d44939085e07f7f96b358"} Feb 02 12:27:20 crc kubenswrapper[4846]: I0202 12:27:20.196754 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs"] Feb 02 12:27:20 crc kubenswrapper[4846]: I0202 12:27:20.998358 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" event={"ID":"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea","Type":"ContainerStarted","Data":"6731dab3a0bc08481dfd1839b1654b21475e63b3637134824afddb1ed357089a"} Feb 02 12:27:20 crc kubenswrapper[4846]: I0202 12:27:20.998719 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" event={"ID":"b08a9394-cfb5-4249-8f6c-a3fbe49bbeea","Type":"ContainerStarted","Data":"e5efca8072649b795c859ecd9f1eb27ae9effdbf47444c9eca4c147d08d4c761"} Feb 02 12:27:20 crc kubenswrapper[4846]: I0202 12:27:20.999078 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:27:21 crc kubenswrapper[4846]: I0202 12:27:21.023688 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" podStartSLOduration=34.023669937 podStartE2EDuration="34.023669937s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:27:21.021794712 +0000 UTC m=+1072.250381585" watchObservedRunningTime="2026-02-02 12:27:21.023669937 +0000 UTC m=+1072.252256800" Feb 02 12:27:23 crc kubenswrapper[4846]: I0202 12:27:23.016150 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" event={"ID":"b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1","Type":"ContainerStarted","Data":"a3ff2760729b3a446e95e04ad9ef7ff053099a3795539c81a3fc487f519e1ec6"} Feb 02 12:27:23 crc kubenswrapper[4846]: I0202 12:27:23.016614 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:27:23 crc kubenswrapper[4846]: I0202 12:27:23.018111 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" event={"ID":"f75d8c59-636e-4496-bd81-e4e53f37a713","Type":"ContainerStarted","Data":"521ba6df00f306a7669fb1309ceb2e6af18403b6c301f01472a672ce32cf4878"} Feb 02 12:27:23 crc kubenswrapper[4846]: I0202 12:27:23.018895 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:27:23 crc kubenswrapper[4846]: I0202 12:27:23.048401 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" podStartSLOduration=33.811841407 podStartE2EDuration="36.04838018s" podCreationTimestamp="2026-02-02 12:26:47 +0000 UTC" firstStartedPulling="2026-02-02 12:27:19.723670391 +0000 UTC m=+1070.952257254" lastFinishedPulling="2026-02-02 12:27:21.960209154 +0000 UTC m=+1073.188796027" observedRunningTime="2026-02-02 12:27:23.039260678 +0000 UTC m=+1074.267847551" watchObservedRunningTime="2026-02-02 12:27:23.04838018 +0000 UTC m=+1074.276967043" Feb 02 12:27:23 crc kubenswrapper[4846]: I0202 12:27:23.061281 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" podStartSLOduration=34.69179809 podStartE2EDuration="37.061256755s" podCreationTimestamp="2026-02-02 12:26:46 +0000 UTC" firstStartedPulling="2026-02-02 12:27:19.589046828 +0000 UTC m=+1070.817633691" lastFinishedPulling="2026-02-02 12:27:21.958505493 +0000 UTC m=+1073.187092356" observedRunningTime="2026-02-02 12:27:23.055559196 +0000 UTC m=+1074.284146079" watchObservedRunningTime="2026-02-02 12:27:23.061256755 +0000 UTC m=+1074.289843628" Feb 02 12:27:29 crc kubenswrapper[4846]: I0202 12:27:29.148346 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79955696d6-c6sk4" Feb 02 12:27:29 crc kubenswrapper[4846]: I0202 12:27:29.513394 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw" Feb 02 12:27:29 crc kubenswrapper[4846]: I0202 12:27:29.782428 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-646f757d77-rlzgs" Feb 02 12:27:30 crc kubenswrapper[4846]: I0202 12:27:30.479305 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:27:30 crc kubenswrapper[4846]: I0202 12:27:30.479654 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.091314 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-f64l5"] Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.097533 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.100612 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-qr27m" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.100888 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.101070 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.101285 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.104144 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-f64l5"] Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.164707 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-z889d"] Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.169434 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8lkh\" (UniqueName: \"kubernetes.io/projected/02e070d5-c4c6-4464-90ce-faeddb3d8d47-kube-api-access-x8lkh\") pod \"dnsmasq-dns-855cbc58c5-f64l5\" (UID: \"02e070d5-c4c6-4464-90ce-faeddb3d8d47\") " pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.169487 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e070d5-c4c6-4464-90ce-faeddb3d8d47-config\") pod \"dnsmasq-dns-855cbc58c5-f64l5\" (UID: \"02e070d5-c4c6-4464-90ce-faeddb3d8d47\") " pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.169971 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.171756 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.183422 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-z889d"] Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.271238 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-config\") pod \"dnsmasq-dns-6fcf94d689-z889d\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.271316 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhpnf\" (UniqueName: \"kubernetes.io/projected/1003a846-f9a3-4562-ad4e-a449e7213dcd-kube-api-access-vhpnf\") pod \"dnsmasq-dns-6fcf94d689-z889d\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.271346 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8lkh\" (UniqueName: \"kubernetes.io/projected/02e070d5-c4c6-4464-90ce-faeddb3d8d47-kube-api-access-x8lkh\") pod \"dnsmasq-dns-855cbc58c5-f64l5\" (UID: \"02e070d5-c4c6-4464-90ce-faeddb3d8d47\") " pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.271430 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e070d5-c4c6-4464-90ce-faeddb3d8d47-config\") pod \"dnsmasq-dns-855cbc58c5-f64l5\" (UID: \"02e070d5-c4c6-4464-90ce-faeddb3d8d47\") " pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.271571 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-dns-svc\") pod \"dnsmasq-dns-6fcf94d689-z889d\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.272359 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e070d5-c4c6-4464-90ce-faeddb3d8d47-config\") pod \"dnsmasq-dns-855cbc58c5-f64l5\" (UID: \"02e070d5-c4c6-4464-90ce-faeddb3d8d47\") " pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.292300 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8lkh\" (UniqueName: \"kubernetes.io/projected/02e070d5-c4c6-4464-90ce-faeddb3d8d47-kube-api-access-x8lkh\") pod \"dnsmasq-dns-855cbc58c5-f64l5\" (UID: \"02e070d5-c4c6-4464-90ce-faeddb3d8d47\") " pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.372964 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhpnf\" (UniqueName: \"kubernetes.io/projected/1003a846-f9a3-4562-ad4e-a449e7213dcd-kube-api-access-vhpnf\") pod \"dnsmasq-dns-6fcf94d689-z889d\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.373036 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-dns-svc\") pod \"dnsmasq-dns-6fcf94d689-z889d\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.373080 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-config\") pod \"dnsmasq-dns-6fcf94d689-z889d\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.374031 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-config\") pod \"dnsmasq-dns-6fcf94d689-z889d\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.374053 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-dns-svc\") pod \"dnsmasq-dns-6fcf94d689-z889d\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.421711 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhpnf\" (UniqueName: \"kubernetes.io/projected/1003a846-f9a3-4562-ad4e-a449e7213dcd-kube-api-access-vhpnf\") pod \"dnsmasq-dns-6fcf94d689-z889d\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.421914 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.489211 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:27:46 crc kubenswrapper[4846]: I0202 12:27:46.936216 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-f64l5"] Feb 02 12:27:47 crc kubenswrapper[4846]: I0202 12:27:47.001833 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-z889d"] Feb 02 12:27:47 crc kubenswrapper[4846]: W0202 12:27:47.002138 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1003a846_f9a3_4562_ad4e_a449e7213dcd.slice/crio-3a008049a5de0fe7c8325a6b979a4b5453c1d151d5db5e4e81fda5a4ce90086b WatchSource:0}: Error finding container 3a008049a5de0fe7c8325a6b979a4b5453c1d151d5db5e4e81fda5a4ce90086b: Status 404 returned error can't find the container with id 3a008049a5de0fe7c8325a6b979a4b5453c1d151d5db5e4e81fda5a4ce90086b Feb 02 12:27:47 crc kubenswrapper[4846]: I0202 12:27:47.192936 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcf94d689-z889d" event={"ID":"1003a846-f9a3-4562-ad4e-a449e7213dcd","Type":"ContainerStarted","Data":"3a008049a5de0fe7c8325a6b979a4b5453c1d151d5db5e4e81fda5a4ce90086b"} Feb 02 12:27:47 crc kubenswrapper[4846]: I0202 12:27:47.193892 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" event={"ID":"02e070d5-c4c6-4464-90ce-faeddb3d8d47","Type":"ContainerStarted","Data":"604f2e4280ffa7168a6361f2e6388ddafe92ed88faa57d12f97c4415791cfebe"} Feb 02 12:27:48 crc kubenswrapper[4846]: I0202 12:27:48.782857 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-z889d"] Feb 02 12:27:48 crc kubenswrapper[4846]: I0202 12:27:48.827185 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-xthhp"] Feb 02 12:27:48 crc kubenswrapper[4846]: I0202 12:27:48.828471 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:48 crc kubenswrapper[4846]: I0202 12:27:48.899316 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-xthhp"] Feb 02 12:27:48 crc kubenswrapper[4846]: I0202 12:27:48.914448 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-config\") pod \"dnsmasq-dns-f54874ffc-xthhp\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:48 crc kubenswrapper[4846]: I0202 12:27:48.914538 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svzjm\" (UniqueName: \"kubernetes.io/projected/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-kube-api-access-svzjm\") pod \"dnsmasq-dns-f54874ffc-xthhp\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:48 crc kubenswrapper[4846]: I0202 12:27:48.914608 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-dns-svc\") pod \"dnsmasq-dns-f54874ffc-xthhp\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.017482 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svzjm\" (UniqueName: \"kubernetes.io/projected/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-kube-api-access-svzjm\") pod \"dnsmasq-dns-f54874ffc-xthhp\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.017580 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-dns-svc\") pod \"dnsmasq-dns-f54874ffc-xthhp\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.017716 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-config\") pod \"dnsmasq-dns-f54874ffc-xthhp\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.018662 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-config\") pod \"dnsmasq-dns-f54874ffc-xthhp\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.018829 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-dns-svc\") pod \"dnsmasq-dns-f54874ffc-xthhp\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.052600 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svzjm\" (UniqueName: \"kubernetes.io/projected/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-kube-api-access-svzjm\") pod \"dnsmasq-dns-f54874ffc-xthhp\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.158878 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.250931 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-f64l5"] Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.283854 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-cfxzp"] Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.288636 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.299134 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-cfxzp"] Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.424451 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpd2g\" (UniqueName: \"kubernetes.io/projected/24a64c0f-4d6c-4f6a-a2c4-659aed878323-kube-api-access-xpd2g\") pod \"dnsmasq-dns-67ff45466c-cfxzp\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.425022 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-config\") pod \"dnsmasq-dns-67ff45466c-cfxzp\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.425085 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-dns-svc\") pod \"dnsmasq-dns-67ff45466c-cfxzp\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.526348 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpd2g\" (UniqueName: \"kubernetes.io/projected/24a64c0f-4d6c-4f6a-a2c4-659aed878323-kube-api-access-xpd2g\") pod \"dnsmasq-dns-67ff45466c-cfxzp\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.526429 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-config\") pod \"dnsmasq-dns-67ff45466c-cfxzp\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.526456 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-dns-svc\") pod \"dnsmasq-dns-67ff45466c-cfxzp\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.528032 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-config\") pod \"dnsmasq-dns-67ff45466c-cfxzp\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.528628 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-dns-svc\") pod \"dnsmasq-dns-67ff45466c-cfxzp\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.546546 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpd2g\" (UniqueName: \"kubernetes.io/projected/24a64c0f-4d6c-4f6a-a2c4-659aed878323-kube-api-access-xpd2g\") pod \"dnsmasq-dns-67ff45466c-cfxzp\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.608421 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.766334 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-xthhp"] Feb 02 12:27:49 crc kubenswrapper[4846]: W0202 12:27:49.782367 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7eee9d79_99e5_4179_a3e1_8a9b9ef8c793.slice/crio-54065d6a56608ac45234fd9b53cf1c5d13a53b7871399058d17629738d730691 WatchSource:0}: Error finding container 54065d6a56608ac45234fd9b53cf1c5d13a53b7871399058d17629738d730691: Status 404 returned error can't find the container with id 54065d6a56608ac45234fd9b53cf1c5d13a53b7871399058d17629738d730691 Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.803107 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 12:27:49 crc kubenswrapper[4846]: I0202 12:27:49.866946 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-cfxzp"] Feb 02 12:27:49 crc kubenswrapper[4846]: W0202 12:27:49.875715 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24a64c0f_4d6c_4f6a_a2c4_659aed878323.slice/crio-441ad332a730307b2b51e7a8416454bce188ea84830313e2c9053c65ee153324 WatchSource:0}: Error finding container 441ad332a730307b2b51e7a8416454bce188ea84830313e2c9053c65ee153324: Status 404 returned error can't find the container with id 441ad332a730307b2b51e7a8416454bce188ea84830313e2c9053c65ee153324 Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.050798 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.052291 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.054973 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.055330 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.055707 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zn2n5" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.055901 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.056093 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.056464 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.057156 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.072669 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136076 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjtpb\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-kube-api-access-tjtpb\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136124 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9cec353a-92f9-4ed4-8ec9-15271cbc6912-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136146 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136166 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136200 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136232 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9cec353a-92f9-4ed4-8ec9-15271cbc6912-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136278 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136311 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136347 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.136404 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.251519 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9cec353a-92f9-4ed4-8ec9-15271cbc6912-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.251564 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.251581 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.251608 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.252373 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9cec353a-92f9-4ed4-8ec9-15271cbc6912-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.252422 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.252449 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.252477 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.252503 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.252518 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.252537 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjtpb\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-kube-api-access-tjtpb\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.254671 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.255536 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.256082 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.256398 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.257363 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.257461 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9cec353a-92f9-4ed4-8ec9-15271cbc6912-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.259259 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.262878 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.263496 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.276122 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" event={"ID":"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793","Type":"ContainerStarted","Data":"54065d6a56608ac45234fd9b53cf1c5d13a53b7871399058d17629738d730691"} Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.277613 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" event={"ID":"24a64c0f-4d6c-4f6a-a2c4-659aed878323","Type":"ContainerStarted","Data":"441ad332a730307b2b51e7a8416454bce188ea84830313e2c9053c65ee153324"} Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.281316 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjtpb\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-kube-api-access-tjtpb\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.290034 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9cec353a-92f9-4ed4-8ec9-15271cbc6912-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.305136 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.390246 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.418913 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.420182 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.427521 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.427549 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-pxfwx" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.427765 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.427941 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.427982 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.428150 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.428439 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.462226 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556227 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/327c8143-b3ed-4431-a206-d2019ddcb606-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556265 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556300 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556362 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5l9t\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-kube-api-access-f5l9t\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556385 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556401 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556420 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556440 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556462 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556484 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.556513 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/327c8143-b3ed-4431-a206-d2019ddcb606-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.657813 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.658195 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.658235 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.658267 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.658310 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/327c8143-b3ed-4431-a206-d2019ddcb606-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.658346 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/327c8143-b3ed-4431-a206-d2019ddcb606-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.658367 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.658399 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.658457 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5l9t\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-kube-api-access-f5l9t\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.658934 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.658962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.660222 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.660536 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.661890 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.662590 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.663304 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.663765 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.665611 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/327c8143-b3ed-4431-a206-d2019ddcb606-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.670999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/327c8143-b3ed-4431-a206-d2019ddcb606-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.676188 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.679237 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.679415 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5l9t\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-kube-api-access-f5l9t\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.697783 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.730232 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 12:27:50 crc kubenswrapper[4846]: I0202 12:27:50.799289 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.295810 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9cec353a-92f9-4ed4-8ec9-15271cbc6912","Type":"ContainerStarted","Data":"4d3faa4533f7f892e5650a5fff8a4ddfbb90fa8747bd72cc9430ed290fd0d96a"} Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.598173 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.600677 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.605201 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-q4cqv" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.605494 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.605751 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.612061 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.626730 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.632807 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.683872 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.683919 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.683964 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.683982 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk6x9\" (UniqueName: \"kubernetes.io/projected/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kube-api-access-fk6x9\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.684002 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.684027 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.684069 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.684106 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.785931 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.786072 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.786105 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.786179 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.786203 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk6x9\" (UniqueName: \"kubernetes.io/projected/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kube-api-access-fk6x9\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.786233 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.786269 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.786323 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.786541 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.786875 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.787035 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-default\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.787462 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kolla-config\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.788274 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.798893 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.809358 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.852447 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.878332 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk6x9\" (UniqueName: \"kubernetes.io/projected/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kube-api-access-fk6x9\") pod \"openstack-galera-0\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " pod="openstack/openstack-galera-0" Feb 02 12:27:51 crc kubenswrapper[4846]: I0202 12:27:51.931223 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.079956 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.083287 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.087348 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.087435 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.088216 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.110312 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.111436 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-cxcsf" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.213297 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.213386 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.213453 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.213471 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.213491 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.213514 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.213534 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6m4b\" (UniqueName: \"kubernetes.io/projected/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kube-api-access-x6m4b\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.213550 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.298046 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.298935 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.301493 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-25cs7" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.301590 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.304561 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.311045 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.314856 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.314894 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.314954 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.314978 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.314999 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6m4b\" (UniqueName: \"kubernetes.io/projected/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kube-api-access-x6m4b\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.315015 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.315062 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.315118 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.322243 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.322755 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.331536 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.332052 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.332149 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.349877 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.356302 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.361516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6m4b\" (UniqueName: \"kubernetes.io/projected/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kube-api-access-x6m4b\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.401017 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.417804 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kolla-config\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.417857 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-config-data\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.417891 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.417925 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.417967 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7vcm\" (UniqueName: \"kubernetes.io/projected/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kube-api-access-h7vcm\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.423741 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.519787 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.519855 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7vcm\" (UniqueName: \"kubernetes.io/projected/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kube-api-access-h7vcm\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.519916 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kolla-config\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.519969 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-config-data\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.520007 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.521261 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kolla-config\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.521587 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-config-data\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.543309 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.543327 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.546540 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7vcm\" (UniqueName: \"kubernetes.io/projected/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kube-api-access-h7vcm\") pod \"memcached-0\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " pod="openstack/memcached-0" Feb 02 12:27:53 crc kubenswrapper[4846]: I0202 12:27:53.614517 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 02 12:27:54 crc kubenswrapper[4846]: I0202 12:27:54.935551 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:27:54 crc kubenswrapper[4846]: I0202 12:27:54.938890 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 12:27:54 crc kubenswrapper[4846]: I0202 12:27:54.950188 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-fqmdw" Feb 02 12:27:54 crc kubenswrapper[4846]: I0202 12:27:54.963238 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:27:55 crc kubenswrapper[4846]: I0202 12:27:55.040893 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4mn2\" (UniqueName: \"kubernetes.io/projected/8c0a429c-a50e-493c-958e-61d8a3a404c6-kube-api-access-q4mn2\") pod \"kube-state-metrics-0\" (UID: \"8c0a429c-a50e-493c-958e-61d8a3a404c6\") " pod="openstack/kube-state-metrics-0" Feb 02 12:27:55 crc kubenswrapper[4846]: I0202 12:27:55.142601 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4mn2\" (UniqueName: \"kubernetes.io/projected/8c0a429c-a50e-493c-958e-61d8a3a404c6-kube-api-access-q4mn2\") pod \"kube-state-metrics-0\" (UID: \"8c0a429c-a50e-493c-958e-61d8a3a404c6\") " pod="openstack/kube-state-metrics-0" Feb 02 12:27:55 crc kubenswrapper[4846]: I0202 12:27:55.164969 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4mn2\" (UniqueName: \"kubernetes.io/projected/8c0a429c-a50e-493c-958e-61d8a3a404c6-kube-api-access-q4mn2\") pod \"kube-state-metrics-0\" (UID: \"8c0a429c-a50e-493c-958e-61d8a3a404c6\") " pod="openstack/kube-state-metrics-0" Feb 02 12:27:55 crc kubenswrapper[4846]: I0202 12:27:55.268380 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.286268 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.444272 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-b2hcc"] Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.445080 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.448088 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-hrgrp" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.449710 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.449945 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.466166 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b2hcc"] Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.473104 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-m79vg"] Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.475022 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.520542 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-m79vg"] Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607655 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-combined-ca-bundle\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607728 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-etc-ovs\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607751 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run-ovn\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607776 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7cs\" (UniqueName: \"kubernetes.io/projected/bda23288-2fee-45ad-81e2-a14654c16589-kube-api-access-md7cs\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607815 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bda23288-2fee-45ad-81e2-a14654c16589-scripts\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607835 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-log\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607868 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-lib\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607890 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-log-ovn\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607928 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-ovn-controller-tls-certs\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607945 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607961 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53dd5f18-d111-4359-8170-102a8b30d0d9-scripts\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607977 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb7p8\" (UniqueName: \"kubernetes.io/projected/53dd5f18-d111-4359-8170-102a8b30d0d9-kube-api-access-lb7p8\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.607997 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-run\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.623970 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.625349 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.627909 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.628244 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.628459 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.628552 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.628701 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-j9mgl" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.647729 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.709751 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md7cs\" (UniqueName: \"kubernetes.io/projected/bda23288-2fee-45ad-81e2-a14654c16589-kube-api-access-md7cs\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.709816 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqw5k\" (UniqueName: \"kubernetes.io/projected/36061d30-6265-47ca-8a87-5275f41d03ed-kube-api-access-dqw5k\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710035 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710061 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bda23288-2fee-45ad-81e2-a14654c16589-scripts\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710112 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710228 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-config\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710261 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710297 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710330 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-combined-ca-bundle\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710360 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-etc-ovs\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710381 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run-ovn\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710441 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-log\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710464 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710543 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-lib\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710575 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-log-ovn\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710617 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-ovn-controller-tls-certs\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710659 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53dd5f18-d111-4359-8170-102a8b30d0d9-scripts\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710688 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb7p8\" (UniqueName: \"kubernetes.io/projected/53dd5f18-d111-4359-8170-102a8b30d0d9-kube-api-access-lb7p8\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710713 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710741 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-run\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710778 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.710942 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.711136 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-log-ovn\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.711293 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-etc-ovs\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.711702 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-lib\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.711702 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-log\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.711776 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-run\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.712184 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run-ovn\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.713583 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53dd5f18-d111-4359-8170-102a8b30d0d9-scripts\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.714928 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bda23288-2fee-45ad-81e2-a14654c16589-scripts\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.721326 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-ovn-controller-tls-certs\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.729970 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb7p8\" (UniqueName: \"kubernetes.io/projected/53dd5f18-d111-4359-8170-102a8b30d0d9-kube-api-access-lb7p8\") pod \"ovn-controller-ovs-m79vg\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.731031 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-combined-ca-bundle\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.733637 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7cs\" (UniqueName: \"kubernetes.io/projected/bda23288-2fee-45ad-81e2-a14654c16589-kube-api-access-md7cs\") pod \"ovn-controller-b2hcc\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.767315 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.811822 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqw5k\" (UniqueName: \"kubernetes.io/projected/36061d30-6265-47ca-8a87-5275f41d03ed-kube-api-access-dqw5k\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.811862 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.811902 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-config\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.811923 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.811956 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.811999 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.812041 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.812068 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.812542 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.812827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-config\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.812993 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.813219 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.817436 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.818030 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.818190 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.821180 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.831170 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqw5k\" (UniqueName: \"kubernetes.io/projected/36061d30-6265-47ca-8a87-5275f41d03ed-kube-api-access-dqw5k\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.832803 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " pod="openstack/ovsdbserver-nb-0" Feb 02 12:27:59 crc kubenswrapper[4846]: I0202 12:27:59.948821 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 02 12:28:00 crc kubenswrapper[4846]: I0202 12:28:00.479597 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:28:00 crc kubenswrapper[4846]: I0202 12:28:00.479731 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:28:00 crc kubenswrapper[4846]: I0202 12:28:00.479786 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:28:00 crc kubenswrapper[4846]: I0202 12:28:00.480430 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4f60d595c3f3559b0b02374c951ace4cd6f20670671fc67c4b4653981928977"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:28:00 crc kubenswrapper[4846]: I0202 12:28:00.480485 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://d4f60d595c3f3559b0b02374c951ace4cd6f20670671fc67c4b4653981928977" gracePeriod=600 Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.431431 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="d4f60d595c3f3559b0b02374c951ace4cd6f20670671fc67c4b4653981928977" exitCode=0 Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.431743 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"d4f60d595c3f3559b0b02374c951ace4cd6f20670671fc67c4b4653981928977"} Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.431770 4846 scope.go:117] "RemoveContainer" containerID="06b5f4ddd20a5b1b0ef8c848baafd57c963bd52aab6eccb2c47d76fc10efc10c" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.726583 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.728047 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.732219 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.732562 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-2fdzk" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.732694 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.732968 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.740114 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.845929 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.846070 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.846124 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.846308 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.846380 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n9nc\" (UniqueName: \"kubernetes.io/projected/797602a8-7d50-4ab4-8ba5-4afff6a34409-kube-api-access-5n9nc\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.846413 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.846586 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-config\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.846791 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.948417 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n9nc\" (UniqueName: \"kubernetes.io/projected/797602a8-7d50-4ab4-8ba5-4afff6a34409-kube-api-access-5n9nc\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.948482 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.948533 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-config\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.948609 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.948666 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.948722 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.948763 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.948798 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.949053 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.949354 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.949643 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-config\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.950065 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.953770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.954340 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.959185 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.967890 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n9nc\" (UniqueName: \"kubernetes.io/projected/797602a8-7d50-4ab4-8ba5-4afff6a34409-kube-api-access-5n9nc\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:01 crc kubenswrapper[4846]: I0202 12:28:01.972215 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:02 crc kubenswrapper[4846]: I0202 12:28:02.092453 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:04 crc kubenswrapper[4846]: W0202 12:28:04.096028 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod327c8143_b3ed_4431_a206_d2019ddcb606.slice/crio-0acf417de61fa10c68d806b0e4b27aef887c00dc4b801daf9610cf72400b0adf WatchSource:0}: Error finding container 0acf417de61fa10c68d806b0e4b27aef887c00dc4b801daf9610cf72400b0adf: Status 404 returned error can't find the container with id 0acf417de61fa10c68d806b0e4b27aef887c00dc4b801daf9610cf72400b0adf Feb 02 12:28:04 crc kubenswrapper[4846]: I0202 12:28:04.469861 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"327c8143-b3ed-4431-a206-d2019ddcb606","Type":"ContainerStarted","Data":"0acf417de61fa10c68d806b0e4b27aef887c00dc4b801daf9610cf72400b0adf"} Feb 02 12:28:04 crc kubenswrapper[4846]: E0202 12:28:04.925083 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2" Feb 02 12:28:04 crc kubenswrapper[4846]: E0202 12:28:04.925237 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vhpnf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6fcf94d689-z889d_openstack(1003a846-f9a3-4562-ad4e-a449e7213dcd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 12:28:04 crc kubenswrapper[4846]: E0202 12:28:04.926615 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6fcf94d689-z889d" podUID="1003a846-f9a3-4562-ad4e-a449e7213dcd" Feb 02 12:28:04 crc kubenswrapper[4846]: E0202 12:28:04.936112 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2" Feb 02 12:28:04 crc kubenswrapper[4846]: E0202 12:28:04.936286 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x8lkh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-855cbc58c5-f64l5_openstack(02e070d5-c4c6-4464-90ce-faeddb3d8d47): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 12:28:04 crc kubenswrapper[4846]: E0202 12:28:04.938894 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" podUID="02e070d5-c4c6-4464-90ce-faeddb3d8d47" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.483678 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.513943 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcf94d689-z889d" event={"ID":"1003a846-f9a3-4562-ad4e-a449e7213dcd","Type":"ContainerDied","Data":"3a008049a5de0fe7c8325a6b979a4b5453c1d151d5db5e4e81fda5a4ce90086b"} Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.514013 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a008049a5de0fe7c8325a6b979a4b5453c1d151d5db5e4e81fda5a4ce90086b" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.517358 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" event={"ID":"02e070d5-c4c6-4464-90ce-faeddb3d8d47","Type":"ContainerDied","Data":"604f2e4280ffa7168a6361f2e6388ddafe92ed88faa57d12f97c4415791cfebe"} Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.517450 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-855cbc58c5-f64l5" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.526398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e070d5-c4c6-4464-90ce-faeddb3d8d47-config\") pod \"02e070d5-c4c6-4464-90ce-faeddb3d8d47\" (UID: \"02e070d5-c4c6-4464-90ce-faeddb3d8d47\") " Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.526533 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8lkh\" (UniqueName: \"kubernetes.io/projected/02e070d5-c4c6-4464-90ce-faeddb3d8d47-kube-api-access-x8lkh\") pod \"02e070d5-c4c6-4464-90ce-faeddb3d8d47\" (UID: \"02e070d5-c4c6-4464-90ce-faeddb3d8d47\") " Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.527789 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02e070d5-c4c6-4464-90ce-faeddb3d8d47-config" (OuterVolumeSpecName: "config") pod "02e070d5-c4c6-4464-90ce-faeddb3d8d47" (UID: "02e070d5-c4c6-4464-90ce-faeddb3d8d47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.542398 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02e070d5-c4c6-4464-90ce-faeddb3d8d47-kube-api-access-x8lkh" (OuterVolumeSpecName: "kube-api-access-x8lkh") pod "02e070d5-c4c6-4464-90ce-faeddb3d8d47" (UID: "02e070d5-c4c6-4464-90ce-faeddb3d8d47"). InnerVolumeSpecName "kube-api-access-x8lkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.546589 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.627932 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-dns-svc\") pod \"1003a846-f9a3-4562-ad4e-a449e7213dcd\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.628451 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1003a846-f9a3-4562-ad4e-a449e7213dcd" (UID: "1003a846-f9a3-4562-ad4e-a449e7213dcd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.628536 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhpnf\" (UniqueName: \"kubernetes.io/projected/1003a846-f9a3-4562-ad4e-a449e7213dcd-kube-api-access-vhpnf\") pod \"1003a846-f9a3-4562-ad4e-a449e7213dcd\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.628684 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-config\") pod \"1003a846-f9a3-4562-ad4e-a449e7213dcd\" (UID: \"1003a846-f9a3-4562-ad4e-a449e7213dcd\") " Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.629376 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.629403 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8lkh\" (UniqueName: \"kubernetes.io/projected/02e070d5-c4c6-4464-90ce-faeddb3d8d47-kube-api-access-x8lkh\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.629418 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e070d5-c4c6-4464-90ce-faeddb3d8d47-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.629851 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-config" (OuterVolumeSpecName: "config") pod "1003a846-f9a3-4562-ad4e-a449e7213dcd" (UID: "1003a846-f9a3-4562-ad4e-a449e7213dcd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.637780 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1003a846-f9a3-4562-ad4e-a449e7213dcd-kube-api-access-vhpnf" (OuterVolumeSpecName: "kube-api-access-vhpnf") pod "1003a846-f9a3-4562-ad4e-a449e7213dcd" (UID: "1003a846-f9a3-4562-ad4e-a449e7213dcd"). InnerVolumeSpecName "kube-api-access-vhpnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.733538 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhpnf\" (UniqueName: \"kubernetes.io/projected/1003a846-f9a3-4562-ad4e-a449e7213dcd-kube-api-access-vhpnf\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.733895 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1003a846-f9a3-4562-ad4e-a449e7213dcd-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:06 crc kubenswrapper[4846]: W0202 12:28:06.926934 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc75e5c92_9e5b_4d18_bb6a_d183c18d8a54.slice/crio-cac3239c4d40ee1f49d1191feea2cd4016924088a291fff1445d44df7bc5e639 WatchSource:0}: Error finding container cac3239c4d40ee1f49d1191feea2cd4016924088a291fff1445d44df7bc5e639: Status 404 returned error can't find the container with id cac3239c4d40ee1f49d1191feea2cd4016924088a291fff1445d44df7bc5e639 Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.934247 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.951593 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-f64l5"] Feb 02 12:28:06 crc kubenswrapper[4846]: I0202 12:28:06.957610 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-855cbc58c5-f64l5"] Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.025163 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.047371 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.058589 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.174147 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:28:07 crc kubenswrapper[4846]: W0202 12:28:07.176112 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c0a429c_a50e_493c_958e_61d8a3a404c6.slice/crio-58310c62e7ca162822f3972aa3f1805f863e8fe7b7d161842fffff5af17b05f9 WatchSource:0}: Error finding container 58310c62e7ca162822f3972aa3f1805f863e8fe7b7d161842fffff5af17b05f9: Status 404 returned error can't find the container with id 58310c62e7ca162822f3972aa3f1805f863e8fe7b7d161842fffff5af17b05f9 Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.209375 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b2hcc"] Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.263347 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-m79vg"] Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.450532 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02e070d5-c4c6-4464-90ce-faeddb3d8d47" path="/var/lib/kubelet/pods/02e070d5-c4c6-4464-90ce-faeddb3d8d47/volumes" Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.534874 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc" event={"ID":"bda23288-2fee-45ad-81e2-a14654c16589","Type":"ContainerStarted","Data":"de6f75ec04c419bcd53e3d033b604026fce641668fdb4d0cd503c6ac74c36ec4"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.538304 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"5342c3479165e81b2ca97736d850f5b8bc8a41ed04e9666168601ccf61d9277c"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.540860 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a7820c87-e5cc-42fd-b3f4-551cd39d3a15","Type":"ContainerStarted","Data":"af4efb7632ab326628f579b99842b6efa732df21f2bd455ab4c2227d445e0985"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.543284 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"36061d30-6265-47ca-8a87-5275f41d03ed","Type":"ContainerStarted","Data":"e56b4c06f56bec2f17b4f1e52b94306479c4be44b0443cd4e2e480cedf3904d7"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.545148 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"327c8143-b3ed-4431-a206-d2019ddcb606","Type":"ContainerStarted","Data":"f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.547481 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54","Type":"ContainerStarted","Data":"cac3239c4d40ee1f49d1191feea2cd4016924088a291fff1445d44df7bc5e639"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.558307 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8c0a429c-a50e-493c-958e-61d8a3a404c6","Type":"ContainerStarted","Data":"58310c62e7ca162822f3972aa3f1805f863e8fe7b7d161842fffff5af17b05f9"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.561767 4846 generic.go:334] "Generic (PLEG): container finished" podID="7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" containerID="b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39" exitCode=0 Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.561833 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" event={"ID":"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793","Type":"ContainerDied","Data":"b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.564613 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m79vg" event={"ID":"53dd5f18-d111-4359-8170-102a8b30d0d9","Type":"ContainerStarted","Data":"80ba3f6f664654c9815bb893f341567415125f9d92fad00f7248ac217eba3452"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.566631 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9cec353a-92f9-4ed4-8ec9-15271cbc6912","Type":"ContainerStarted","Data":"0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.569934 4846 generic.go:334] "Generic (PLEG): container finished" podID="24a64c0f-4d6c-4f6a-a2c4-659aed878323" containerID="91d2dedc4d1014aae9e5126721547aa83c88d706e8419d0ae73e4cf1c125efd2" exitCode=0 Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.570041 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" event={"ID":"24a64c0f-4d6c-4f6a-a2c4-659aed878323","Type":"ContainerDied","Data":"91d2dedc4d1014aae9e5126721547aa83c88d706e8419d0ae73e4cf1c125efd2"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.574420 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcf94d689-z889d" Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.574401 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a","Type":"ContainerStarted","Data":"d616050f9a822b3fb0ebb372d1934eca8cdf9b1fa2823a510a1a7315eb62b648"} Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.719849 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-z889d"] Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.728191 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6fcf94d689-z889d"] Feb 02 12:28:07 crc kubenswrapper[4846]: E0202 12:28:07.821440 4846 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Feb 02 12:28:07 crc kubenswrapper[4846]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 02 12:28:07 crc kubenswrapper[4846]: > podSandboxID="54065d6a56608ac45234fd9b53cf1c5d13a53b7871399058d17629738d730691" Feb 02 12:28:07 crc kubenswrapper[4846]: E0202 12:28:07.821874 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:28:07 crc kubenswrapper[4846]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-svzjm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-f54874ffc-xthhp_openstack(7eee9d79-99e5-4179-a3e1-8a9b9ef8c793): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 02 12:28:07 crc kubenswrapper[4846]: > logger="UnhandledError" Feb 02 12:28:07 crc kubenswrapper[4846]: E0202 12:28:07.823382 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" podUID="7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" Feb 02 12:28:07 crc kubenswrapper[4846]: I0202 12:28:07.832818 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 12:28:07 crc kubenswrapper[4846]: W0202 12:28:07.946732 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod797602a8_7d50_4ab4_8ba5_4afff6a34409.slice/crio-b4e9070cc9cc12c08941b4b81e99ada4d05d93e6997c9d731c3f64fef3b9183c WatchSource:0}: Error finding container b4e9070cc9cc12c08941b4b81e99ada4d05d93e6997c9d731c3f64fef3b9183c: Status 404 returned error can't find the container with id b4e9070cc9cc12c08941b4b81e99ada4d05d93e6997c9d731c3f64fef3b9183c Feb 02 12:28:08 crc kubenswrapper[4846]: I0202 12:28:08.590863 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" event={"ID":"24a64c0f-4d6c-4f6a-a2c4-659aed878323","Type":"ContainerStarted","Data":"e9d6f021bfb567f68a287dd7ad47757893a73f5c8f145cce18bb4eb3fd0776f8"} Feb 02 12:28:08 crc kubenswrapper[4846]: I0202 12:28:08.591156 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:28:08 crc kubenswrapper[4846]: I0202 12:28:08.592208 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"797602a8-7d50-4ab4-8ba5-4afff6a34409","Type":"ContainerStarted","Data":"b4e9070cc9cc12c08941b4b81e99ada4d05d93e6997c9d731c3f64fef3b9183c"} Feb 02 12:28:08 crc kubenswrapper[4846]: I0202 12:28:08.621677 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" podStartSLOduration=3.030819642 podStartE2EDuration="19.62165806s" podCreationTimestamp="2026-02-02 12:27:49 +0000 UTC" firstStartedPulling="2026-02-02 12:27:49.877717288 +0000 UTC m=+1101.106304151" lastFinishedPulling="2026-02-02 12:28:06.468555696 +0000 UTC m=+1117.697142569" observedRunningTime="2026-02-02 12:28:08.619067515 +0000 UTC m=+1119.847654378" watchObservedRunningTime="2026-02-02 12:28:08.62165806 +0000 UTC m=+1119.850244923" Feb 02 12:28:09 crc kubenswrapper[4846]: I0202 12:28:09.433875 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1003a846-f9a3-4562-ad4e-a449e7213dcd" path="/var/lib/kubelet/pods/1003a846-f9a3-4562-ad4e-a449e7213dcd/volumes" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.176073 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-nlrm4"] Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.177699 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.190322 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.198331 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nlrm4"] Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.322176 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-xthhp"] Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.327865 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cbhc\" (UniqueName: \"kubernetes.io/projected/fd01ba51-8a4c-4b74-a7ab-382600f93e68-kube-api-access-8cbhc\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.327943 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.328014 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd01ba51-8a4c-4b74-a7ab-382600f93e68-config\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.328064 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovn-rundir\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.328105 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovs-rundir\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.328146 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-combined-ca-bundle\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.365429 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-f6ngv"] Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.367041 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.376890 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.380741 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-f6ngv"] Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.431608 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cbhc\" (UniqueName: \"kubernetes.io/projected/fd01ba51-8a4c-4b74-a7ab-382600f93e68-kube-api-access-8cbhc\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.431718 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.431789 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd01ba51-8a4c-4b74-a7ab-382600f93e68-config\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.431838 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovn-rundir\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.431877 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovs-rundir\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.431911 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-combined-ca-bundle\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.433101 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd01ba51-8a4c-4b74-a7ab-382600f93e68-config\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.435754 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovn-rundir\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.435847 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovs-rundir\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.438197 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.438641 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-combined-ca-bundle\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.463517 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cbhc\" (UniqueName: \"kubernetes.io/projected/fd01ba51-8a4c-4b74-a7ab-382600f93e68-kube-api-access-8cbhc\") pod \"ovn-controller-metrics-nlrm4\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.474179 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-cfxzp"] Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.474383 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" podUID="24a64c0f-4d6c-4f6a-a2c4-659aed878323" containerName="dnsmasq-dns" containerID="cri-o://e9d6f021bfb567f68a287dd7ad47757893a73f5c8f145cce18bb4eb3fd0776f8" gracePeriod=10 Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.498698 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.511428 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-jlqjt"] Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.513310 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.518782 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.522449 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-jlqjt"] Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.533530 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-dns-svc\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.533869 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-config\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.534078 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g56wq\" (UniqueName: \"kubernetes.io/projected/2b7825cd-54bc-49e9-981f-9eb82b346bf9-kube-api-access-g56wq\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.534239 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-ovsdbserver-nb\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.614707 4846 generic.go:334] "Generic (PLEG): container finished" podID="24a64c0f-4d6c-4f6a-a2c4-659aed878323" containerID="e9d6f021bfb567f68a287dd7ad47757893a73f5c8f145cce18bb4eb3fd0776f8" exitCode=0 Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.614897 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" event={"ID":"24a64c0f-4d6c-4f6a-a2c4-659aed878323","Type":"ContainerDied","Data":"e9d6f021bfb567f68a287dd7ad47757893a73f5c8f145cce18bb4eb3fd0776f8"} Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.636004 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-dns-svc\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.636080 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-config\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.636119 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-dns-svc\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.636153 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npjlm\" (UniqueName: \"kubernetes.io/projected/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-kube-api-access-npjlm\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.636174 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-nb\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.636195 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-sb\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.636232 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g56wq\" (UniqueName: \"kubernetes.io/projected/2b7825cd-54bc-49e9-981f-9eb82b346bf9-kube-api-access-g56wq\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.636277 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-config\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.636315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-ovsdbserver-nb\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.637013 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-config\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.637140 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-ovsdbserver-nb\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.637305 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-dns-svc\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.655545 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g56wq\" (UniqueName: \"kubernetes.io/projected/2b7825cd-54bc-49e9-981f-9eb82b346bf9-kube-api-access-g56wq\") pod \"dnsmasq-dns-64f7f48db9-f6ngv\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.692968 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.738966 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-sb\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.739070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-config\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.739355 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-dns-svc\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.739384 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npjlm\" (UniqueName: \"kubernetes.io/projected/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-kube-api-access-npjlm\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.739403 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-nb\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.739948 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-sb\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.740216 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-config\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.740363 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-dns-svc\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.740683 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-nb\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.755325 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npjlm\" (UniqueName: \"kubernetes.io/projected/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-kube-api-access-npjlm\") pod \"dnsmasq-dns-56df986d9c-jlqjt\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:11 crc kubenswrapper[4846]: I0202 12:28:11.890951 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.591070 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.703461 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpd2g\" (UniqueName: \"kubernetes.io/projected/24a64c0f-4d6c-4f6a-a2c4-659aed878323-kube-api-access-xpd2g\") pod \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.704095 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-config\") pod \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.704193 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-dns-svc\") pod \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\" (UID: \"24a64c0f-4d6c-4f6a-a2c4-659aed878323\") " Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.718333 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24a64c0f-4d6c-4f6a-a2c4-659aed878323-kube-api-access-xpd2g" (OuterVolumeSpecName: "kube-api-access-xpd2g") pod "24a64c0f-4d6c-4f6a-a2c4-659aed878323" (UID: "24a64c0f-4d6c-4f6a-a2c4-659aed878323"). InnerVolumeSpecName "kube-api-access-xpd2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.729552 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" event={"ID":"24a64c0f-4d6c-4f6a-a2c4-659aed878323","Type":"ContainerDied","Data":"441ad332a730307b2b51e7a8416454bce188ea84830313e2c9053c65ee153324"} Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.730009 4846 scope.go:117] "RemoveContainer" containerID="e9d6f021bfb567f68a287dd7ad47757893a73f5c8f145cce18bb4eb3fd0776f8" Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.730129 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67ff45466c-cfxzp" Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.742492 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" podUID="7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" containerName="dnsmasq-dns" containerID="cri-o://a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4" gracePeriod=10 Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.742842 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.781843 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" podStartSLOduration=10.068406187 podStartE2EDuration="26.781823161s" podCreationTimestamp="2026-02-02 12:27:48 +0000 UTC" firstStartedPulling="2026-02-02 12:27:49.802864997 +0000 UTC m=+1101.031451860" lastFinishedPulling="2026-02-02 12:28:06.516281971 +0000 UTC m=+1117.744868834" observedRunningTime="2026-02-02 12:28:14.773222376 +0000 UTC m=+1126.001809259" watchObservedRunningTime="2026-02-02 12:28:14.781823161 +0000 UTC m=+1126.010410024" Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.805887 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpd2g\" (UniqueName: \"kubernetes.io/projected/24a64c0f-4d6c-4f6a-a2c4-659aed878323-kube-api-access-xpd2g\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.829943 4846 scope.go:117] "RemoveContainer" containerID="91d2dedc4d1014aae9e5126721547aa83c88d706e8419d0ae73e4cf1c125efd2" Feb 02 12:28:14 crc kubenswrapper[4846]: I0202 12:28:14.985614 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-config" (OuterVolumeSpecName: "config") pod "24a64c0f-4d6c-4f6a-a2c4-659aed878323" (UID: "24a64c0f-4d6c-4f6a-a2c4-659aed878323"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.009050 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.052659 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "24a64c0f-4d6c-4f6a-a2c4-659aed878323" (UID: "24a64c0f-4d6c-4f6a-a2c4-659aed878323"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.080513 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-f6ngv"] Feb 02 12:28:15 crc kubenswrapper[4846]: W0202 12:28:15.091132 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b7825cd_54bc_49e9_981f_9eb82b346bf9.slice/crio-a7d3649122de3c2178921734275209ec73e1122a15b8302b727188a0d5f48a61 WatchSource:0}: Error finding container a7d3649122de3c2178921734275209ec73e1122a15b8302b727188a0d5f48a61: Status 404 returned error can't find the container with id a7d3649122de3c2178921734275209ec73e1122a15b8302b727188a0d5f48a61 Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.112459 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24a64c0f-4d6c-4f6a-a2c4-659aed878323-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.202467 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-jlqjt"] Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.230276 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nlrm4"] Feb 02 12:28:15 crc kubenswrapper[4846]: W0202 12:28:15.282935 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30067520_bdb6_4fd5_b35e_0a4a2f8dd149.slice/crio-977bfb05d7dae5886fb12759dcc2baf23083b9e560e54e2ac79c18585bdb8ec4 WatchSource:0}: Error finding container 977bfb05d7dae5886fb12759dcc2baf23083b9e560e54e2ac79c18585bdb8ec4: Status 404 returned error can't find the container with id 977bfb05d7dae5886fb12759dcc2baf23083b9e560e54e2ac79c18585bdb8ec4 Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.384534 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.397860 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-cfxzp"] Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.423442 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svzjm\" (UniqueName: \"kubernetes.io/projected/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-kube-api-access-svzjm\") pod \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.423657 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-dns-svc\") pod \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.423695 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-config\") pod \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\" (UID: \"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793\") " Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.479381 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-kube-api-access-svzjm" (OuterVolumeSpecName: "kube-api-access-svzjm") pod "7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" (UID: "7eee9d79-99e5-4179-a3e1-8a9b9ef8c793"). InnerVolumeSpecName "kube-api-access-svzjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.505275 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67ff45466c-cfxzp"] Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.527454 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svzjm\" (UniqueName: \"kubernetes.io/projected/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-kube-api-access-svzjm\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.768875 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a","Type":"ContainerStarted","Data":"2849d11fa15affc0fa45246ebd204917e430eb9bef2190522d139fa1508c3879"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.786972 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nlrm4" event={"ID":"fd01ba51-8a4c-4b74-a7ab-382600f93e68","Type":"ContainerStarted","Data":"14b0c39bd579b305c315f664d839d215ca357c7387892d18fcf466087cdd5b44"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.805060 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"36061d30-6265-47ca-8a87-5275f41d03ed","Type":"ContainerStarted","Data":"b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.811424 4846 generic.go:334] "Generic (PLEG): container finished" podID="7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" containerID="a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4" exitCode=0 Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.811489 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" event={"ID":"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793","Type":"ContainerDied","Data":"a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.811512 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" event={"ID":"7eee9d79-99e5-4179-a3e1-8a9b9ef8c793","Type":"ContainerDied","Data":"54065d6a56608ac45234fd9b53cf1c5d13a53b7871399058d17629738d730691"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.811528 4846 scope.go:117] "RemoveContainer" containerID="a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.811699 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f54874ffc-xthhp" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.821661 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc" event={"ID":"bda23288-2fee-45ad-81e2-a14654c16589","Type":"ContainerStarted","Data":"ebe246dd795f71a1ad40ab382816989fc7965fac63ac05fe50ffa37d6e3182a0"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.822607 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-b2hcc" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.828266 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"797602a8-7d50-4ab4-8ba5-4afff6a34409","Type":"ContainerStarted","Data":"ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.845095 4846 generic.go:334] "Generic (PLEG): container finished" podID="2b7825cd-54bc-49e9-981f-9eb82b346bf9" containerID="8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2" exitCode=0 Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.845177 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" event={"ID":"2b7825cd-54bc-49e9-981f-9eb82b346bf9","Type":"ContainerDied","Data":"8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.845209 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" event={"ID":"2b7825cd-54bc-49e9-981f-9eb82b346bf9","Type":"ContainerStarted","Data":"a7d3649122de3c2178921734275209ec73e1122a15b8302b727188a0d5f48a61"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.852843 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-b2hcc" podStartSLOduration=9.482888284 podStartE2EDuration="16.85282499s" podCreationTimestamp="2026-02-02 12:27:59 +0000 UTC" firstStartedPulling="2026-02-02 12:28:07.210033119 +0000 UTC m=+1118.438619982" lastFinishedPulling="2026-02-02 12:28:14.579969835 +0000 UTC m=+1125.808556688" observedRunningTime="2026-02-02 12:28:15.852134752 +0000 UTC m=+1127.080721615" watchObservedRunningTime="2026-02-02 12:28:15.85282499 +0000 UTC m=+1127.081411843" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.858390 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m79vg" event={"ID":"53dd5f18-d111-4359-8170-102a8b30d0d9","Type":"ContainerStarted","Data":"5f54c9521f3ba3a9cace7a633d97e116ed3332a7c43fda1e5db49eeb3187a18b"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.870536 4846 scope.go:117] "RemoveContainer" containerID="b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.878965 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a7820c87-e5cc-42fd-b3f4-551cd39d3a15","Type":"ContainerStarted","Data":"20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.880758 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.902762 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" event={"ID":"30067520-bdb6-4fd5-b35e-0a4a2f8dd149","Type":"ContainerStarted","Data":"977bfb05d7dae5886fb12759dcc2baf23083b9e560e54e2ac79c18585bdb8ec4"} Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.920612 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=16.011205848 podStartE2EDuration="22.920593327s" podCreationTimestamp="2026-02-02 12:27:53 +0000 UTC" firstStartedPulling="2026-02-02 12:28:07.064441502 +0000 UTC m=+1118.293028355" lastFinishedPulling="2026-02-02 12:28:13.973828971 +0000 UTC m=+1125.202415834" observedRunningTime="2026-02-02 12:28:15.919457228 +0000 UTC m=+1127.148044091" watchObservedRunningTime="2026-02-02 12:28:15.920593327 +0000 UTC m=+1127.149180190" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.948313 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" (UID: "7eee9d79-99e5-4179-a3e1-8a9b9ef8c793"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:15 crc kubenswrapper[4846]: I0202 12:28:15.960661 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-config" (OuterVolumeSpecName: "config") pod "7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" (UID: "7eee9d79-99e5-4179-a3e1-8a9b9ef8c793"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.027395 4846 scope.go:117] "RemoveContainer" containerID="a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4" Feb 02 12:28:16 crc kubenswrapper[4846]: E0202 12:28:16.043640 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4\": container with ID starting with a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4 not found: ID does not exist" containerID="a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4" Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.043685 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4"} err="failed to get container status \"a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4\": rpc error: code = NotFound desc = could not find container \"a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4\": container with ID starting with a6e33d88a374edd2b58c08107716f559b808d0e2895f77388ab7ebb61043dfa4 not found: ID does not exist" Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.043710 4846 scope.go:117] "RemoveContainer" containerID="b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39" Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.044869 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.044900 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:16 crc kubenswrapper[4846]: E0202 12:28:16.045006 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39\": container with ID starting with b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39 not found: ID does not exist" containerID="b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39" Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.045083 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39"} err="failed to get container status \"b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39\": rpc error: code = NotFound desc = could not find container \"b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39\": container with ID starting with b0995899262c223d7da3d5a8ac41de249dce9e44897ef3bdd9e65f13f436da39 not found: ID does not exist" Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.178196 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-xthhp"] Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.185290 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f54874ffc-xthhp"] Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.914512 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" event={"ID":"2b7825cd-54bc-49e9-981f-9eb82b346bf9","Type":"ContainerStarted","Data":"6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb"} Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.914887 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.920828 4846 generic.go:334] "Generic (PLEG): container finished" podID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerID="5f54c9521f3ba3a9cace7a633d97e116ed3332a7c43fda1e5db49eeb3187a18b" exitCode=0 Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.920948 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m79vg" event={"ID":"53dd5f18-d111-4359-8170-102a8b30d0d9","Type":"ContainerDied","Data":"5f54c9521f3ba3a9cace7a633d97e116ed3332a7c43fda1e5db49eeb3187a18b"} Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.922799 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54","Type":"ContainerStarted","Data":"eb2f98a1c13c7bf89f3e7b636cf147de5beecff45f5771d9fc6fe245fe4a6f37"} Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.927809 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8c0a429c-a50e-493c-958e-61d8a3a404c6","Type":"ContainerStarted","Data":"1d67d357e4081d3bfaa9f34fd59e7809db1eb45b38eb505b82cd78ccaa91a725"} Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.928667 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.945786 4846 generic.go:334] "Generic (PLEG): container finished" podID="30067520-bdb6-4fd5-b35e-0a4a2f8dd149" containerID="7499ec96816d972ab810dad1499d7dfc7b95cd7d2d4b1476f511cb3440c77e9b" exitCode=0 Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.946492 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" event={"ID":"30067520-bdb6-4fd5-b35e-0a4a2f8dd149","Type":"ContainerDied","Data":"7499ec96816d972ab810dad1499d7dfc7b95cd7d2d4b1476f511cb3440c77e9b"} Feb 02 12:28:16 crc kubenswrapper[4846]: I0202 12:28:16.948526 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" podStartSLOduration=5.948513496 podStartE2EDuration="5.948513496s" podCreationTimestamp="2026-02-02 12:28:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:28:16.939168173 +0000 UTC m=+1128.167755046" watchObservedRunningTime="2026-02-02 12:28:16.948513496 +0000 UTC m=+1128.177100359" Feb 02 12:28:17 crc kubenswrapper[4846]: I0202 12:28:17.033209 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.51984051 podStartE2EDuration="23.033181748s" podCreationTimestamp="2026-02-02 12:27:54 +0000 UTC" firstStartedPulling="2026-02-02 12:28:07.178471289 +0000 UTC m=+1118.407058152" lastFinishedPulling="2026-02-02 12:28:14.691812527 +0000 UTC m=+1125.920399390" observedRunningTime="2026-02-02 12:28:16.994490658 +0000 UTC m=+1128.223077541" watchObservedRunningTime="2026-02-02 12:28:17.033181748 +0000 UTC m=+1128.261768631" Feb 02 12:28:17 crc kubenswrapper[4846]: I0202 12:28:17.434470 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24a64c0f-4d6c-4f6a-a2c4-659aed878323" path="/var/lib/kubelet/pods/24a64c0f-4d6c-4f6a-a2c4-659aed878323/volumes" Feb 02 12:28:17 crc kubenswrapper[4846]: I0202 12:28:17.435072 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" path="/var/lib/kubelet/pods/7eee9d79-99e5-4179-a3e1-8a9b9ef8c793/volumes" Feb 02 12:28:18 crc kubenswrapper[4846]: I0202 12:28:18.964386 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"36061d30-6265-47ca-8a87-5275f41d03ed","Type":"ContainerStarted","Data":"0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341"} Feb 02 12:28:18 crc kubenswrapper[4846]: I0202 12:28:18.966163 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" event={"ID":"30067520-bdb6-4fd5-b35e-0a4a2f8dd149","Type":"ContainerStarted","Data":"8b9a073035c47ad01504300f68639a51137c3f30e2c366ae687c78ea2b4a5d7a"} Feb 02 12:28:18 crc kubenswrapper[4846]: I0202 12:28:18.966308 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:18 crc kubenswrapper[4846]: I0202 12:28:18.967915 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nlrm4" event={"ID":"fd01ba51-8a4c-4b74-a7ab-382600f93e68","Type":"ContainerStarted","Data":"0448b20de6b4b9dfd9b17913d840a0f18245b4ede9e0f0f9853c064551787006"} Feb 02 12:28:18 crc kubenswrapper[4846]: I0202 12:28:18.970016 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"797602a8-7d50-4ab4-8ba5-4afff6a34409","Type":"ContainerStarted","Data":"8cd3e31565aad9a3a479cdc1ab4236b7a0085ec750392646a09a6087a0894db0"} Feb 02 12:28:18 crc kubenswrapper[4846]: I0202 12:28:18.972166 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m79vg" event={"ID":"53dd5f18-d111-4359-8170-102a8b30d0d9","Type":"ContainerStarted","Data":"7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481"} Feb 02 12:28:18 crc kubenswrapper[4846]: I0202 12:28:18.988027 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=9.440542364 podStartE2EDuration="20.988008295s" podCreationTimestamp="2026-02-02 12:27:58 +0000 UTC" firstStartedPulling="2026-02-02 12:28:07.035822326 +0000 UTC m=+1118.264409189" lastFinishedPulling="2026-02-02 12:28:18.583288257 +0000 UTC m=+1129.811875120" observedRunningTime="2026-02-02 12:28:18.986701573 +0000 UTC m=+1130.215288436" watchObservedRunningTime="2026-02-02 12:28:18.988008295 +0000 UTC m=+1130.216595158" Feb 02 12:28:19 crc kubenswrapper[4846]: I0202 12:28:19.038909 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=8.37774924 podStartE2EDuration="19.03888883s" podCreationTimestamp="2026-02-02 12:28:00 +0000 UTC" firstStartedPulling="2026-02-02 12:28:07.950024616 +0000 UTC m=+1119.178611479" lastFinishedPulling="2026-02-02 12:28:18.611164206 +0000 UTC m=+1129.839751069" observedRunningTime="2026-02-02 12:28:19.021353311 +0000 UTC m=+1130.249940194" watchObservedRunningTime="2026-02-02 12:28:19.03888883 +0000 UTC m=+1130.267475693" Feb 02 12:28:19 crc kubenswrapper[4846]: I0202 12:28:19.048554 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-nlrm4" podStartSLOduration=4.744790052 podStartE2EDuration="8.048534991s" podCreationTimestamp="2026-02-02 12:28:11 +0000 UTC" firstStartedPulling="2026-02-02 12:28:15.291814686 +0000 UTC m=+1126.520401549" lastFinishedPulling="2026-02-02 12:28:18.595559635 +0000 UTC m=+1129.824146488" observedRunningTime="2026-02-02 12:28:19.046808018 +0000 UTC m=+1130.275394891" watchObservedRunningTime="2026-02-02 12:28:19.048534991 +0000 UTC m=+1130.277121864" Feb 02 12:28:19 crc kubenswrapper[4846]: I0202 12:28:19.103903 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" podStartSLOduration=8.103881428 podStartE2EDuration="8.103881428s" podCreationTimestamp="2026-02-02 12:28:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:28:19.075650941 +0000 UTC m=+1130.304237804" watchObservedRunningTime="2026-02-02 12:28:19.103881428 +0000 UTC m=+1130.332468291" Feb 02 12:28:19 crc kubenswrapper[4846]: I0202 12:28:19.949406 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 02 12:28:19 crc kubenswrapper[4846]: I0202 12:28:19.982510 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m79vg" event={"ID":"53dd5f18-d111-4359-8170-102a8b30d0d9","Type":"ContainerStarted","Data":"c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a"} Feb 02 12:28:19 crc kubenswrapper[4846]: I0202 12:28:19.982773 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:28:19 crc kubenswrapper[4846]: I0202 12:28:19.985251 4846 generic.go:334] "Generic (PLEG): container finished" podID="c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" containerID="eb2f98a1c13c7bf89f3e7b636cf147de5beecff45f5771d9fc6fe245fe4a6f37" exitCode=0 Feb 02 12:28:19 crc kubenswrapper[4846]: I0202 12:28:19.985352 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54","Type":"ContainerDied","Data":"eb2f98a1c13c7bf89f3e7b636cf147de5beecff45f5771d9fc6fe245fe4a6f37"} Feb 02 12:28:19 crc kubenswrapper[4846]: I0202 12:28:19.993224 4846 generic.go:334] "Generic (PLEG): container finished" podID="b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" containerID="2849d11fa15affc0fa45246ebd204917e430eb9bef2190522d139fa1508c3879" exitCode=0 Feb 02 12:28:19 crc kubenswrapper[4846]: I0202 12:28:19.993277 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a","Type":"ContainerDied","Data":"2849d11fa15affc0fa45246ebd204917e430eb9bef2190522d139fa1508c3879"} Feb 02 12:28:20 crc kubenswrapper[4846]: I0202 12:28:20.021273 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-m79vg" podStartSLOduration=14.286828743 podStartE2EDuration="21.021249858s" podCreationTimestamp="2026-02-02 12:27:59 +0000 UTC" firstStartedPulling="2026-02-02 12:28:07.266854393 +0000 UTC m=+1118.495441256" lastFinishedPulling="2026-02-02 12:28:14.001275498 +0000 UTC m=+1125.229862371" observedRunningTime="2026-02-02 12:28:20.00617709 +0000 UTC m=+1131.234763953" watchObservedRunningTime="2026-02-02 12:28:20.021249858 +0000 UTC m=+1131.249836731" Feb 02 12:28:20 crc kubenswrapper[4846]: I0202 12:28:20.092894 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:20 crc kubenswrapper[4846]: I0202 12:28:20.144116 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:20 crc kubenswrapper[4846]: I0202 12:28:20.949223 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 02 12:28:20 crc kubenswrapper[4846]: I0202 12:28:20.986394 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.003828 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a","Type":"ContainerStarted","Data":"e484c823759e2a5e61def70b96f67d07c7974e29d5c4f9fc17eee40d86e83c71"} Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.011594 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54","Type":"ContainerStarted","Data":"481554f18c651c06a0af22ab57eb2b59897fae18ef156d9d05b24af20a349d9a"} Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.011849 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.012255 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.026548 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=23.956822666 podStartE2EDuration="31.02653024s" podCreationTimestamp="2026-02-02 12:27:50 +0000 UTC" firstStartedPulling="2026-02-02 12:28:07.064141175 +0000 UTC m=+1118.292728038" lastFinishedPulling="2026-02-02 12:28:14.133848749 +0000 UTC m=+1125.362435612" observedRunningTime="2026-02-02 12:28:21.02451057 +0000 UTC m=+1132.253097443" watchObservedRunningTime="2026-02-02 12:28:21.02653024 +0000 UTC m=+1132.255117103" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.056809 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.058034 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.41405055 podStartE2EDuration="29.05802068s" podCreationTimestamp="2026-02-02 12:27:52 +0000 UTC" firstStartedPulling="2026-02-02 12:28:06.93580328 +0000 UTC m=+1118.164390143" lastFinishedPulling="2026-02-02 12:28:14.5797734 +0000 UTC m=+1125.808360273" observedRunningTime="2026-02-02 12:28:21.042584772 +0000 UTC m=+1132.271171655" watchObservedRunningTime="2026-02-02 12:28:21.05802068 +0000 UTC m=+1132.286607553" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.060214 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.420585 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 02 12:28:21 crc kubenswrapper[4846]: E0202 12:28:21.420917 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a64c0f-4d6c-4f6a-a2c4-659aed878323" containerName="dnsmasq-dns" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.420928 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a64c0f-4d6c-4f6a-a2c4-659aed878323" containerName="dnsmasq-dns" Feb 02 12:28:21 crc kubenswrapper[4846]: E0202 12:28:21.420957 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" containerName="init" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.420963 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" containerName="init" Feb 02 12:28:21 crc kubenswrapper[4846]: E0202 12:28:21.420982 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" containerName="dnsmasq-dns" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.420989 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" containerName="dnsmasq-dns" Feb 02 12:28:21 crc kubenswrapper[4846]: E0202 12:28:21.421006 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a64c0f-4d6c-4f6a-a2c4-659aed878323" containerName="init" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.421011 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a64c0f-4d6c-4f6a-a2c4-659aed878323" containerName="init" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.421166 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eee9d79-99e5-4179-a3e1-8a9b9ef8c793" containerName="dnsmasq-dns" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.421180 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="24a64c0f-4d6c-4f6a-a2c4-659aed878323" containerName="dnsmasq-dns" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.421891 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.431599 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.431641 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.431848 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.431957 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-npjxc" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.453421 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-scripts\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.453507 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.453658 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wsbq\" (UniqueName: \"kubernetes.io/projected/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-kube-api-access-5wsbq\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.453742 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-config\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.453788 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.453830 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.453867 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.459845 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.555863 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wsbq\" (UniqueName: \"kubernetes.io/projected/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-kube-api-access-5wsbq\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.555922 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-config\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.555953 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.555976 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.556016 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.556047 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-scripts\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.556078 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.556748 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-config\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.557317 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-scripts\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.557455 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.563796 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.568477 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.568850 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.570893 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wsbq\" (UniqueName: \"kubernetes.io/projected/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-kube-api-access-5wsbq\") pod \"ovn-northd-0\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.694824 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.770998 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.931354 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 02 12:28:21 crc kubenswrapper[4846]: I0202 12:28:21.931601 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 02 12:28:22 crc kubenswrapper[4846]: I0202 12:28:22.311588 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 02 12:28:22 crc kubenswrapper[4846]: W0202 12:28:22.315285 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0ec2404_1bf0_4fc1_bc66_1b2023c7cd9e.slice/crio-be0202f872b03082898f8eac24324b031c5a25cacce1fb82f674633ee4244ce3 WatchSource:0}: Error finding container be0202f872b03082898f8eac24324b031c5a25cacce1fb82f674633ee4244ce3: Status 404 returned error can't find the container with id be0202f872b03082898f8eac24324b031c5a25cacce1fb82f674633ee4244ce3 Feb 02 12:28:23 crc kubenswrapper[4846]: I0202 12:28:23.026224 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e","Type":"ContainerStarted","Data":"be0202f872b03082898f8eac24324b031c5a25cacce1fb82f674633ee4244ce3"} Feb 02 12:28:23 crc kubenswrapper[4846]: I0202 12:28:23.435334 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 02 12:28:23 crc kubenswrapper[4846]: I0202 12:28:23.435684 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 02 12:28:23 crc kubenswrapper[4846]: I0202 12:28:23.615808 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 02 12:28:24 crc kubenswrapper[4846]: I0202 12:28:24.034547 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e","Type":"ContainerStarted","Data":"a6d2a885c9e46febf0a5624587138a27dc292d388c2adfac2d4f55ce324f5505"} Feb 02 12:28:24 crc kubenswrapper[4846]: I0202 12:28:24.034970 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 02 12:28:24 crc kubenswrapper[4846]: I0202 12:28:24.034985 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e","Type":"ContainerStarted","Data":"a60702e8aab62dc62e3d8fd855c2cb15d42dc1d569a8a5f56df75bb229c8672e"} Feb 02 12:28:24 crc kubenswrapper[4846]: I0202 12:28:24.064326 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.282200654 podStartE2EDuration="3.064310406s" podCreationTimestamp="2026-02-02 12:28:21 +0000 UTC" firstStartedPulling="2026-02-02 12:28:22.321701824 +0000 UTC m=+1133.550288687" lastFinishedPulling="2026-02-02 12:28:23.103811566 +0000 UTC m=+1134.332398439" observedRunningTime="2026-02-02 12:28:24.05850284 +0000 UTC m=+1135.287089723" watchObservedRunningTime="2026-02-02 12:28:24.064310406 +0000 UTC m=+1135.292897269" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.197716 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-jlqjt"] Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.197997 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" podUID="30067520-bdb6-4fd5-b35e-0a4a2f8dd149" containerName="dnsmasq-dns" containerID="cri-o://8b9a073035c47ad01504300f68639a51137c3f30e2c366ae687c78ea2b4a5d7a" gracePeriod=10 Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.205323 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.317939 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.355150 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-hzrlk"] Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.360362 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.388640 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-hzrlk"] Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.525508 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-config\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.525549 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvg2z\" (UniqueName: \"kubernetes.io/projected/240e8c55-ad59-4bee-a92e-93876302d7f5-kube-api-access-hvg2z\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.525577 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-dns-svc\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.525611 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-nb\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.525656 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-sb\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.627581 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-config\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.627646 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvg2z\" (UniqueName: \"kubernetes.io/projected/240e8c55-ad59-4bee-a92e-93876302d7f5-kube-api-access-hvg2z\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.627684 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-dns-svc\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.627712 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-nb\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.627746 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-sb\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.628743 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-config\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.628776 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-dns-svc\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.628789 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-sb\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.629007 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-nb\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.647800 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvg2z\" (UniqueName: \"kubernetes.io/projected/240e8c55-ad59-4bee-a92e-93876302d7f5-kube-api-access-hvg2z\") pod \"dnsmasq-dns-66b577f8c-hzrlk\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:25 crc kubenswrapper[4846]: I0202 12:28:25.748500 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.055818 4846 generic.go:334] "Generic (PLEG): container finished" podID="30067520-bdb6-4fd5-b35e-0a4a2f8dd149" containerID="8b9a073035c47ad01504300f68639a51137c3f30e2c366ae687c78ea2b4a5d7a" exitCode=0 Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.055944 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" event={"ID":"30067520-bdb6-4fd5-b35e-0a4a2f8dd149","Type":"ContainerDied","Data":"8b9a073035c47ad01504300f68639a51137c3f30e2c366ae687c78ea2b4a5d7a"} Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.194507 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-hzrlk"] Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.519191 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.526741 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.529236 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.529615 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-jqj6n" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.529618 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.530307 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.554494 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.647238 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-cache\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.647360 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2ctx\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-kube-api-access-b2ctx\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.647491 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.647837 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.647898 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-lock\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.648035 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309e8cc0-3b70-4961-b69c-1abf373f3c61-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.749394 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309e8cc0-3b70-4961-b69c-1abf373f3c61-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.749476 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-cache\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.749541 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2ctx\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-kube-api-access-b2ctx\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.749570 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.749606 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.749650 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-lock\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.750385 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-lock\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.751448 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-cache\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.751554 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: E0202 12:28:26.751570 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 02 12:28:26 crc kubenswrapper[4846]: E0202 12:28:26.751586 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 02 12:28:26 crc kubenswrapper[4846]: E0202 12:28:26.751655 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift podName:309e8cc0-3b70-4961-b69c-1abf373f3c61 nodeName:}" failed. No retries permitted until 2026-02-02 12:28:27.251610362 +0000 UTC m=+1138.480197225 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift") pod "swift-storage-0" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61") : configmap "swift-ring-files" not found Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.755031 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309e8cc0-3b70-4961-b69c-1abf373f3c61-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.771042 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2ctx\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-kube-api-access-b2ctx\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.774294 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:26 crc kubenswrapper[4846]: I0202 12:28:26.892596 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" podUID="30067520-bdb6-4fd5-b35e-0a4a2f8dd149" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.055216 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-wj5sj"] Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.056478 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.062342 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.065516 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" event={"ID":"240e8c55-ad59-4bee-a92e-93876302d7f5","Type":"ContainerStarted","Data":"41f357ff8f68cee95cec3690836cfc9d309b40430d83a1ef97613e78e69f6740"} Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.066055 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.068354 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wj5sj"] Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.069079 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.157100 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-swiftconf\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.157157 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-scripts\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.157179 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-ring-data-devices\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.157212 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-combined-ca-bundle\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.157232 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ece3e0cc-5193-4c71-982b-84fe26ad56e1-etc-swift\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.157316 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qrkt\" (UniqueName: \"kubernetes.io/projected/ece3e0cc-5193-4c71-982b-84fe26ad56e1-kube-api-access-9qrkt\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.157369 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-dispersionconf\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.255468 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.259446 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-swiftconf\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.259552 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-scripts\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.259597 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-ring-data-devices\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.259680 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-combined-ca-bundle\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.259739 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ece3e0cc-5193-4c71-982b-84fe26ad56e1-etc-swift\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.259782 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qrkt\" (UniqueName: \"kubernetes.io/projected/ece3e0cc-5193-4c71-982b-84fe26ad56e1-kube-api-access-9qrkt\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.259806 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-dispersionconf\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.259851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:27 crc kubenswrapper[4846]: E0202 12:28:27.260056 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 02 12:28:27 crc kubenswrapper[4846]: E0202 12:28:27.260080 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 02 12:28:27 crc kubenswrapper[4846]: E0202 12:28:27.260130 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift podName:309e8cc0-3b70-4961-b69c-1abf373f3c61 nodeName:}" failed. No retries permitted until 2026-02-02 12:28:28.26011417 +0000 UTC m=+1139.488701043 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift") pod "swift-storage-0" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61") : configmap "swift-ring-files" not found Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.260563 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-scripts\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.260836 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ece3e0cc-5193-4c71-982b-84fe26ad56e1-etc-swift\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.260999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-ring-data-devices\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.265326 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-combined-ca-bundle\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.265895 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-dispersionconf\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.268403 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-swiftconf\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.290755 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qrkt\" (UniqueName: \"kubernetes.io/projected/ece3e0cc-5193-4c71-982b-84fe26ad56e1-kube-api-access-9qrkt\") pod \"swift-ring-rebalance-wj5sj\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.359829 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" containerName="galera" probeResult="failure" output=< Feb 02 12:28:27 crc kubenswrapper[4846]: wsrep_local_state_comment (Joined) differs from Synced Feb 02 12:28:27 crc kubenswrapper[4846]: > Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.378928 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:27 crc kubenswrapper[4846]: I0202 12:28:27.823137 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wj5sj"] Feb 02 12:28:28 crc kubenswrapper[4846]: I0202 12:28:28.074354 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wj5sj" event={"ID":"ece3e0cc-5193-4c71-982b-84fe26ad56e1","Type":"ContainerStarted","Data":"10dae633a3b235287ec29708ef6c3fd4a234cd1f0512003b0227bae4f6e3ef2d"} Feb 02 12:28:28 crc kubenswrapper[4846]: I0202 12:28:28.276930 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:28 crc kubenswrapper[4846]: E0202 12:28:28.277100 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 02 12:28:28 crc kubenswrapper[4846]: E0202 12:28:28.277128 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 02 12:28:28 crc kubenswrapper[4846]: E0202 12:28:28.277199 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift podName:309e8cc0-3b70-4961-b69c-1abf373f3c61 nodeName:}" failed. No retries permitted until 2026-02-02 12:28:30.277176968 +0000 UTC m=+1141.505763831 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift") pod "swift-storage-0" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61") : configmap "swift-ring-files" not found Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.541438 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.604124 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-nb\") pod \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.604193 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-config\") pod \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.604277 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-sb\") pod \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.604373 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-dns-svc\") pod \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.604440 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npjlm\" (UniqueName: \"kubernetes.io/projected/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-kube-api-access-npjlm\") pod \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\" (UID: \"30067520-bdb6-4fd5-b35e-0a4a2f8dd149\") " Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.610230 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-kube-api-access-npjlm" (OuterVolumeSpecName: "kube-api-access-npjlm") pod "30067520-bdb6-4fd5-b35e-0a4a2f8dd149" (UID: "30067520-bdb6-4fd5-b35e-0a4a2f8dd149"). InnerVolumeSpecName "kube-api-access-npjlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.650500 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "30067520-bdb6-4fd5-b35e-0a4a2f8dd149" (UID: "30067520-bdb6-4fd5-b35e-0a4a2f8dd149"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.650909 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "30067520-bdb6-4fd5-b35e-0a4a2f8dd149" (UID: "30067520-bdb6-4fd5-b35e-0a4a2f8dd149"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.651461 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "30067520-bdb6-4fd5-b35e-0a4a2f8dd149" (UID: "30067520-bdb6-4fd5-b35e-0a4a2f8dd149"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.652690 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-config" (OuterVolumeSpecName: "config") pod "30067520-bdb6-4fd5-b35e-0a4a2f8dd149" (UID: "30067520-bdb6-4fd5-b35e-0a4a2f8dd149"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.707755 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.707795 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.707806 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.707817 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:29 crc kubenswrapper[4846]: I0202 12:28:29.707829 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npjlm\" (UniqueName: \"kubernetes.io/projected/30067520-bdb6-4fd5-b35e-0a4a2f8dd149-kube-api-access-npjlm\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.045443 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.092450 4846 generic.go:334] "Generic (PLEG): container finished" podID="240e8c55-ad59-4bee-a92e-93876302d7f5" containerID="be8dea291f883c78f6a9fe42a55b49ce1c88d18d046987607886c951725e9c18" exitCode=0 Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.092552 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" event={"ID":"240e8c55-ad59-4bee-a92e-93876302d7f5","Type":"ContainerDied","Data":"be8dea291f883c78f6a9fe42a55b49ce1c88d18d046987607886c951725e9c18"} Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.097268 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" event={"ID":"30067520-bdb6-4fd5-b35e-0a4a2f8dd149","Type":"ContainerDied","Data":"977bfb05d7dae5886fb12759dcc2baf23083b9e560e54e2ac79c18585bdb8ec4"} Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.097339 4846 scope.go:117] "RemoveContainer" containerID="8b9a073035c47ad01504300f68639a51137c3f30e2c366ae687c78ea2b4a5d7a" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.097446 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df986d9c-jlqjt" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.134023 4846 scope.go:117] "RemoveContainer" containerID="7499ec96816d972ab810dad1499d7dfc7b95cd7d2d4b1476f511cb3440c77e9b" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.134204 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.187309 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-jlqjt"] Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.193904 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df986d9c-jlqjt"] Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.337390 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:30 crc kubenswrapper[4846]: E0202 12:28:30.337591 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 02 12:28:30 crc kubenswrapper[4846]: E0202 12:28:30.337607 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 02 12:28:30 crc kubenswrapper[4846]: E0202 12:28:30.337680 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift podName:309e8cc0-3b70-4961-b69c-1abf373f3c61 nodeName:}" failed. No retries permitted until 2026-02-02 12:28:34.337661222 +0000 UTC m=+1145.566248085 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift") pod "swift-storage-0" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61") : configmap "swift-ring-files" not found Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.651402 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-cpnvx"] Feb 02 12:28:30 crc kubenswrapper[4846]: E0202 12:28:30.651998 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30067520-bdb6-4fd5-b35e-0a4a2f8dd149" containerName="dnsmasq-dns" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.652011 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="30067520-bdb6-4fd5-b35e-0a4a2f8dd149" containerName="dnsmasq-dns" Feb 02 12:28:30 crc kubenswrapper[4846]: E0202 12:28:30.652031 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30067520-bdb6-4fd5-b35e-0a4a2f8dd149" containerName="init" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.652036 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="30067520-bdb6-4fd5-b35e-0a4a2f8dd149" containerName="init" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.652210 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="30067520-bdb6-4fd5-b35e-0a4a2f8dd149" containerName="dnsmasq-dns" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.652761 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cpnvx" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.656187 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.660492 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-cpnvx"] Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.745017 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cea47b6-6835-4c98-9006-c7d10e0e7457-operator-scripts\") pod \"root-account-create-update-cpnvx\" (UID: \"0cea47b6-6835-4c98-9006-c7d10e0e7457\") " pod="openstack/root-account-create-update-cpnvx" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.745074 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvjj2\" (UniqueName: \"kubernetes.io/projected/0cea47b6-6835-4c98-9006-c7d10e0e7457-kube-api-access-cvjj2\") pod \"root-account-create-update-cpnvx\" (UID: \"0cea47b6-6835-4c98-9006-c7d10e0e7457\") " pod="openstack/root-account-create-update-cpnvx" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.847403 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cea47b6-6835-4c98-9006-c7d10e0e7457-operator-scripts\") pod \"root-account-create-update-cpnvx\" (UID: \"0cea47b6-6835-4c98-9006-c7d10e0e7457\") " pod="openstack/root-account-create-update-cpnvx" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.847451 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvjj2\" (UniqueName: \"kubernetes.io/projected/0cea47b6-6835-4c98-9006-c7d10e0e7457-kube-api-access-cvjj2\") pod \"root-account-create-update-cpnvx\" (UID: \"0cea47b6-6835-4c98-9006-c7d10e0e7457\") " pod="openstack/root-account-create-update-cpnvx" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.848544 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cea47b6-6835-4c98-9006-c7d10e0e7457-operator-scripts\") pod \"root-account-create-update-cpnvx\" (UID: \"0cea47b6-6835-4c98-9006-c7d10e0e7457\") " pod="openstack/root-account-create-update-cpnvx" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.894366 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvjj2\" (UniqueName: \"kubernetes.io/projected/0cea47b6-6835-4c98-9006-c7d10e0e7457-kube-api-access-cvjj2\") pod \"root-account-create-update-cpnvx\" (UID: \"0cea47b6-6835-4c98-9006-c7d10e0e7457\") " pod="openstack/root-account-create-update-cpnvx" Feb 02 12:28:30 crc kubenswrapper[4846]: I0202 12:28:30.971307 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cpnvx" Feb 02 12:28:31 crc kubenswrapper[4846]: I0202 12:28:31.444795 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30067520-bdb6-4fd5-b35e-0a4a2f8dd149" path="/var/lib/kubelet/pods/30067520-bdb6-4fd5-b35e-0a4a2f8dd149/volumes" Feb 02 12:28:32 crc kubenswrapper[4846]: I0202 12:28:32.805566 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-cpnvx"] Feb 02 12:28:32 crc kubenswrapper[4846]: W0202 12:28:32.813107 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cea47b6_6835_4c98_9006_c7d10e0e7457.slice/crio-53128b23d4642df840c5e5050e30b515dce977829e94e30b456c93bfc3db8645 WatchSource:0}: Error finding container 53128b23d4642df840c5e5050e30b515dce977829e94e30b456c93bfc3db8645: Status 404 returned error can't find the container with id 53128b23d4642df840c5e5050e30b515dce977829e94e30b456c93bfc3db8645 Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.126787 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wj5sj" event={"ID":"ece3e0cc-5193-4c71-982b-84fe26ad56e1","Type":"ContainerStarted","Data":"7e20049cbc1c96a32d8a5a8a599eef2e575b73f6df35000995ce4b76e4f2a280"} Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.129761 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-cpnvx" event={"ID":"0cea47b6-6835-4c98-9006-c7d10e0e7457","Type":"ContainerStarted","Data":"9baac131245f62b73851f82758eac3beae324a624e97690d13a59f6f6d3e2bd8"} Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.129953 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-cpnvx" event={"ID":"0cea47b6-6835-4c98-9006-c7d10e0e7457","Type":"ContainerStarted","Data":"53128b23d4642df840c5e5050e30b515dce977829e94e30b456c93bfc3db8645"} Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.132252 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" event={"ID":"240e8c55-ad59-4bee-a92e-93876302d7f5","Type":"ContainerStarted","Data":"6e21400ac8e3c4e79243ecc8b3b8272649285daca0a5f4b7d6f4cc9592ac49c6"} Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.132939 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.149429 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-wj5sj" podStartSLOduration=1.593855179 podStartE2EDuration="6.149406775s" podCreationTimestamp="2026-02-02 12:28:27 +0000 UTC" firstStartedPulling="2026-02-02 12:28:27.834012966 +0000 UTC m=+1139.062599849" lastFinishedPulling="2026-02-02 12:28:32.389564582 +0000 UTC m=+1143.618151445" observedRunningTime="2026-02-02 12:28:33.143936808 +0000 UTC m=+1144.372523681" watchObservedRunningTime="2026-02-02 12:28:33.149406775 +0000 UTC m=+1144.377993638" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.164756 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" podStartSLOduration=8.164736739 podStartE2EDuration="8.164736739s" podCreationTimestamp="2026-02-02 12:28:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:28:33.162479983 +0000 UTC m=+1144.391066856" watchObservedRunningTime="2026-02-02 12:28:33.164736739 +0000 UTC m=+1144.393323622" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.232192 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-xvdwm"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.233715 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xvdwm" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.245966 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xvdwm"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.295420 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-operator-scripts\") pod \"keystone-db-create-xvdwm\" (UID: \"534b394f-d62c-4c1d-8528-3a8aa37d5d7c\") " pod="openstack/keystone-db-create-xvdwm" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.295519 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjbww\" (UniqueName: \"kubernetes.io/projected/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-kube-api-access-sjbww\") pod \"keystone-db-create-xvdwm\" (UID: \"534b394f-d62c-4c1d-8528-3a8aa37d5d7c\") " pod="openstack/keystone-db-create-xvdwm" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.332295 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5b76-account-create-update-gm2qj"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.333291 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b76-account-create-update-gm2qj" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.334971 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.348912 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b76-account-create-update-gm2qj"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.396776 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-operator-scripts\") pod \"keystone-db-create-xvdwm\" (UID: \"534b394f-d62c-4c1d-8528-3a8aa37d5d7c\") " pod="openstack/keystone-db-create-xvdwm" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.396893 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d858780-98c0-4167-9bc5-8a4cc0e054da-operator-scripts\") pod \"keystone-5b76-account-create-update-gm2qj\" (UID: \"0d858780-98c0-4167-9bc5-8a4cc0e054da\") " pod="openstack/keystone-5b76-account-create-update-gm2qj" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.396975 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjbww\" (UniqueName: \"kubernetes.io/projected/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-kube-api-access-sjbww\") pod \"keystone-db-create-xvdwm\" (UID: \"534b394f-d62c-4c1d-8528-3a8aa37d5d7c\") " pod="openstack/keystone-db-create-xvdwm" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.397063 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xw72\" (UniqueName: \"kubernetes.io/projected/0d858780-98c0-4167-9bc5-8a4cc0e054da-kube-api-access-7xw72\") pod \"keystone-5b76-account-create-update-gm2qj\" (UID: \"0d858780-98c0-4167-9bc5-8a4cc0e054da\") " pod="openstack/keystone-5b76-account-create-update-gm2qj" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.397940 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-operator-scripts\") pod \"keystone-db-create-xvdwm\" (UID: \"534b394f-d62c-4c1d-8528-3a8aa37d5d7c\") " pod="openstack/keystone-db-create-xvdwm" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.416391 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjbww\" (UniqueName: \"kubernetes.io/projected/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-kube-api-access-sjbww\") pod \"keystone-db-create-xvdwm\" (UID: \"534b394f-d62c-4c1d-8528-3a8aa37d5d7c\") " pod="openstack/keystone-db-create-xvdwm" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.493250 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.498524 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d858780-98c0-4167-9bc5-8a4cc0e054da-operator-scripts\") pod \"keystone-5b76-account-create-update-gm2qj\" (UID: \"0d858780-98c0-4167-9bc5-8a4cc0e054da\") " pod="openstack/keystone-5b76-account-create-update-gm2qj" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.499221 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d858780-98c0-4167-9bc5-8a4cc0e054da-operator-scripts\") pod \"keystone-5b76-account-create-update-gm2qj\" (UID: \"0d858780-98c0-4167-9bc5-8a4cc0e054da\") " pod="openstack/keystone-5b76-account-create-update-gm2qj" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.499417 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xw72\" (UniqueName: \"kubernetes.io/projected/0d858780-98c0-4167-9bc5-8a4cc0e054da-kube-api-access-7xw72\") pod \"keystone-5b76-account-create-update-gm2qj\" (UID: \"0d858780-98c0-4167-9bc5-8a4cc0e054da\") " pod="openstack/keystone-5b76-account-create-update-gm2qj" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.529166 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xw72\" (UniqueName: \"kubernetes.io/projected/0d858780-98c0-4167-9bc5-8a4cc0e054da-kube-api-access-7xw72\") pod \"keystone-5b76-account-create-update-gm2qj\" (UID: \"0d858780-98c0-4167-9bc5-8a4cc0e054da\") " pod="openstack/keystone-5b76-account-create-update-gm2qj" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.552733 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xvdwm" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.590686 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-crpfz"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.591867 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-crpfz" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.621013 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-crpfz"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.707347 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1269aae-c267-46d0-8adb-cdbe01430a31-operator-scripts\") pod \"placement-db-create-crpfz\" (UID: \"c1269aae-c267-46d0-8adb-cdbe01430a31\") " pod="openstack/placement-db-create-crpfz" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.707818 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlcdm\" (UniqueName: \"kubernetes.io/projected/c1269aae-c267-46d0-8adb-cdbe01430a31-kube-api-access-hlcdm\") pod \"placement-db-create-crpfz\" (UID: \"c1269aae-c267-46d0-8adb-cdbe01430a31\") " pod="openstack/placement-db-create-crpfz" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.712923 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a0f2-account-create-update-pmqhw"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.714039 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0f2-account-create-update-pmqhw" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.719734 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.720273 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b76-account-create-update-gm2qj" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.721442 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a0f2-account-create-update-pmqhw"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.785551 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-c64lq"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.786712 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c64lq" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.794211 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-c64lq"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.809771 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7zgd\" (UniqueName: \"kubernetes.io/projected/8ad01c88-75f3-49ef-be2a-19ca711923bd-kube-api-access-h7zgd\") pod \"placement-a0f2-account-create-update-pmqhw\" (UID: \"8ad01c88-75f3-49ef-be2a-19ca711923bd\") " pod="openstack/placement-a0f2-account-create-update-pmqhw" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.809823 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1269aae-c267-46d0-8adb-cdbe01430a31-operator-scripts\") pod \"placement-db-create-crpfz\" (UID: \"c1269aae-c267-46d0-8adb-cdbe01430a31\") " pod="openstack/placement-db-create-crpfz" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.809919 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlcdm\" (UniqueName: \"kubernetes.io/projected/c1269aae-c267-46d0-8adb-cdbe01430a31-kube-api-access-hlcdm\") pod \"placement-db-create-crpfz\" (UID: \"c1269aae-c267-46d0-8adb-cdbe01430a31\") " pod="openstack/placement-db-create-crpfz" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.809960 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ad01c88-75f3-49ef-be2a-19ca711923bd-operator-scripts\") pod \"placement-a0f2-account-create-update-pmqhw\" (UID: \"8ad01c88-75f3-49ef-be2a-19ca711923bd\") " pod="openstack/placement-a0f2-account-create-update-pmqhw" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.810571 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1269aae-c267-46d0-8adb-cdbe01430a31-operator-scripts\") pod \"placement-db-create-crpfz\" (UID: \"c1269aae-c267-46d0-8adb-cdbe01430a31\") " pod="openstack/placement-db-create-crpfz" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.828971 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlcdm\" (UniqueName: \"kubernetes.io/projected/c1269aae-c267-46d0-8adb-cdbe01430a31-kube-api-access-hlcdm\") pod \"placement-db-create-crpfz\" (UID: \"c1269aae-c267-46d0-8adb-cdbe01430a31\") " pod="openstack/placement-db-create-crpfz" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.888222 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-f94a-account-create-update-vwrmd"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.889226 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f94a-account-create-update-vwrmd" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.892655 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.900531 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f94a-account-create-update-vwrmd"] Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.911602 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t775\" (UniqueName: \"kubernetes.io/projected/17190291-9e13-48b3-aff4-fc7e84e966bd-kube-api-access-6t775\") pod \"glance-db-create-c64lq\" (UID: \"17190291-9e13-48b3-aff4-fc7e84e966bd\") " pod="openstack/glance-db-create-c64lq" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.911747 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7zgd\" (UniqueName: \"kubernetes.io/projected/8ad01c88-75f3-49ef-be2a-19ca711923bd-kube-api-access-h7zgd\") pod \"placement-a0f2-account-create-update-pmqhw\" (UID: \"8ad01c88-75f3-49ef-be2a-19ca711923bd\") " pod="openstack/placement-a0f2-account-create-update-pmqhw" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.911795 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17190291-9e13-48b3-aff4-fc7e84e966bd-operator-scripts\") pod \"glance-db-create-c64lq\" (UID: \"17190291-9e13-48b3-aff4-fc7e84e966bd\") " pod="openstack/glance-db-create-c64lq" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.911813 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ad01c88-75f3-49ef-be2a-19ca711923bd-operator-scripts\") pod \"placement-a0f2-account-create-update-pmqhw\" (UID: \"8ad01c88-75f3-49ef-be2a-19ca711923bd\") " pod="openstack/placement-a0f2-account-create-update-pmqhw" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.912466 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ad01c88-75f3-49ef-be2a-19ca711923bd-operator-scripts\") pod \"placement-a0f2-account-create-update-pmqhw\" (UID: \"8ad01c88-75f3-49ef-be2a-19ca711923bd\") " pod="openstack/placement-a0f2-account-create-update-pmqhw" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.930110 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7zgd\" (UniqueName: \"kubernetes.io/projected/8ad01c88-75f3-49ef-be2a-19ca711923bd-kube-api-access-h7zgd\") pod \"placement-a0f2-account-create-update-pmqhw\" (UID: \"8ad01c88-75f3-49ef-be2a-19ca711923bd\") " pod="openstack/placement-a0f2-account-create-update-pmqhw" Feb 02 12:28:33 crc kubenswrapper[4846]: I0202 12:28:33.988100 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-crpfz" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.013489 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs7hr\" (UniqueName: \"kubernetes.io/projected/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-kube-api-access-gs7hr\") pod \"glance-f94a-account-create-update-vwrmd\" (UID: \"93eff246-ccb3-472c-bf51-f8ca43a5c4a2\") " pod="openstack/glance-f94a-account-create-update-vwrmd" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.013847 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t775\" (UniqueName: \"kubernetes.io/projected/17190291-9e13-48b3-aff4-fc7e84e966bd-kube-api-access-6t775\") pod \"glance-db-create-c64lq\" (UID: \"17190291-9e13-48b3-aff4-fc7e84e966bd\") " pod="openstack/glance-db-create-c64lq" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.013906 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-operator-scripts\") pod \"glance-f94a-account-create-update-vwrmd\" (UID: \"93eff246-ccb3-472c-bf51-f8ca43a5c4a2\") " pod="openstack/glance-f94a-account-create-update-vwrmd" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.014154 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17190291-9e13-48b3-aff4-fc7e84e966bd-operator-scripts\") pod \"glance-db-create-c64lq\" (UID: \"17190291-9e13-48b3-aff4-fc7e84e966bd\") " pod="openstack/glance-db-create-c64lq" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.016826 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17190291-9e13-48b3-aff4-fc7e84e966bd-operator-scripts\") pod \"glance-db-create-c64lq\" (UID: \"17190291-9e13-48b3-aff4-fc7e84e966bd\") " pod="openstack/glance-db-create-c64lq" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.031430 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t775\" (UniqueName: \"kubernetes.io/projected/17190291-9e13-48b3-aff4-fc7e84e966bd-kube-api-access-6t775\") pod \"glance-db-create-c64lq\" (UID: \"17190291-9e13-48b3-aff4-fc7e84e966bd\") " pod="openstack/glance-db-create-c64lq" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.043085 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0f2-account-create-update-pmqhw" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.105707 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c64lq" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.115474 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs7hr\" (UniqueName: \"kubernetes.io/projected/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-kube-api-access-gs7hr\") pod \"glance-f94a-account-create-update-vwrmd\" (UID: \"93eff246-ccb3-472c-bf51-f8ca43a5c4a2\") " pod="openstack/glance-f94a-account-create-update-vwrmd" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.115863 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-operator-scripts\") pod \"glance-f94a-account-create-update-vwrmd\" (UID: \"93eff246-ccb3-472c-bf51-f8ca43a5c4a2\") " pod="openstack/glance-f94a-account-create-update-vwrmd" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.117363 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-operator-scripts\") pod \"glance-f94a-account-create-update-vwrmd\" (UID: \"93eff246-ccb3-472c-bf51-f8ca43a5c4a2\") " pod="openstack/glance-f94a-account-create-update-vwrmd" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.129522 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xvdwm"] Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.135283 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs7hr\" (UniqueName: \"kubernetes.io/projected/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-kube-api-access-gs7hr\") pod \"glance-f94a-account-create-update-vwrmd\" (UID: \"93eff246-ccb3-472c-bf51-f8ca43a5c4a2\") " pod="openstack/glance-f94a-account-create-update-vwrmd" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.153397 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xvdwm" event={"ID":"534b394f-d62c-4c1d-8528-3a8aa37d5d7c","Type":"ContainerStarted","Data":"ebe6eb149b34169ef5de661e76d4415aea88e9a65060ff83f25f284f3669f92d"} Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.167214 4846 generic.go:334] "Generic (PLEG): container finished" podID="0cea47b6-6835-4c98-9006-c7d10e0e7457" containerID="9baac131245f62b73851f82758eac3beae324a624e97690d13a59f6f6d3e2bd8" exitCode=0 Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.168132 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-cpnvx" event={"ID":"0cea47b6-6835-4c98-9006-c7d10e0e7457","Type":"ContainerDied","Data":"9baac131245f62b73851f82758eac3beae324a624e97690d13a59f6f6d3e2bd8"} Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.209952 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f94a-account-create-update-vwrmd" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.326897 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b76-account-create-update-gm2qj"] Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.438827 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:34 crc kubenswrapper[4846]: E0202 12:28:34.439226 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 02 12:28:34 crc kubenswrapper[4846]: E0202 12:28:34.439304 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 02 12:28:34 crc kubenswrapper[4846]: E0202 12:28:34.439399 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift podName:309e8cc0-3b70-4961-b69c-1abf373f3c61 nodeName:}" failed. No retries permitted until 2026-02-02 12:28:42.439383869 +0000 UTC m=+1153.667970732 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift") pod "swift-storage-0" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61") : configmap "swift-ring-files" not found Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.486169 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-crpfz"] Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.738458 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cpnvx" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.812349 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a0f2-account-create-update-pmqhw"] Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.830220 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-c64lq"] Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.845235 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cea47b6-6835-4c98-9006-c7d10e0e7457-operator-scripts\") pod \"0cea47b6-6835-4c98-9006-c7d10e0e7457\" (UID: \"0cea47b6-6835-4c98-9006-c7d10e0e7457\") " Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.845351 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvjj2\" (UniqueName: \"kubernetes.io/projected/0cea47b6-6835-4c98-9006-c7d10e0e7457-kube-api-access-cvjj2\") pod \"0cea47b6-6835-4c98-9006-c7d10e0e7457\" (UID: \"0cea47b6-6835-4c98-9006-c7d10e0e7457\") " Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.848081 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cea47b6-6835-4c98-9006-c7d10e0e7457-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0cea47b6-6835-4c98-9006-c7d10e0e7457" (UID: "0cea47b6-6835-4c98-9006-c7d10e0e7457"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.857818 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cea47b6-6835-4c98-9006-c7d10e0e7457-kube-api-access-cvjj2" (OuterVolumeSpecName: "kube-api-access-cvjj2") pod "0cea47b6-6835-4c98-9006-c7d10e0e7457" (UID: "0cea47b6-6835-4c98-9006-c7d10e0e7457"). InnerVolumeSpecName "kube-api-access-cvjj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:34 crc kubenswrapper[4846]: W0202 12:28:34.873575 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17190291_9e13_48b3_aff4_fc7e84e966bd.slice/crio-9afe6e320c46d13f3428dff83befad3d73de5e6e1012ebae2ae3f97ae0ca108e WatchSource:0}: Error finding container 9afe6e320c46d13f3428dff83befad3d73de5e6e1012ebae2ae3f97ae0ca108e: Status 404 returned error can't find the container with id 9afe6e320c46d13f3428dff83befad3d73de5e6e1012ebae2ae3f97ae0ca108e Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.948005 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cea47b6-6835-4c98-9006-c7d10e0e7457-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.948037 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvjj2\" (UniqueName: \"kubernetes.io/projected/0cea47b6-6835-4c98-9006-c7d10e0e7457-kube-api-access-cvjj2\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:34 crc kubenswrapper[4846]: I0202 12:28:34.974849 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f94a-account-create-update-vwrmd"] Feb 02 12:28:35 crc kubenswrapper[4846]: W0202 12:28:35.067553 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93eff246_ccb3_472c_bf51_f8ca43a5c4a2.slice/crio-6634e2427ed21c9b564026d48faa3f358d37cc9a56efd549d0262b70838117fe WatchSource:0}: Error finding container 6634e2427ed21c9b564026d48faa3f358d37cc9a56efd549d0262b70838117fe: Status 404 returned error can't find the container with id 6634e2427ed21c9b564026d48faa3f358d37cc9a56efd549d0262b70838117fe Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.183201 4846 generic.go:334] "Generic (PLEG): container finished" podID="17190291-9e13-48b3-aff4-fc7e84e966bd" containerID="8317181145f1669cb74990cc6c775fa122e6ab9aacd49fce8490f28352e2c57f" exitCode=0 Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.183271 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c64lq" event={"ID":"17190291-9e13-48b3-aff4-fc7e84e966bd","Type":"ContainerDied","Data":"8317181145f1669cb74990cc6c775fa122e6ab9aacd49fce8490f28352e2c57f"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.183301 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c64lq" event={"ID":"17190291-9e13-48b3-aff4-fc7e84e966bd","Type":"ContainerStarted","Data":"9afe6e320c46d13f3428dff83befad3d73de5e6e1012ebae2ae3f97ae0ca108e"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.186123 4846 generic.go:334] "Generic (PLEG): container finished" podID="0d858780-98c0-4167-9bc5-8a4cc0e054da" containerID="aa70f3c1cb4d9ccf0a47e1cabbf6e0b6f5b91bf416c3d8de28edf8ad6a45b351" exitCode=0 Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.186172 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b76-account-create-update-gm2qj" event={"ID":"0d858780-98c0-4167-9bc5-8a4cc0e054da","Type":"ContainerDied","Data":"aa70f3c1cb4d9ccf0a47e1cabbf6e0b6f5b91bf416c3d8de28edf8ad6a45b351"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.186197 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b76-account-create-update-gm2qj" event={"ID":"0d858780-98c0-4167-9bc5-8a4cc0e054da","Type":"ContainerStarted","Data":"0356e8c8b7e555dae39fa7a1629f7eab1794c7de6964c0c6f81a57c40211f124"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.189813 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0f2-account-create-update-pmqhw" event={"ID":"8ad01c88-75f3-49ef-be2a-19ca711923bd","Type":"ContainerStarted","Data":"af62214010ef6c9ebee8a7b9733f31b2fde1d938f00d53e230f39c1f885fa12e"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.189882 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0f2-account-create-update-pmqhw" event={"ID":"8ad01c88-75f3-49ef-be2a-19ca711923bd","Type":"ContainerStarted","Data":"3199c9ff1f9cba4637cf4bd363ecf5e0ed47c3b61f29bc1de35c561c4f4e6060"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.193120 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f94a-account-create-update-vwrmd" event={"ID":"93eff246-ccb3-472c-bf51-f8ca43a5c4a2","Type":"ContainerStarted","Data":"6634e2427ed21c9b564026d48faa3f358d37cc9a56efd549d0262b70838117fe"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.211122 4846 generic.go:334] "Generic (PLEG): container finished" podID="534b394f-d62c-4c1d-8528-3a8aa37d5d7c" containerID="c9dc6241864be92bed2674514ffa4f13c2775741f26cfa057ce276cb569c8226" exitCode=0 Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.211223 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xvdwm" event={"ID":"534b394f-d62c-4c1d-8528-3a8aa37d5d7c","Type":"ContainerDied","Data":"c9dc6241864be92bed2674514ffa4f13c2775741f26cfa057ce276cb569c8226"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.214459 4846 generic.go:334] "Generic (PLEG): container finished" podID="c1269aae-c267-46d0-8adb-cdbe01430a31" containerID="6ed6bdf55d9e56b0bdeb412f1c47fec912bfa311c417c14a15454e3c3c8395eb" exitCode=0 Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.214595 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-crpfz" event={"ID":"c1269aae-c267-46d0-8adb-cdbe01430a31","Type":"ContainerDied","Data":"6ed6bdf55d9e56b0bdeb412f1c47fec912bfa311c417c14a15454e3c3c8395eb"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.214676 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-crpfz" event={"ID":"c1269aae-c267-46d0-8adb-cdbe01430a31","Type":"ContainerStarted","Data":"6f6746bfdc53403ad3f1934759bd5b3055ee3250227715a5d01583bdb7a55984"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.219530 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-cpnvx" Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.219643 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-cpnvx" event={"ID":"0cea47b6-6835-4c98-9006-c7d10e0e7457","Type":"ContainerDied","Data":"53128b23d4642df840c5e5050e30b515dce977829e94e30b456c93bfc3db8645"} Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.219692 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53128b23d4642df840c5e5050e30b515dce977829e94e30b456c93bfc3db8645" Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.227092 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-f94a-account-create-update-vwrmd" podStartSLOduration=2.22707434 podStartE2EDuration="2.22707434s" podCreationTimestamp="2026-02-02 12:28:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:28:35.226907746 +0000 UTC m=+1146.455494609" watchObservedRunningTime="2026-02-02 12:28:35.22707434 +0000 UTC m=+1146.455661203" Feb 02 12:28:35 crc kubenswrapper[4846]: I0202 12:28:35.250090 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-a0f2-account-create-update-pmqhw" podStartSLOduration=2.250069307 podStartE2EDuration="2.250069307s" podCreationTimestamp="2026-02-02 12:28:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:28:35.240597969 +0000 UTC m=+1146.469184832" watchObservedRunningTime="2026-02-02 12:28:35.250069307 +0000 UTC m=+1146.478656170" Feb 02 12:28:36 crc kubenswrapper[4846]: I0202 12:28:36.228387 4846 generic.go:334] "Generic (PLEG): container finished" podID="8ad01c88-75f3-49ef-be2a-19ca711923bd" containerID="af62214010ef6c9ebee8a7b9733f31b2fde1d938f00d53e230f39c1f885fa12e" exitCode=0 Feb 02 12:28:36 crc kubenswrapper[4846]: I0202 12:28:36.228478 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0f2-account-create-update-pmqhw" event={"ID":"8ad01c88-75f3-49ef-be2a-19ca711923bd","Type":"ContainerDied","Data":"af62214010ef6c9ebee8a7b9733f31b2fde1d938f00d53e230f39c1f885fa12e"} Feb 02 12:28:36 crc kubenswrapper[4846]: I0202 12:28:36.230557 4846 generic.go:334] "Generic (PLEG): container finished" podID="93eff246-ccb3-472c-bf51-f8ca43a5c4a2" containerID="980217af7f9420706d5f9938b02bc23ed28998d36f06b2cc585c6fa0c1914b8e" exitCode=0 Feb 02 12:28:36 crc kubenswrapper[4846]: I0202 12:28:36.230607 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f94a-account-create-update-vwrmd" event={"ID":"93eff246-ccb3-472c-bf51-f8ca43a5c4a2","Type":"ContainerDied","Data":"980217af7f9420706d5f9938b02bc23ed28998d36f06b2cc585c6fa0c1914b8e"} Feb 02 12:28:37 crc kubenswrapper[4846]: E0202 12:28:37.667666 4846 kubelet_node_status.go:756] "Failed to set some node status fields" err="failed to validate nodeIP: route ip+net: no such network interface" node="crc" Feb 02 12:28:37 crc kubenswrapper[4846]: I0202 12:28:37.779819 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-crpfz" Feb 02 12:28:37 crc kubenswrapper[4846]: I0202 12:28:37.932643 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlcdm\" (UniqueName: \"kubernetes.io/projected/c1269aae-c267-46d0-8adb-cdbe01430a31-kube-api-access-hlcdm\") pod \"c1269aae-c267-46d0-8adb-cdbe01430a31\" (UID: \"c1269aae-c267-46d0-8adb-cdbe01430a31\") " Feb 02 12:28:37 crc kubenswrapper[4846]: I0202 12:28:37.932807 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1269aae-c267-46d0-8adb-cdbe01430a31-operator-scripts\") pod \"c1269aae-c267-46d0-8adb-cdbe01430a31\" (UID: \"c1269aae-c267-46d0-8adb-cdbe01430a31\") " Feb 02 12:28:37 crc kubenswrapper[4846]: I0202 12:28:37.933515 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1269aae-c267-46d0-8adb-cdbe01430a31-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c1269aae-c267-46d0-8adb-cdbe01430a31" (UID: "c1269aae-c267-46d0-8adb-cdbe01430a31"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:37 crc kubenswrapper[4846]: I0202 12:28:37.939541 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1269aae-c267-46d0-8adb-cdbe01430a31-kube-api-access-hlcdm" (OuterVolumeSpecName: "kube-api-access-hlcdm") pod "c1269aae-c267-46d0-8adb-cdbe01430a31" (UID: "c1269aae-c267-46d0-8adb-cdbe01430a31"). InnerVolumeSpecName "kube-api-access-hlcdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.020061 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xvdwm" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.028186 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b76-account-create-update-gm2qj" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.029946 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c64lq" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.040540 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1269aae-c267-46d0-8adb-cdbe01430a31-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.040911 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlcdm\" (UniqueName: \"kubernetes.io/projected/c1269aae-c267-46d0-8adb-cdbe01430a31-kube-api-access-hlcdm\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.096171 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0f2-account-create-update-pmqhw" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.141578 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t775\" (UniqueName: \"kubernetes.io/projected/17190291-9e13-48b3-aff4-fc7e84e966bd-kube-api-access-6t775\") pod \"17190291-9e13-48b3-aff4-fc7e84e966bd\" (UID: \"17190291-9e13-48b3-aff4-fc7e84e966bd\") " Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.142328 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjbww\" (UniqueName: \"kubernetes.io/projected/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-kube-api-access-sjbww\") pod \"534b394f-d62c-4c1d-8528-3a8aa37d5d7c\" (UID: \"534b394f-d62c-4c1d-8528-3a8aa37d5d7c\") " Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.142393 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17190291-9e13-48b3-aff4-fc7e84e966bd-operator-scripts\") pod \"17190291-9e13-48b3-aff4-fc7e84e966bd\" (UID: \"17190291-9e13-48b3-aff4-fc7e84e966bd\") " Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.142443 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-operator-scripts\") pod \"534b394f-d62c-4c1d-8528-3a8aa37d5d7c\" (UID: \"534b394f-d62c-4c1d-8528-3a8aa37d5d7c\") " Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.142482 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xw72\" (UniqueName: \"kubernetes.io/projected/0d858780-98c0-4167-9bc5-8a4cc0e054da-kube-api-access-7xw72\") pod \"0d858780-98c0-4167-9bc5-8a4cc0e054da\" (UID: \"0d858780-98c0-4167-9bc5-8a4cc0e054da\") " Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.142522 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d858780-98c0-4167-9bc5-8a4cc0e054da-operator-scripts\") pod \"0d858780-98c0-4167-9bc5-8a4cc0e054da\" (UID: \"0d858780-98c0-4167-9bc5-8a4cc0e054da\") " Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.143315 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17190291-9e13-48b3-aff4-fc7e84e966bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "17190291-9e13-48b3-aff4-fc7e84e966bd" (UID: "17190291-9e13-48b3-aff4-fc7e84e966bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.143319 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "534b394f-d62c-4c1d-8528-3a8aa37d5d7c" (UID: "534b394f-d62c-4c1d-8528-3a8aa37d5d7c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.143827 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d858780-98c0-4167-9bc5-8a4cc0e054da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d858780-98c0-4167-9bc5-8a4cc0e054da" (UID: "0d858780-98c0-4167-9bc5-8a4cc0e054da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.146236 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d858780-98c0-4167-9bc5-8a4cc0e054da-kube-api-access-7xw72" (OuterVolumeSpecName: "kube-api-access-7xw72") pod "0d858780-98c0-4167-9bc5-8a4cc0e054da" (UID: "0d858780-98c0-4167-9bc5-8a4cc0e054da"). InnerVolumeSpecName "kube-api-access-7xw72". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.146433 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17190291-9e13-48b3-aff4-fc7e84e966bd-kube-api-access-6t775" (OuterVolumeSpecName: "kube-api-access-6t775") pod "17190291-9e13-48b3-aff4-fc7e84e966bd" (UID: "17190291-9e13-48b3-aff4-fc7e84e966bd"). InnerVolumeSpecName "kube-api-access-6t775". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.147597 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-kube-api-access-sjbww" (OuterVolumeSpecName: "kube-api-access-sjbww") pod "534b394f-d62c-4c1d-8528-3a8aa37d5d7c" (UID: "534b394f-d62c-4c1d-8528-3a8aa37d5d7c"). InnerVolumeSpecName "kube-api-access-sjbww". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.153729 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f94a-account-create-update-vwrmd" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.244080 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ad01c88-75f3-49ef-be2a-19ca711923bd-operator-scripts\") pod \"8ad01c88-75f3-49ef-be2a-19ca711923bd\" (UID: \"8ad01c88-75f3-49ef-be2a-19ca711923bd\") " Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.244195 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7zgd\" (UniqueName: \"kubernetes.io/projected/8ad01c88-75f3-49ef-be2a-19ca711923bd-kube-api-access-h7zgd\") pod \"8ad01c88-75f3-49ef-be2a-19ca711923bd\" (UID: \"8ad01c88-75f3-49ef-be2a-19ca711923bd\") " Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.244332 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-operator-scripts\") pod \"93eff246-ccb3-472c-bf51-f8ca43a5c4a2\" (UID: \"93eff246-ccb3-472c-bf51-f8ca43a5c4a2\") " Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.244487 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad01c88-75f3-49ef-be2a-19ca711923bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8ad01c88-75f3-49ef-be2a-19ca711923bd" (UID: "8ad01c88-75f3-49ef-be2a-19ca711923bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.244789 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "93eff246-ccb3-472c-bf51-f8ca43a5c4a2" (UID: "93eff246-ccb3-472c-bf51-f8ca43a5c4a2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.244853 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs7hr\" (UniqueName: \"kubernetes.io/projected/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-kube-api-access-gs7hr\") pod \"93eff246-ccb3-472c-bf51-f8ca43a5c4a2\" (UID: \"93eff246-ccb3-472c-bf51-f8ca43a5c4a2\") " Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.245497 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.245519 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t775\" (UniqueName: \"kubernetes.io/projected/17190291-9e13-48b3-aff4-fc7e84e966bd-kube-api-access-6t775\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.245529 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjbww\" (UniqueName: \"kubernetes.io/projected/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-kube-api-access-sjbww\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.245537 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17190291-9e13-48b3-aff4-fc7e84e966bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.245545 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ad01c88-75f3-49ef-be2a-19ca711923bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.245553 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/534b394f-d62c-4c1d-8528-3a8aa37d5d7c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.245561 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xw72\" (UniqueName: \"kubernetes.io/projected/0d858780-98c0-4167-9bc5-8a4cc0e054da-kube-api-access-7xw72\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.245569 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d858780-98c0-4167-9bc5-8a4cc0e054da-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.246987 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad01c88-75f3-49ef-be2a-19ca711923bd-kube-api-access-h7zgd" (OuterVolumeSpecName: "kube-api-access-h7zgd") pod "8ad01c88-75f3-49ef-be2a-19ca711923bd" (UID: "8ad01c88-75f3-49ef-be2a-19ca711923bd"). InnerVolumeSpecName "kube-api-access-h7zgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.247427 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-kube-api-access-gs7hr" (OuterVolumeSpecName: "kube-api-access-gs7hr") pod "93eff246-ccb3-472c-bf51-f8ca43a5c4a2" (UID: "93eff246-ccb3-472c-bf51-f8ca43a5c4a2"). InnerVolumeSpecName "kube-api-access-gs7hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.347388 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7zgd\" (UniqueName: \"kubernetes.io/projected/8ad01c88-75f3-49ef-be2a-19ca711923bd-kube-api-access-h7zgd\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.347459 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs7hr\" (UniqueName: \"kubernetes.io/projected/93eff246-ccb3-472c-bf51-f8ca43a5c4a2-kube-api-access-gs7hr\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.642572 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xvdwm" event={"ID":"534b394f-d62c-4c1d-8528-3a8aa37d5d7c","Type":"ContainerDied","Data":"ebe6eb149b34169ef5de661e76d4415aea88e9a65060ff83f25f284f3669f92d"} Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.642790 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebe6eb149b34169ef5de661e76d4415aea88e9a65060ff83f25f284f3669f92d" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.642895 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xvdwm" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.648768 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-crpfz" event={"ID":"c1269aae-c267-46d0-8adb-cdbe01430a31","Type":"ContainerDied","Data":"6f6746bfdc53403ad3f1934759bd5b3055ee3250227715a5d01583bdb7a55984"} Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.648810 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f6746bfdc53403ad3f1934759bd5b3055ee3250227715a5d01583bdb7a55984" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.648778 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-crpfz" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.651195 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-c64lq" event={"ID":"17190291-9e13-48b3-aff4-fc7e84e966bd","Type":"ContainerDied","Data":"9afe6e320c46d13f3428dff83befad3d73de5e6e1012ebae2ae3f97ae0ca108e"} Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.651227 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9afe6e320c46d13f3428dff83befad3d73de5e6e1012ebae2ae3f97ae0ca108e" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.651202 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-c64lq" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.653731 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b76-account-create-update-gm2qj" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.653731 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b76-account-create-update-gm2qj" event={"ID":"0d858780-98c0-4167-9bc5-8a4cc0e054da","Type":"ContainerDied","Data":"0356e8c8b7e555dae39fa7a1629f7eab1794c7de6964c0c6f81a57c40211f124"} Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.654097 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0356e8c8b7e555dae39fa7a1629f7eab1794c7de6964c0c6f81a57c40211f124" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.655744 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0f2-account-create-update-pmqhw" event={"ID":"8ad01c88-75f3-49ef-be2a-19ca711923bd","Type":"ContainerDied","Data":"3199c9ff1f9cba4637cf4bd363ecf5e0ed47c3b61f29bc1de35c561c4f4e6060"} Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.655839 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3199c9ff1f9cba4637cf4bd363ecf5e0ed47c3b61f29bc1de35c561c4f4e6060" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.655734 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0f2-account-create-update-pmqhw" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.657033 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f94a-account-create-update-vwrmd" event={"ID":"93eff246-ccb3-472c-bf51-f8ca43a5c4a2","Type":"ContainerDied","Data":"6634e2427ed21c9b564026d48faa3f358d37cc9a56efd549d0262b70838117fe"} Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.657062 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6634e2427ed21c9b564026d48faa3f358d37cc9a56efd549d0262b70838117fe" Feb 02 12:28:38 crc kubenswrapper[4846]: I0202 12:28:38.657105 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f94a-account-create-update-vwrmd" Feb 02 12:28:39 crc kubenswrapper[4846]: I0202 12:28:39.665748 4846 generic.go:334] "Generic (PLEG): container finished" podID="327c8143-b3ed-4431-a206-d2019ddcb606" containerID="f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb" exitCode=0 Feb 02 12:28:39 crc kubenswrapper[4846]: I0202 12:28:39.665845 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"327c8143-b3ed-4431-a206-d2019ddcb606","Type":"ContainerDied","Data":"f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb"} Feb 02 12:28:39 crc kubenswrapper[4846]: I0202 12:28:39.671101 4846 generic.go:334] "Generic (PLEG): container finished" podID="9cec353a-92f9-4ed4-8ec9-15271cbc6912" containerID="0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f" exitCode=0 Feb 02 12:28:39 crc kubenswrapper[4846]: I0202 12:28:39.671136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9cec353a-92f9-4ed4-8ec9-15271cbc6912","Type":"ContainerDied","Data":"0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f"} Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.685946 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9cec353a-92f9-4ed4-8ec9-15271cbc6912","Type":"ContainerStarted","Data":"488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2"} Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.686341 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.688921 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"327c8143-b3ed-4431-a206-d2019ddcb606","Type":"ContainerStarted","Data":"ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56"} Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.689220 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.690705 4846 generic.go:334] "Generic (PLEG): container finished" podID="ece3e0cc-5193-4c71-982b-84fe26ad56e1" containerID="7e20049cbc1c96a32d8a5a8a599eef2e575b73f6df35000995ce4b76e4f2a280" exitCode=0 Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.690743 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wj5sj" event={"ID":"ece3e0cc-5193-4c71-982b-84fe26ad56e1","Type":"ContainerDied","Data":"7e20049cbc1c96a32d8a5a8a599eef2e575b73f6df35000995ce4b76e4f2a280"} Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.731653 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.005927593 podStartE2EDuration="51.731636429s" podCreationTimestamp="2026-02-02 12:27:49 +0000 UTC" firstStartedPulling="2026-02-02 12:27:50.742363667 +0000 UTC m=+1101.970950530" lastFinishedPulling="2026-02-02 12:28:06.468072493 +0000 UTC m=+1117.696659366" observedRunningTime="2026-02-02 12:28:40.711246368 +0000 UTC m=+1151.939833241" watchObservedRunningTime="2026-02-02 12:28:40.731636429 +0000 UTC m=+1151.960223292" Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.747210 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=49.376514634 podStartE2EDuration="51.747186838s" podCreationTimestamp="2026-02-02 12:27:49 +0000 UTC" firstStartedPulling="2026-02-02 12:28:04.11141679 +0000 UTC m=+1115.340003653" lastFinishedPulling="2026-02-02 12:28:06.482088994 +0000 UTC m=+1117.710675857" observedRunningTime="2026-02-02 12:28:40.745029164 +0000 UTC m=+1151.973616027" watchObservedRunningTime="2026-02-02 12:28:40.747186838 +0000 UTC m=+1151.975773701" Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.750816 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.802402 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-f6ngv"] Feb 02 12:28:40 crc kubenswrapper[4846]: I0202 12:28:40.802764 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" podUID="2b7825cd-54bc-49e9-981f-9eb82b346bf9" containerName="dnsmasq-dns" containerID="cri-o://6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb" gracePeriod=10 Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.398953 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.501424 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-dns-svc\") pod \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.501846 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-ovsdbserver-nb\") pod \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.502122 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g56wq\" (UniqueName: \"kubernetes.io/projected/2b7825cd-54bc-49e9-981f-9eb82b346bf9-kube-api-access-g56wq\") pod \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.502234 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-config\") pod \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\" (UID: \"2b7825cd-54bc-49e9-981f-9eb82b346bf9\") " Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.510993 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b7825cd-54bc-49e9-981f-9eb82b346bf9-kube-api-access-g56wq" (OuterVolumeSpecName: "kube-api-access-g56wq") pod "2b7825cd-54bc-49e9-981f-9eb82b346bf9" (UID: "2b7825cd-54bc-49e9-981f-9eb82b346bf9"). InnerVolumeSpecName "kube-api-access-g56wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.536708 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2b7825cd-54bc-49e9-981f-9eb82b346bf9" (UID: "2b7825cd-54bc-49e9-981f-9eb82b346bf9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.540414 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2b7825cd-54bc-49e9-981f-9eb82b346bf9" (UID: "2b7825cd-54bc-49e9-981f-9eb82b346bf9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.549911 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-config" (OuterVolumeSpecName: "config") pod "2b7825cd-54bc-49e9-981f-9eb82b346bf9" (UID: "2b7825cd-54bc-49e9-981f-9eb82b346bf9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.603837 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g56wq\" (UniqueName: \"kubernetes.io/projected/2b7825cd-54bc-49e9-981f-9eb82b346bf9-kube-api-access-g56wq\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.603868 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.603877 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.603886 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b7825cd-54bc-49e9-981f-9eb82b346bf9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.698751 4846 generic.go:334] "Generic (PLEG): container finished" podID="2b7825cd-54bc-49e9-981f-9eb82b346bf9" containerID="6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb" exitCode=0 Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.698817 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" event={"ID":"2b7825cd-54bc-49e9-981f-9eb82b346bf9","Type":"ContainerDied","Data":"6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb"} Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.698872 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" event={"ID":"2b7825cd-54bc-49e9-981f-9eb82b346bf9","Type":"ContainerDied","Data":"a7d3649122de3c2178921734275209ec73e1122a15b8302b727188a0d5f48a61"} Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.698899 4846 scope.go:117] "RemoveContainer" containerID="6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.699083 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f7f48db9-f6ngv" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.726219 4846 scope.go:117] "RemoveContainer" containerID="8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.728353 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-f6ngv"] Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.738935 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64f7f48db9-f6ngv"] Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.749383 4846 scope.go:117] "RemoveContainer" containerID="6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb" Feb 02 12:28:41 crc kubenswrapper[4846]: E0202 12:28:41.749868 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb\": container with ID starting with 6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb not found: ID does not exist" containerID="6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.749923 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb"} err="failed to get container status \"6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb\": rpc error: code = NotFound desc = could not find container \"6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb\": container with ID starting with 6f87284dd34f142ac24851c2e3c1fa7d4f39b81dea5e774d7b9ce25a377d0efb not found: ID does not exist" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.749956 4846 scope.go:117] "RemoveContainer" containerID="8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2" Feb 02 12:28:41 crc kubenswrapper[4846]: E0202 12:28:41.750601 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2\": container with ID starting with 8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2 not found: ID does not exist" containerID="8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.750650 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2"} err="failed to get container status \"8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2\": rpc error: code = NotFound desc = could not find container \"8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2\": container with ID starting with 8afd2399e66147b4790730baf2d78749c400fe05106233b832f7ecc39b28abd2 not found: ID does not exist" Feb 02 12:28:41 crc kubenswrapper[4846]: I0202 12:28:41.854188 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.071367 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-cpnvx"] Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.083603 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-cpnvx"] Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.092991 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.152646 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-c26kj"] Feb 02 12:28:42 crc kubenswrapper[4846]: E0202 12:28:42.152944 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cea47b6-6835-4c98-9006-c7d10e0e7457" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.152959 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cea47b6-6835-4c98-9006-c7d10e0e7457" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: E0202 12:28:42.152973 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93eff246-ccb3-472c-bf51-f8ca43a5c4a2" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.152980 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93eff246-ccb3-472c-bf51-f8ca43a5c4a2" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: E0202 12:28:42.152990 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1269aae-c267-46d0-8adb-cdbe01430a31" containerName="mariadb-database-create" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.152997 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1269aae-c267-46d0-8adb-cdbe01430a31" containerName="mariadb-database-create" Feb 02 12:28:42 crc kubenswrapper[4846]: E0202 12:28:42.153008 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad01c88-75f3-49ef-be2a-19ca711923bd" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153015 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad01c88-75f3-49ef-be2a-19ca711923bd" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: E0202 12:28:42.153029 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534b394f-d62c-4c1d-8528-3a8aa37d5d7c" containerName="mariadb-database-create" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153034 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="534b394f-d62c-4c1d-8528-3a8aa37d5d7c" containerName="mariadb-database-create" Feb 02 12:28:42 crc kubenswrapper[4846]: E0202 12:28:42.153045 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7825cd-54bc-49e9-981f-9eb82b346bf9" containerName="init" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153050 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7825cd-54bc-49e9-981f-9eb82b346bf9" containerName="init" Feb 02 12:28:42 crc kubenswrapper[4846]: E0202 12:28:42.153059 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17190291-9e13-48b3-aff4-fc7e84e966bd" containerName="mariadb-database-create" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153066 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="17190291-9e13-48b3-aff4-fc7e84e966bd" containerName="mariadb-database-create" Feb 02 12:28:42 crc kubenswrapper[4846]: E0202 12:28:42.153077 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece3e0cc-5193-4c71-982b-84fe26ad56e1" containerName="swift-ring-rebalance" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153083 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece3e0cc-5193-4c71-982b-84fe26ad56e1" containerName="swift-ring-rebalance" Feb 02 12:28:42 crc kubenswrapper[4846]: E0202 12:28:42.153094 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d858780-98c0-4167-9bc5-8a4cc0e054da" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153100 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d858780-98c0-4167-9bc5-8a4cc0e054da" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: E0202 12:28:42.153113 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b7825cd-54bc-49e9-981f-9eb82b346bf9" containerName="dnsmasq-dns" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153119 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b7825cd-54bc-49e9-981f-9eb82b346bf9" containerName="dnsmasq-dns" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153255 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d858780-98c0-4167-9bc5-8a4cc0e054da" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153272 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b7825cd-54bc-49e9-981f-9eb82b346bf9" containerName="dnsmasq-dns" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153281 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ece3e0cc-5193-4c71-982b-84fe26ad56e1" containerName="swift-ring-rebalance" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153289 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="534b394f-d62c-4c1d-8528-3a8aa37d5d7c" containerName="mariadb-database-create" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153300 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="17190291-9e13-48b3-aff4-fc7e84e966bd" containerName="mariadb-database-create" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153308 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1269aae-c267-46d0-8adb-cdbe01430a31" containerName="mariadb-database-create" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153319 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93eff246-ccb3-472c-bf51-f8ca43a5c4a2" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153329 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cea47b6-6835-4c98-9006-c7d10e0e7457" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153337 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad01c88-75f3-49ef-be2a-19ca711923bd" containerName="mariadb-account-create-update" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.153888 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c26kj" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.156047 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.186400 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-c26kj"] Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.214074 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-swiftconf\") pod \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.214176 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-combined-ca-bundle\") pod \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.214207 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-scripts\") pod \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.214237 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qrkt\" (UniqueName: \"kubernetes.io/projected/ece3e0cc-5193-4c71-982b-84fe26ad56e1-kube-api-access-9qrkt\") pod \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.214286 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-ring-data-devices\") pod \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.214321 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ece3e0cc-5193-4c71-982b-84fe26ad56e1-etc-swift\") pod \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.214410 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-dispersionconf\") pod \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\" (UID: \"ece3e0cc-5193-4c71-982b-84fe26ad56e1\") " Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.221768 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ece3e0cc-5193-4c71-982b-84fe26ad56e1-kube-api-access-9qrkt" (OuterVolumeSpecName: "kube-api-access-9qrkt") pod "ece3e0cc-5193-4c71-982b-84fe26ad56e1" (UID: "ece3e0cc-5193-4c71-982b-84fe26ad56e1"). InnerVolumeSpecName "kube-api-access-9qrkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.222487 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-operator-scripts\") pod \"root-account-create-update-c26kj\" (UID: \"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f\") " pod="openstack/root-account-create-update-c26kj" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.222704 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svm56\" (UniqueName: \"kubernetes.io/projected/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-kube-api-access-svm56\") pod \"root-account-create-update-c26kj\" (UID: \"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f\") " pod="openstack/root-account-create-update-c26kj" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.222862 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qrkt\" (UniqueName: \"kubernetes.io/projected/ece3e0cc-5193-4c71-982b-84fe26ad56e1-kube-api-access-9qrkt\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.223560 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ece3e0cc-5193-4c71-982b-84fe26ad56e1" (UID: "ece3e0cc-5193-4c71-982b-84fe26ad56e1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.224038 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ece3e0cc-5193-4c71-982b-84fe26ad56e1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ece3e0cc-5193-4c71-982b-84fe26ad56e1" (UID: "ece3e0cc-5193-4c71-982b-84fe26ad56e1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.247826 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ece3e0cc-5193-4c71-982b-84fe26ad56e1" (UID: "ece3e0cc-5193-4c71-982b-84fe26ad56e1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.247851 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ece3e0cc-5193-4c71-982b-84fe26ad56e1" (UID: "ece3e0cc-5193-4c71-982b-84fe26ad56e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.250615 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ece3e0cc-5193-4c71-982b-84fe26ad56e1" (UID: "ece3e0cc-5193-4c71-982b-84fe26ad56e1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.265290 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-scripts" (OuterVolumeSpecName: "scripts") pod "ece3e0cc-5193-4c71-982b-84fe26ad56e1" (UID: "ece3e0cc-5193-4c71-982b-84fe26ad56e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.323864 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svm56\" (UniqueName: \"kubernetes.io/projected/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-kube-api-access-svm56\") pod \"root-account-create-update-c26kj\" (UID: \"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f\") " pod="openstack/root-account-create-update-c26kj" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.324029 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-operator-scripts\") pod \"root-account-create-update-c26kj\" (UID: \"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f\") " pod="openstack/root-account-create-update-c26kj" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.324133 4846 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.324149 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.324164 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.324175 4846 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ece3e0cc-5193-4c71-982b-84fe26ad56e1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.324185 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ece3e0cc-5193-4c71-982b-84fe26ad56e1-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.324196 4846 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ece3e0cc-5193-4c71-982b-84fe26ad56e1-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.325270 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-operator-scripts\") pod \"root-account-create-update-c26kj\" (UID: \"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f\") " pod="openstack/root-account-create-update-c26kj" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.346727 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svm56\" (UniqueName: \"kubernetes.io/projected/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-kube-api-access-svm56\") pod \"root-account-create-update-c26kj\" (UID: \"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f\") " pod="openstack/root-account-create-update-c26kj" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.480696 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c26kj" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.526578 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.532016 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift\") pod \"swift-storage-0\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " pod="openstack/swift-storage-0" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.713868 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wj5sj" event={"ID":"ece3e0cc-5193-4c71-982b-84fe26ad56e1","Type":"ContainerDied","Data":"10dae633a3b235287ec29708ef6c3fd4a234cd1f0512003b0227bae4f6e3ef2d"} Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.713919 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10dae633a3b235287ec29708ef6c3fd4a234cd1f0512003b0227bae4f6e3ef2d" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.714001 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wj5sj" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.743871 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 02 12:28:42 crc kubenswrapper[4846]: I0202 12:28:42.933422 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-c26kj"] Feb 02 12:28:43 crc kubenswrapper[4846]: I0202 12:28:43.249843 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 02 12:28:43 crc kubenswrapper[4846]: W0202 12:28:43.258931 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod309e8cc0_3b70_4961_b69c_1abf373f3c61.slice/crio-a3a0e95c20fcca3b1fd57686cceb9b66f09a011e6eaf01e68d9543c52646d542 WatchSource:0}: Error finding container a3a0e95c20fcca3b1fd57686cceb9b66f09a011e6eaf01e68d9543c52646d542: Status 404 returned error can't find the container with id a3a0e95c20fcca3b1fd57686cceb9b66f09a011e6eaf01e68d9543c52646d542 Feb 02 12:28:43 crc kubenswrapper[4846]: I0202 12:28:43.434676 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cea47b6-6835-4c98-9006-c7d10e0e7457" path="/var/lib/kubelet/pods/0cea47b6-6835-4c98-9006-c7d10e0e7457/volumes" Feb 02 12:28:43 crc kubenswrapper[4846]: I0202 12:28:43.435729 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b7825cd-54bc-49e9-981f-9eb82b346bf9" path="/var/lib/kubelet/pods/2b7825cd-54bc-49e9-981f-9eb82b346bf9/volumes" Feb 02 12:28:43 crc kubenswrapper[4846]: I0202 12:28:43.721747 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"a3a0e95c20fcca3b1fd57686cceb9b66f09a011e6eaf01e68d9543c52646d542"} Feb 02 12:28:43 crc kubenswrapper[4846]: I0202 12:28:43.722909 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f" containerID="0c052e8fc632fa57acc71fa4082dddaf8f00051c9a1fdbd9fa203c55647113eb" exitCode=0 Feb 02 12:28:43 crc kubenswrapper[4846]: I0202 12:28:43.722951 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c26kj" event={"ID":"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f","Type":"ContainerDied","Data":"0c052e8fc632fa57acc71fa4082dddaf8f00051c9a1fdbd9fa203c55647113eb"} Feb 02 12:28:43 crc kubenswrapper[4846]: I0202 12:28:43.722977 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c26kj" event={"ID":"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f","Type":"ContainerStarted","Data":"821013087cf191dc584a80435e988597a46edf96a6e03362ce4461e45334d5ee"} Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.103120 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-lwh8p"] Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.110678 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.112610 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.112610 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-c2sk4" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.131450 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-lwh8p"] Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.258111 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-db-sync-config-data\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.258503 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-config-data\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.258543 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-combined-ca-bundle\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.258610 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hckkz\" (UniqueName: \"kubernetes.io/projected/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-kube-api-access-hckkz\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.360013 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-db-sync-config-data\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.360091 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-config-data\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.360127 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-combined-ca-bundle\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.360184 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hckkz\" (UniqueName: \"kubernetes.io/projected/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-kube-api-access-hckkz\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.370456 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-db-sync-config-data\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.370505 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-combined-ca-bundle\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.372162 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-config-data\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.381647 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hckkz\" (UniqueName: \"kubernetes.io/projected/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-kube-api-access-hckkz\") pod \"glance-db-sync-lwh8p\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.443418 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lwh8p" Feb 02 12:28:44 crc kubenswrapper[4846]: I0202 12:28:44.746030 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d"} Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.047295 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c26kj" Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.104978 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-lwh8p"] Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.176016 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svm56\" (UniqueName: \"kubernetes.io/projected/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-kube-api-access-svm56\") pod \"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f\" (UID: \"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f\") " Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.176070 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-operator-scripts\") pod \"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f\" (UID: \"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f\") " Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.177464 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f" (UID: "ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.184821 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-kube-api-access-svm56" (OuterVolumeSpecName: "kube-api-access-svm56") pod "ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f" (UID: "ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f"). InnerVolumeSpecName "kube-api-access-svm56". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.278229 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svm56\" (UniqueName: \"kubernetes.io/projected/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-kube-api-access-svm56\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.278271 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.754504 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lwh8p" event={"ID":"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757","Type":"ContainerStarted","Data":"063dee3c9dc61f2259f99b3fecc01cf936ec4c545c8ecaf0ab9e93128aca1780"} Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.755936 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c26kj" event={"ID":"ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f","Type":"ContainerDied","Data":"821013087cf191dc584a80435e988597a46edf96a6e03362ce4461e45334d5ee"} Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.755973 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="821013087cf191dc584a80435e988597a46edf96a6e03362ce4461e45334d5ee" Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.756024 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c26kj" Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.759376 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3"} Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.759408 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864"} Feb 02 12:28:45 crc kubenswrapper[4846]: I0202 12:28:45.759419 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f"} Feb 02 12:28:46 crc kubenswrapper[4846]: I0202 12:28:46.773439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445"} Feb 02 12:28:46 crc kubenswrapper[4846]: I0202 12:28:46.773853 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728"} Feb 02 12:28:46 crc kubenswrapper[4846]: I0202 12:28:46.773870 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d"} Feb 02 12:28:47 crc kubenswrapper[4846]: I0202 12:28:47.789751 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7"} Feb 02 12:28:48 crc kubenswrapper[4846]: I0202 12:28:48.802741 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558"} Feb 02 12:28:48 crc kubenswrapper[4846]: I0202 12:28:48.803010 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891"} Feb 02 12:28:48 crc kubenswrapper[4846]: I0202 12:28:48.803019 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c"} Feb 02 12:28:48 crc kubenswrapper[4846]: I0202 12:28:48.803027 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3"} Feb 02 12:28:49 crc kubenswrapper[4846]: I0202 12:28:49.806752 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-b2hcc" podUID="bda23288-2fee-45ad-81e2-a14654c16589" containerName="ovn-controller" probeResult="failure" output=< Feb 02 12:28:49 crc kubenswrapper[4846]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 02 12:28:49 crc kubenswrapper[4846]: > Feb 02 12:28:49 crc kubenswrapper[4846]: I0202 12:28:49.816991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455"} Feb 02 12:28:49 crc kubenswrapper[4846]: I0202 12:28:49.817030 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064"} Feb 02 12:28:49 crc kubenswrapper[4846]: I0202 12:28:49.817040 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerStarted","Data":"8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae"} Feb 02 12:28:49 crc kubenswrapper[4846]: I0202 12:28:49.864548 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.217300513 podStartE2EDuration="24.864520835s" podCreationTimestamp="2026-02-02 12:28:25 +0000 UTC" firstStartedPulling="2026-02-02 12:28:43.260952947 +0000 UTC m=+1154.489539810" lastFinishedPulling="2026-02-02 12:28:47.908173269 +0000 UTC m=+1159.136760132" observedRunningTime="2026-02-02 12:28:49.856948315 +0000 UTC m=+1161.085535178" watchObservedRunningTime="2026-02-02 12:28:49.864520835 +0000 UTC m=+1161.093107698" Feb 02 12:28:49 crc kubenswrapper[4846]: I0202 12:28:49.867992 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:28:49 crc kubenswrapper[4846]: I0202 12:28:49.874092 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.167935 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lw7hb"] Feb 02 12:28:50 crc kubenswrapper[4846]: E0202 12:28:50.168555 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f" containerName="mariadb-account-create-update" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.168572 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f" containerName="mariadb-account-create-update" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.168765 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f" containerName="mariadb-account-create-update" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.170106 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.174847 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.187379 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lw7hb"] Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.244606 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-b2hcc-config-9f76v"] Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.245546 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.249402 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.255038 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b2hcc-config-9f76v"] Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.260720 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfkq6\" (UniqueName: \"kubernetes.io/projected/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-kube-api-access-sfkq6\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.260774 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-svc\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.260806 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-sb\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.260850 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-config\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.260890 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-swift-storage-0\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.260909 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-nb\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.362682 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run-ovn\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.362727 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-additional-scripts\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.362768 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-config\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.362827 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-scripts\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.362866 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-swift-storage-0\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.362896 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-nb\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.362920 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.362948 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-log-ovn\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.363047 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfkq6\" (UniqueName: \"kubernetes.io/projected/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-kube-api-access-sfkq6\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.363079 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-svc\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.363119 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-sb\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.363165 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgpch\" (UniqueName: \"kubernetes.io/projected/a00f0a2e-5571-4d2b-9be9-700e7186b93a-kube-api-access-tgpch\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.364355 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-config\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.364366 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-nb\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.364589 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-swift-storage-0\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.364659 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-sb\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.367285 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-svc\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.381956 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfkq6\" (UniqueName: \"kubernetes.io/projected/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-kube-api-access-sfkq6\") pod \"dnsmasq-dns-5799dddbf-lw7hb\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.393991 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.468010 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgpch\" (UniqueName: \"kubernetes.io/projected/a00f0a2e-5571-4d2b-9be9-700e7186b93a-kube-api-access-tgpch\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.468064 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run-ovn\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.468083 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-additional-scripts\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.468129 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-scripts\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.468170 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.468199 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-log-ovn\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.468467 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-log-ovn\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.471079 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run-ovn\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.472490 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-additional-scripts\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.475188 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-scripts\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.475340 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.494507 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.495047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgpch\" (UniqueName: \"kubernetes.io/projected/a00f0a2e-5571-4d2b-9be9-700e7186b93a-kube-api-access-tgpch\") pod \"ovn-controller-b2hcc-config-9f76v\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.617017 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.656699 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-tb9t6"] Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.662140 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tb9t6" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.680871 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tb9t6"] Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.754952 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-nm26b"] Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.756497 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nm26b" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.771690 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nm26b"] Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.777898 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbe30827-4587-4d5a-a651-df65604399ee-operator-scripts\") pod \"barbican-db-create-nm26b\" (UID: \"bbe30827-4587-4d5a-a651-df65604399ee\") " pod="openstack/barbican-db-create-nm26b" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.777962 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnvh2\" (UniqueName: \"kubernetes.io/projected/bbe30827-4587-4d5a-a651-df65604399ee-kube-api-access-nnvh2\") pod \"barbican-db-create-nm26b\" (UID: \"bbe30827-4587-4d5a-a651-df65604399ee\") " pod="openstack/barbican-db-create-nm26b" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.778019 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjhvt\" (UniqueName: \"kubernetes.io/projected/d670b5c2-8eea-4413-90df-0f4b72d8acd7-kube-api-access-qjhvt\") pod \"cinder-db-create-tb9t6\" (UID: \"d670b5c2-8eea-4413-90df-0f4b72d8acd7\") " pod="openstack/cinder-db-create-tb9t6" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.778107 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d670b5c2-8eea-4413-90df-0f4b72d8acd7-operator-scripts\") pod \"cinder-db-create-tb9t6\" (UID: \"d670b5c2-8eea-4413-90df-0f4b72d8acd7\") " pod="openstack/cinder-db-create-tb9t6" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.804768 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.870242 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c2bc-account-create-update-2r5pb"] Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.871657 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c2bc-account-create-update-2r5pb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.876609 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.885788 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjhvt\" (UniqueName: \"kubernetes.io/projected/d670b5c2-8eea-4413-90df-0f4b72d8acd7-kube-api-access-qjhvt\") pod \"cinder-db-create-tb9t6\" (UID: \"d670b5c2-8eea-4413-90df-0f4b72d8acd7\") " pod="openstack/cinder-db-create-tb9t6" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.885946 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d670b5c2-8eea-4413-90df-0f4b72d8acd7-operator-scripts\") pod \"cinder-db-create-tb9t6\" (UID: \"d670b5c2-8eea-4413-90df-0f4b72d8acd7\") " pod="openstack/cinder-db-create-tb9t6" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.886068 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbe30827-4587-4d5a-a651-df65604399ee-operator-scripts\") pod \"barbican-db-create-nm26b\" (UID: \"bbe30827-4587-4d5a-a651-df65604399ee\") " pod="openstack/barbican-db-create-nm26b" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.886121 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnvh2\" (UniqueName: \"kubernetes.io/projected/bbe30827-4587-4d5a-a651-df65604399ee-kube-api-access-nnvh2\") pod \"barbican-db-create-nm26b\" (UID: \"bbe30827-4587-4d5a-a651-df65604399ee\") " pod="openstack/barbican-db-create-nm26b" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.888737 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbe30827-4587-4d5a-a651-df65604399ee-operator-scripts\") pod \"barbican-db-create-nm26b\" (UID: \"bbe30827-4587-4d5a-a651-df65604399ee\") " pod="openstack/barbican-db-create-nm26b" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.888823 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d670b5c2-8eea-4413-90df-0f4b72d8acd7-operator-scripts\") pod \"cinder-db-create-tb9t6\" (UID: \"d670b5c2-8eea-4413-90df-0f4b72d8acd7\") " pod="openstack/cinder-db-create-tb9t6" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.909563 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c2bc-account-create-update-2r5pb"] Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.941212 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjhvt\" (UniqueName: \"kubernetes.io/projected/d670b5c2-8eea-4413-90df-0f4b72d8acd7-kube-api-access-qjhvt\") pod \"cinder-db-create-tb9t6\" (UID: \"d670b5c2-8eea-4413-90df-0f4b72d8acd7\") " pod="openstack/cinder-db-create-tb9t6" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.942024 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnvh2\" (UniqueName: \"kubernetes.io/projected/bbe30827-4587-4d5a-a651-df65604399ee-kube-api-access-nnvh2\") pod \"barbican-db-create-nm26b\" (UID: \"bbe30827-4587-4d5a-a651-df65604399ee\") " pod="openstack/barbican-db-create-nm26b" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.986274 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tb9t6" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.987593 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9h82\" (UniqueName: \"kubernetes.io/projected/a488c453-b188-4d57-a136-0a002cd342bb-kube-api-access-q9h82\") pod \"barbican-c2bc-account-create-update-2r5pb\" (UID: \"a488c453-b188-4d57-a136-0a002cd342bb\") " pod="openstack/barbican-c2bc-account-create-update-2r5pb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.987700 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a488c453-b188-4d57-a136-0a002cd342bb-operator-scripts\") pod \"barbican-c2bc-account-create-update-2r5pb\" (UID: \"a488c453-b188-4d57-a136-0a002cd342bb\") " pod="openstack/barbican-c2bc-account-create-update-2r5pb" Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.992259 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-cd5d-account-create-update-7xmpx"] Feb 02 12:28:50 crc kubenswrapper[4846]: I0202 12:28:50.993369 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cd5d-account-create-update-7xmpx" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.007046 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.028148 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cd5d-account-create-update-7xmpx"] Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.074837 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nm26b" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.089393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9h82\" (UniqueName: \"kubernetes.io/projected/a488c453-b188-4d57-a136-0a002cd342bb-kube-api-access-q9h82\") pod \"barbican-c2bc-account-create-update-2r5pb\" (UID: \"a488c453-b188-4d57-a136-0a002cd342bb\") " pod="openstack/barbican-c2bc-account-create-update-2r5pb" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.089815 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a488c453-b188-4d57-a136-0a002cd342bb-operator-scripts\") pod \"barbican-c2bc-account-create-update-2r5pb\" (UID: \"a488c453-b188-4d57-a136-0a002cd342bb\") " pod="openstack/barbican-c2bc-account-create-update-2r5pb" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.090736 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a488c453-b188-4d57-a136-0a002cd342bb-operator-scripts\") pod \"barbican-c2bc-account-create-update-2r5pb\" (UID: \"a488c453-b188-4d57-a136-0a002cd342bb\") " pod="openstack/barbican-c2bc-account-create-update-2r5pb" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.111507 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9h82\" (UniqueName: \"kubernetes.io/projected/a488c453-b188-4d57-a136-0a002cd342bb-kube-api-access-q9h82\") pod \"barbican-c2bc-account-create-update-2r5pb\" (UID: \"a488c453-b188-4d57-a136-0a002cd342bb\") " pod="openstack/barbican-c2bc-account-create-update-2r5pb" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.158732 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-pnzdq"] Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.160555 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pnzdq" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.180441 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pnzdq"] Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.191161 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-operator-scripts\") pod \"neutron-db-create-pnzdq\" (UID: \"3dd98e97-bd56-4dde-8592-8836bbc2a0ee\") " pod="openstack/neutron-db-create-pnzdq" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.191218 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94a78e75-b5be-43f1-ac04-911885a75395-operator-scripts\") pod \"cinder-cd5d-account-create-update-7xmpx\" (UID: \"94a78e75-b5be-43f1-ac04-911885a75395\") " pod="openstack/cinder-cd5d-account-create-update-7xmpx" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.191243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkgvk\" (UniqueName: \"kubernetes.io/projected/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-kube-api-access-xkgvk\") pod \"neutron-db-create-pnzdq\" (UID: \"3dd98e97-bd56-4dde-8592-8836bbc2a0ee\") " pod="openstack/neutron-db-create-pnzdq" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.191683 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x84fh\" (UniqueName: \"kubernetes.io/projected/94a78e75-b5be-43f1-ac04-911885a75395-kube-api-access-x84fh\") pod \"cinder-cd5d-account-create-update-7xmpx\" (UID: \"94a78e75-b5be-43f1-ac04-911885a75395\") " pod="openstack/cinder-cd5d-account-create-update-7xmpx" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.195999 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c2bc-account-create-update-2r5pb" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.218569 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xf7xw"] Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.219749 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.224265 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9wp9t" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.224598 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.225466 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.225594 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.240027 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xf7xw"] Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.293337 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-config-data\") pod \"keystone-db-sync-xf7xw\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.293398 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-combined-ca-bundle\") pod \"keystone-db-sync-xf7xw\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.293469 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x84fh\" (UniqueName: \"kubernetes.io/projected/94a78e75-b5be-43f1-ac04-911885a75395-kube-api-access-x84fh\") pod \"cinder-cd5d-account-create-update-7xmpx\" (UID: \"94a78e75-b5be-43f1-ac04-911885a75395\") " pod="openstack/cinder-cd5d-account-create-update-7xmpx" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.293611 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-operator-scripts\") pod \"neutron-db-create-pnzdq\" (UID: \"3dd98e97-bd56-4dde-8592-8836bbc2a0ee\") " pod="openstack/neutron-db-create-pnzdq" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.293706 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djxp7\" (UniqueName: \"kubernetes.io/projected/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-kube-api-access-djxp7\") pod \"keystone-db-sync-xf7xw\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.293771 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94a78e75-b5be-43f1-ac04-911885a75395-operator-scripts\") pod \"cinder-cd5d-account-create-update-7xmpx\" (UID: \"94a78e75-b5be-43f1-ac04-911885a75395\") " pod="openstack/cinder-cd5d-account-create-update-7xmpx" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.293820 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkgvk\" (UniqueName: \"kubernetes.io/projected/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-kube-api-access-xkgvk\") pod \"neutron-db-create-pnzdq\" (UID: \"3dd98e97-bd56-4dde-8592-8836bbc2a0ee\") " pod="openstack/neutron-db-create-pnzdq" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.294642 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94a78e75-b5be-43f1-ac04-911885a75395-operator-scripts\") pod \"cinder-cd5d-account-create-update-7xmpx\" (UID: \"94a78e75-b5be-43f1-ac04-911885a75395\") " pod="openstack/cinder-cd5d-account-create-update-7xmpx" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.294721 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-operator-scripts\") pod \"neutron-db-create-pnzdq\" (UID: \"3dd98e97-bd56-4dde-8592-8836bbc2a0ee\") " pod="openstack/neutron-db-create-pnzdq" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.311743 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x84fh\" (UniqueName: \"kubernetes.io/projected/94a78e75-b5be-43f1-ac04-911885a75395-kube-api-access-x84fh\") pod \"cinder-cd5d-account-create-update-7xmpx\" (UID: \"94a78e75-b5be-43f1-ac04-911885a75395\") " pod="openstack/cinder-cd5d-account-create-update-7xmpx" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.312020 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkgvk\" (UniqueName: \"kubernetes.io/projected/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-kube-api-access-xkgvk\") pod \"neutron-db-create-pnzdq\" (UID: \"3dd98e97-bd56-4dde-8592-8836bbc2a0ee\") " pod="openstack/neutron-db-create-pnzdq" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.324849 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cd5d-account-create-update-7xmpx" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.360958 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-33ff-account-create-update-mpf9l"] Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.362208 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-33ff-account-create-update-mpf9l" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.364502 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.390966 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-33ff-account-create-update-mpf9l"] Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.395650 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djxp7\" (UniqueName: \"kubernetes.io/projected/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-kube-api-access-djxp7\") pod \"keystone-db-sync-xf7xw\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.395793 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-config-data\") pod \"keystone-db-sync-xf7xw\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.395822 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-combined-ca-bundle\") pod \"keystone-db-sync-xf7xw\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.399549 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-combined-ca-bundle\") pod \"keystone-db-sync-xf7xw\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.400917 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-config-data\") pod \"keystone-db-sync-xf7xw\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.413474 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djxp7\" (UniqueName: \"kubernetes.io/projected/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-kube-api-access-djxp7\") pod \"keystone-db-sync-xf7xw\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.485560 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pnzdq" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.497558 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-operator-scripts\") pod \"neutron-33ff-account-create-update-mpf9l\" (UID: \"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1\") " pod="openstack/neutron-33ff-account-create-update-mpf9l" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.497614 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gpnc\" (UniqueName: \"kubernetes.io/projected/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-kube-api-access-5gpnc\") pod \"neutron-33ff-account-create-update-mpf9l\" (UID: \"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1\") " pod="openstack/neutron-33ff-account-create-update-mpf9l" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.544276 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.599717 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-operator-scripts\") pod \"neutron-33ff-account-create-update-mpf9l\" (UID: \"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1\") " pod="openstack/neutron-33ff-account-create-update-mpf9l" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.599778 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gpnc\" (UniqueName: \"kubernetes.io/projected/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-kube-api-access-5gpnc\") pod \"neutron-33ff-account-create-update-mpf9l\" (UID: \"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1\") " pod="openstack/neutron-33ff-account-create-update-mpf9l" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.600829 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-operator-scripts\") pod \"neutron-33ff-account-create-update-mpf9l\" (UID: \"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1\") " pod="openstack/neutron-33ff-account-create-update-mpf9l" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.620209 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gpnc\" (UniqueName: \"kubernetes.io/projected/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-kube-api-access-5gpnc\") pod \"neutron-33ff-account-create-update-mpf9l\" (UID: \"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1\") " pod="openstack/neutron-33ff-account-create-update-mpf9l" Feb 02 12:28:51 crc kubenswrapper[4846]: I0202 12:28:51.692415 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-33ff-account-create-update-mpf9l" Feb 02 12:28:54 crc kubenswrapper[4846]: I0202 12:28:54.810470 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-b2hcc" podUID="bda23288-2fee-45ad-81e2-a14654c16589" containerName="ovn-controller" probeResult="failure" output=< Feb 02 12:28:54 crc kubenswrapper[4846]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 02 12:28:54 crc kubenswrapper[4846]: > Feb 02 12:28:59 crc kubenswrapper[4846]: I0202 12:28:59.815391 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-b2hcc" podUID="bda23288-2fee-45ad-81e2-a14654c16589" containerName="ovn-controller" probeResult="failure" output=< Feb 02 12:28:59 crc kubenswrapper[4846]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 02 12:28:59 crc kubenswrapper[4846]: > Feb 02 12:29:00 crc kubenswrapper[4846]: W0202 12:29:00.071252 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dd98e97_bd56_4dde_8592_8836bbc2a0ee.slice/crio-bf3cd3627f20fb5bce2625d78cb349835c021051fb6cb49e0ef354ae18ceb449 WatchSource:0}: Error finding container bf3cd3627f20fb5bce2625d78cb349835c021051fb6cb49e0ef354ae18ceb449: Status 404 returned error can't find the container with id bf3cd3627f20fb5bce2625d78cb349835c021051fb6cb49e0ef354ae18ceb449 Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.074640 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-33ff-account-create-update-mpf9l"] Feb 02 12:29:00 crc kubenswrapper[4846]: W0202 12:29:00.080592 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc39fb0e2_eb96_4f4e_a8d2_a268b0d8b4f1.slice/crio-a0e4edba90d667e4357289ccbefaae12e144f228081c9d9969b2f407c32b6a97 WatchSource:0}: Error finding container a0e4edba90d667e4357289ccbefaae12e144f228081c9d9969b2f407c32b6a97: Status 404 returned error can't find the container with id a0e4edba90d667e4357289ccbefaae12e144f228081c9d9969b2f407c32b6a97 Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.084490 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-tb9t6"] Feb 02 12:29:00 crc kubenswrapper[4846]: W0202 12:29:00.085442 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda00f0a2e_5571_4d2b_9be9_700e7186b93a.slice/crio-5254cbc22f86c6859ea483797c83efad8117822b9d8c00153bf37afd3953049c WatchSource:0}: Error finding container 5254cbc22f86c6859ea483797c83efad8117822b9d8c00153bf37afd3953049c: Status 404 returned error can't find the container with id 5254cbc22f86c6859ea483797c83efad8117822b9d8c00153bf37afd3953049c Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.100362 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b2hcc-config-9f76v"] Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.107402 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pnzdq"] Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.264611 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xf7xw"] Feb 02 12:29:00 crc kubenswrapper[4846]: W0202 12:29:00.280743 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5fa8c0f_b2ae_49c3_a2d6_6ff48830d757.slice/crio-cc2f9ab12743b74111196c4c26f621d8c66e5e26cdb67ab799e309f0433783cd WatchSource:0}: Error finding container cc2f9ab12743b74111196c4c26f621d8c66e5e26cdb67ab799e309f0433783cd: Status 404 returned error can't find the container with id cc2f9ab12743b74111196c4c26f621d8c66e5e26cdb67ab799e309f0433783cd Feb 02 12:29:00 crc kubenswrapper[4846]: W0202 12:29:00.287532 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99c7f7c1_1b76_4d26_84e7_a49eaa21a006.slice/crio-fc123523fe75b12a27a66f778e3b01d4407e7d406fd37186a25da89ff4aaa2f1 WatchSource:0}: Error finding container fc123523fe75b12a27a66f778e3b01d4407e7d406fd37186a25da89ff4aaa2f1: Status 404 returned error can't find the container with id fc123523fe75b12a27a66f778e3b01d4407e7d406fd37186a25da89ff4aaa2f1 Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.292151 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lw7hb"] Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.305468 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cd5d-account-create-update-7xmpx"] Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.320805 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c2bc-account-create-update-2r5pb"] Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.340721 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-nm26b"] Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.921902 4846 generic.go:334] "Generic (PLEG): container finished" podID="d670b5c2-8eea-4413-90df-0f4b72d8acd7" containerID="82c7e76bff98e984e34c1c47e1a89029be4cfa11944cf5e9fa3ccf1450ac3a6e" exitCode=0 Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.921995 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tb9t6" event={"ID":"d670b5c2-8eea-4413-90df-0f4b72d8acd7","Type":"ContainerDied","Data":"82c7e76bff98e984e34c1c47e1a89029be4cfa11944cf5e9fa3ccf1450ac3a6e"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.922290 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tb9t6" event={"ID":"d670b5c2-8eea-4413-90df-0f4b72d8acd7","Type":"ContainerStarted","Data":"cba43c57f0a7ec30b5e679b0287e921396a9ec483977f8bbe73ae7392d635e21"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.925747 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lwh8p" event={"ID":"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757","Type":"ContainerStarted","Data":"c62f51d03fd9f435dd3c6722b2d45f51aea94cdc0dc8bea514ed230dba6a1885"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.929662 4846 generic.go:334] "Generic (PLEG): container finished" podID="3dd98e97-bd56-4dde-8592-8836bbc2a0ee" containerID="64fad55f3bd27a34f1da83a5dc31008972283b2c847e98e400a85649b868215e" exitCode=0 Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.929766 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pnzdq" event={"ID":"3dd98e97-bd56-4dde-8592-8836bbc2a0ee","Type":"ContainerDied","Data":"64fad55f3bd27a34f1da83a5dc31008972283b2c847e98e400a85649b868215e"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.929796 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pnzdq" event={"ID":"3dd98e97-bd56-4dde-8592-8836bbc2a0ee","Type":"ContainerStarted","Data":"bf3cd3627f20fb5bce2625d78cb349835c021051fb6cb49e0ef354ae18ceb449"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.931232 4846 generic.go:334] "Generic (PLEG): container finished" podID="94a78e75-b5be-43f1-ac04-911885a75395" containerID="0badaa9fb9c046619f3874d44014b8116430f9025733765fa1163fcae1ed3d30" exitCode=0 Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.931278 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cd5d-account-create-update-7xmpx" event={"ID":"94a78e75-b5be-43f1-ac04-911885a75395","Type":"ContainerDied","Data":"0badaa9fb9c046619f3874d44014b8116430f9025733765fa1163fcae1ed3d30"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.931298 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cd5d-account-create-update-7xmpx" event={"ID":"94a78e75-b5be-43f1-ac04-911885a75395","Type":"ContainerStarted","Data":"1864a8c9fced459a4f013e854d65a92c0c5e04ca483d785f69cc236e32a88c19"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.932535 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xf7xw" event={"ID":"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757","Type":"ContainerStarted","Data":"cc2f9ab12743b74111196c4c26f621d8c66e5e26cdb67ab799e309f0433783cd"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.933916 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc-config-9f76v" event={"ID":"a00f0a2e-5571-4d2b-9be9-700e7186b93a","Type":"ContainerStarted","Data":"6dbe9840669d94a9aa14d90cfc98a8e3f74cc3f30a489967e47e07ffa42a5e82"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.933937 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc-config-9f76v" event={"ID":"a00f0a2e-5571-4d2b-9be9-700e7186b93a","Type":"ContainerStarted","Data":"5254cbc22f86c6859ea483797c83efad8117822b9d8c00153bf37afd3953049c"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.941958 4846 generic.go:334] "Generic (PLEG): container finished" podID="bbe30827-4587-4d5a-a651-df65604399ee" containerID="1aca36bbf97855cd72821d99a63c0d8ff444d4e356394fb4c879014ec3b75fd5" exitCode=0 Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.942063 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nm26b" event={"ID":"bbe30827-4587-4d5a-a651-df65604399ee","Type":"ContainerDied","Data":"1aca36bbf97855cd72821d99a63c0d8ff444d4e356394fb4c879014ec3b75fd5"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.942093 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nm26b" event={"ID":"bbe30827-4587-4d5a-a651-df65604399ee","Type":"ContainerStarted","Data":"f343503c6fece3528c8854591dea29eb94684b06326c690d79eced45bbeea3e3"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.955507 4846 generic.go:334] "Generic (PLEG): container finished" podID="99c7f7c1-1b76-4d26-84e7-a49eaa21a006" containerID="3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748" exitCode=0 Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.955647 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" event={"ID":"99c7f7c1-1b76-4d26-84e7-a49eaa21a006","Type":"ContainerDied","Data":"3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.955700 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" event={"ID":"99c7f7c1-1b76-4d26-84e7-a49eaa21a006","Type":"ContainerStarted","Data":"fc123523fe75b12a27a66f778e3b01d4407e7d406fd37186a25da89ff4aaa2f1"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.957806 4846 generic.go:334] "Generic (PLEG): container finished" podID="a488c453-b188-4d57-a136-0a002cd342bb" containerID="6131396e4bf56b3a277f471ab103da5529eb4c3c786f308bb7fbd7abf06db02d" exitCode=0 Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.957867 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c2bc-account-create-update-2r5pb" event={"ID":"a488c453-b188-4d57-a136-0a002cd342bb","Type":"ContainerDied","Data":"6131396e4bf56b3a277f471ab103da5529eb4c3c786f308bb7fbd7abf06db02d"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.957887 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c2bc-account-create-update-2r5pb" event={"ID":"a488c453-b188-4d57-a136-0a002cd342bb","Type":"ContainerStarted","Data":"5faeb8b832b4c5792683dc0f84c3887f3849f6e64cdf573933662409e9bd39c9"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.960917 4846 generic.go:334] "Generic (PLEG): container finished" podID="c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1" containerID="8774cd543521690df2b64182c81e025daf2b49960381c918588bd9af86e76a03" exitCode=0 Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.961049 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-33ff-account-create-update-mpf9l" event={"ID":"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1","Type":"ContainerDied","Data":"8774cd543521690df2b64182c81e025daf2b49960381c918588bd9af86e76a03"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.961099 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-33ff-account-create-update-mpf9l" event={"ID":"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1","Type":"ContainerStarted","Data":"a0e4edba90d667e4357289ccbefaae12e144f228081c9d9969b2f407c32b6a97"} Feb 02 12:29:00 crc kubenswrapper[4846]: I0202 12:29:00.962355 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-lwh8p" podStartSLOduration=2.598366809 podStartE2EDuration="16.962340395s" podCreationTimestamp="2026-02-02 12:28:44 +0000 UTC" firstStartedPulling="2026-02-02 12:28:45.114531109 +0000 UTC m=+1156.343117972" lastFinishedPulling="2026-02-02 12:28:59.478504685 +0000 UTC m=+1170.707091558" observedRunningTime="2026-02-02 12:29:00.954232581 +0000 UTC m=+1172.182819444" watchObservedRunningTime="2026-02-02 12:29:00.962340395 +0000 UTC m=+1172.190927258" Feb 02 12:29:01 crc kubenswrapper[4846]: I0202 12:29:01.008464 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-b2hcc-config-9f76v" podStartSLOduration=11.00844627 podStartE2EDuration="11.00844627s" podCreationTimestamp="2026-02-02 12:28:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:01.002436298 +0000 UTC m=+1172.231023161" watchObservedRunningTime="2026-02-02 12:29:01.00844627 +0000 UTC m=+1172.237033133" Feb 02 12:29:01 crc kubenswrapper[4846]: I0202 12:29:01.983217 4846 generic.go:334] "Generic (PLEG): container finished" podID="a00f0a2e-5571-4d2b-9be9-700e7186b93a" containerID="6dbe9840669d94a9aa14d90cfc98a8e3f74cc3f30a489967e47e07ffa42a5e82" exitCode=0 Feb 02 12:29:01 crc kubenswrapper[4846]: I0202 12:29:01.983304 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc-config-9f76v" event={"ID":"a00f0a2e-5571-4d2b-9be9-700e7186b93a","Type":"ContainerDied","Data":"6dbe9840669d94a9aa14d90cfc98a8e3f74cc3f30a489967e47e07ffa42a5e82"} Feb 02 12:29:01 crc kubenswrapper[4846]: I0202 12:29:01.988993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" event={"ID":"99c7f7c1-1b76-4d26-84e7-a49eaa21a006","Type":"ContainerStarted","Data":"f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18"} Feb 02 12:29:02 crc kubenswrapper[4846]: I0202 12:29:02.024512 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" podStartSLOduration=12.024489841 podStartE2EDuration="12.024489841s" podCreationTimestamp="2026-02-02 12:28:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:02.018916231 +0000 UTC m=+1173.247503114" watchObservedRunningTime="2026-02-02 12:29:02.024489841 +0000 UTC m=+1173.253076704" Feb 02 12:29:02 crc kubenswrapper[4846]: I0202 12:29:02.998278 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:29:04 crc kubenswrapper[4846]: I0202 12:29:04.823709 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-b2hcc" Feb 02 12:29:04 crc kubenswrapper[4846]: I0202 12:29:04.992820 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cd5d-account-create-update-7xmpx" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.016089 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pnzdq" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.019020 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tb9t6" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.032090 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pnzdq" event={"ID":"3dd98e97-bd56-4dde-8592-8836bbc2a0ee","Type":"ContainerDied","Data":"bf3cd3627f20fb5bce2625d78cb349835c021051fb6cb49e0ef354ae18ceb449"} Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.032147 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf3cd3627f20fb5bce2625d78cb349835c021051fb6cb49e0ef354ae18ceb449" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.032104 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pnzdq" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.032386 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-33ff-account-create-update-mpf9l" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.034134 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-nm26b" event={"ID":"bbe30827-4587-4d5a-a651-df65604399ee","Type":"ContainerDied","Data":"f343503c6fece3528c8854591dea29eb94684b06326c690d79eced45bbeea3e3"} Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.034171 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f343503c6fece3528c8854591dea29eb94684b06326c690d79eced45bbeea3e3" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.046250 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c2bc-account-create-update-2r5pb" event={"ID":"a488c453-b188-4d57-a136-0a002cd342bb","Type":"ContainerDied","Data":"5faeb8b832b4c5792683dc0f84c3887f3849f6e64cdf573933662409e9bd39c9"} Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.046335 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5faeb8b832b4c5792683dc0f84c3887f3849f6e64cdf573933662409e9bd39c9" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.051900 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.052419 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c2bc-account-create-update-2r5pb" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.053045 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cd5d-account-create-update-7xmpx" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.053070 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cd5d-account-create-update-7xmpx" event={"ID":"94a78e75-b5be-43f1-ac04-911885a75395","Type":"ContainerDied","Data":"1864a8c9fced459a4f013e854d65a92c0c5e04ca483d785f69cc236e32a88c19"} Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.053103 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1864a8c9fced459a4f013e854d65a92c0c5e04ca483d785f69cc236e32a88c19" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.072600 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc-config-9f76v" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.072652 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nm26b" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.072729 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc-config-9f76v" event={"ID":"a00f0a2e-5571-4d2b-9be9-700e7186b93a","Type":"ContainerDied","Data":"5254cbc22f86c6859ea483797c83efad8117822b9d8c00153bf37afd3953049c"} Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.072778 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5254cbc22f86c6859ea483797c83efad8117822b9d8c00153bf37afd3953049c" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.083667 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-tb9t6" event={"ID":"d670b5c2-8eea-4413-90df-0f4b72d8acd7","Type":"ContainerDied","Data":"cba43c57f0a7ec30b5e679b0287e921396a9ec483977f8bbe73ae7392d635e21"} Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.083714 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cba43c57f0a7ec30b5e679b0287e921396a9ec483977f8bbe73ae7392d635e21" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.083762 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-tb9t6" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.094571 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-33ff-account-create-update-mpf9l" event={"ID":"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1","Type":"ContainerDied","Data":"a0e4edba90d667e4357289ccbefaae12e144f228081c9d9969b2f407c32b6a97"} Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.094637 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0e4edba90d667e4357289ccbefaae12e144f228081c9d9969b2f407c32b6a97" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.094728 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-33ff-account-create-update-mpf9l" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148071 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d670b5c2-8eea-4413-90df-0f4b72d8acd7-operator-scripts\") pod \"d670b5c2-8eea-4413-90df-0f4b72d8acd7\" (UID: \"d670b5c2-8eea-4413-90df-0f4b72d8acd7\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148136 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkgvk\" (UniqueName: \"kubernetes.io/projected/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-kube-api-access-xkgvk\") pod \"3dd98e97-bd56-4dde-8592-8836bbc2a0ee\" (UID: \"3dd98e97-bd56-4dde-8592-8836bbc2a0ee\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148164 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gpnc\" (UniqueName: \"kubernetes.io/projected/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-kube-api-access-5gpnc\") pod \"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1\" (UID: \"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148252 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-operator-scripts\") pod \"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1\" (UID: \"c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148287 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a488c453-b188-4d57-a136-0a002cd342bb-operator-scripts\") pod \"a488c453-b188-4d57-a136-0a002cd342bb\" (UID: \"a488c453-b188-4d57-a136-0a002cd342bb\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148335 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjhvt\" (UniqueName: \"kubernetes.io/projected/d670b5c2-8eea-4413-90df-0f4b72d8acd7-kube-api-access-qjhvt\") pod \"d670b5c2-8eea-4413-90df-0f4b72d8acd7\" (UID: \"d670b5c2-8eea-4413-90df-0f4b72d8acd7\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148369 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9h82\" (UniqueName: \"kubernetes.io/projected/a488c453-b188-4d57-a136-0a002cd342bb-kube-api-access-q9h82\") pod \"a488c453-b188-4d57-a136-0a002cd342bb\" (UID: \"a488c453-b188-4d57-a136-0a002cd342bb\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148401 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-additional-scripts\") pod \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148455 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-scripts\") pod \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148526 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run-ovn\") pod \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148544 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x84fh\" (UniqueName: \"kubernetes.io/projected/94a78e75-b5be-43f1-ac04-911885a75395-kube-api-access-x84fh\") pod \"94a78e75-b5be-43f1-ac04-911885a75395\" (UID: \"94a78e75-b5be-43f1-ac04-911885a75395\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148567 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgpch\" (UniqueName: \"kubernetes.io/projected/a00f0a2e-5571-4d2b-9be9-700e7186b93a-kube-api-access-tgpch\") pod \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148593 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-log-ovn\") pod \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148614 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-operator-scripts\") pod \"3dd98e97-bd56-4dde-8592-8836bbc2a0ee\" (UID: \"3dd98e97-bd56-4dde-8592-8836bbc2a0ee\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148707 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run\") pod \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\" (UID: \"a00f0a2e-5571-4d2b-9be9-700e7186b93a\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.148738 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94a78e75-b5be-43f1-ac04-911885a75395-operator-scripts\") pod \"94a78e75-b5be-43f1-ac04-911885a75395\" (UID: \"94a78e75-b5be-43f1-ac04-911885a75395\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149025 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1" (UID: "c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149079 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d670b5c2-8eea-4413-90df-0f4b72d8acd7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d670b5c2-8eea-4413-90df-0f4b72d8acd7" (UID: "d670b5c2-8eea-4413-90df-0f4b72d8acd7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149102 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a00f0a2e-5571-4d2b-9be9-700e7186b93a" (UID: "a00f0a2e-5571-4d2b-9be9-700e7186b93a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149375 4846 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149395 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d670b5c2-8eea-4413-90df-0f4b72d8acd7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149410 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149262 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a488c453-b188-4d57-a136-0a002cd342bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a488c453-b188-4d57-a136-0a002cd342bb" (UID: "a488c453-b188-4d57-a136-0a002cd342bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149457 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a00f0a2e-5571-4d2b-9be9-700e7186b93a" (UID: "a00f0a2e-5571-4d2b-9be9-700e7186b93a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149487 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94a78e75-b5be-43f1-ac04-911885a75395-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "94a78e75-b5be-43f1-ac04-911885a75395" (UID: "94a78e75-b5be-43f1-ac04-911885a75395"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149760 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run" (OuterVolumeSpecName: "var-run") pod "a00f0a2e-5571-4d2b-9be9-700e7186b93a" (UID: "a00f0a2e-5571-4d2b-9be9-700e7186b93a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.149912 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3dd98e97-bd56-4dde-8592-8836bbc2a0ee" (UID: "3dd98e97-bd56-4dde-8592-8836bbc2a0ee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.150000 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a00f0a2e-5571-4d2b-9be9-700e7186b93a" (UID: "a00f0a2e-5571-4d2b-9be9-700e7186b93a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.150754 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-scripts" (OuterVolumeSpecName: "scripts") pod "a00f0a2e-5571-4d2b-9be9-700e7186b93a" (UID: "a00f0a2e-5571-4d2b-9be9-700e7186b93a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.163868 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a78e75-b5be-43f1-ac04-911885a75395-kube-api-access-x84fh" (OuterVolumeSpecName: "kube-api-access-x84fh") pod "94a78e75-b5be-43f1-ac04-911885a75395" (UID: "94a78e75-b5be-43f1-ac04-911885a75395"). InnerVolumeSpecName "kube-api-access-x84fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.163978 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-kube-api-access-5gpnc" (OuterVolumeSpecName: "kube-api-access-5gpnc") pod "c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1" (UID: "c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1"). InnerVolumeSpecName "kube-api-access-5gpnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.164067 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-kube-api-access-xkgvk" (OuterVolumeSpecName: "kube-api-access-xkgvk") pod "3dd98e97-bd56-4dde-8592-8836bbc2a0ee" (UID: "3dd98e97-bd56-4dde-8592-8836bbc2a0ee"). InnerVolumeSpecName "kube-api-access-xkgvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.164448 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d670b5c2-8eea-4413-90df-0f4b72d8acd7-kube-api-access-qjhvt" (OuterVolumeSpecName: "kube-api-access-qjhvt") pod "d670b5c2-8eea-4413-90df-0f4b72d8acd7" (UID: "d670b5c2-8eea-4413-90df-0f4b72d8acd7"). InnerVolumeSpecName "kube-api-access-qjhvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.164867 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a488c453-b188-4d57-a136-0a002cd342bb-kube-api-access-q9h82" (OuterVolumeSpecName: "kube-api-access-q9h82") pod "a488c453-b188-4d57-a136-0a002cd342bb" (UID: "a488c453-b188-4d57-a136-0a002cd342bb"). InnerVolumeSpecName "kube-api-access-q9h82". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.167226 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a00f0a2e-5571-4d2b-9be9-700e7186b93a-kube-api-access-tgpch" (OuterVolumeSpecName: "kube-api-access-tgpch") pod "a00f0a2e-5571-4d2b-9be9-700e7186b93a" (UID: "a00f0a2e-5571-4d2b-9be9-700e7186b93a"). InnerVolumeSpecName "kube-api-access-tgpch". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.250240 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnvh2\" (UniqueName: \"kubernetes.io/projected/bbe30827-4587-4d5a-a651-df65604399ee-kube-api-access-nnvh2\") pod \"bbe30827-4587-4d5a-a651-df65604399ee\" (UID: \"bbe30827-4587-4d5a-a651-df65604399ee\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.250671 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbe30827-4587-4d5a-a651-df65604399ee-operator-scripts\") pod \"bbe30827-4587-4d5a-a651-df65604399ee\" (UID: \"bbe30827-4587-4d5a-a651-df65604399ee\") " Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251083 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94a78e75-b5be-43f1-ac04-911885a75395-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251107 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkgvk\" (UniqueName: \"kubernetes.io/projected/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-kube-api-access-xkgvk\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251118 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gpnc\" (UniqueName: \"kubernetes.io/projected/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1-kube-api-access-5gpnc\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251127 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a488c453-b188-4d57-a136-0a002cd342bb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251136 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjhvt\" (UniqueName: \"kubernetes.io/projected/d670b5c2-8eea-4413-90df-0f4b72d8acd7-kube-api-access-qjhvt\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251146 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9h82\" (UniqueName: \"kubernetes.io/projected/a488c453-b188-4d57-a136-0a002cd342bb-kube-api-access-q9h82\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251155 4846 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251165 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a00f0a2e-5571-4d2b-9be9-700e7186b93a-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251174 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x84fh\" (UniqueName: \"kubernetes.io/projected/94a78e75-b5be-43f1-ac04-911885a75395-kube-api-access-x84fh\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251181 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgpch\" (UniqueName: \"kubernetes.io/projected/a00f0a2e-5571-4d2b-9be9-700e7186b93a-kube-api-access-tgpch\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251189 4846 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251198 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3dd98e97-bd56-4dde-8592-8836bbc2a0ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251207 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a00f0a2e-5571-4d2b-9be9-700e7186b93a-var-run\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.251247 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbe30827-4587-4d5a-a651-df65604399ee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bbe30827-4587-4d5a-a651-df65604399ee" (UID: "bbe30827-4587-4d5a-a651-df65604399ee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.253285 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbe30827-4587-4d5a-a651-df65604399ee-kube-api-access-nnvh2" (OuterVolumeSpecName: "kube-api-access-nnvh2") pod "bbe30827-4587-4d5a-a651-df65604399ee" (UID: "bbe30827-4587-4d5a-a651-df65604399ee"). InnerVolumeSpecName "kube-api-access-nnvh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.353007 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnvh2\" (UniqueName: \"kubernetes.io/projected/bbe30827-4587-4d5a-a651-df65604399ee-kube-api-access-nnvh2\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:05 crc kubenswrapper[4846]: I0202 12:29:05.353042 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbe30827-4587-4d5a-a651-df65604399ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.103788 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c2bc-account-create-update-2r5pb" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.103766 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xf7xw" event={"ID":"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757","Type":"ContainerStarted","Data":"66243d036d20a71042e4208e8b9c88b63381e5c9b16d891c7d371cdcefee7420"} Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.103796 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-nm26b" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.134345 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xf7xw" podStartSLOduration=10.600791058 podStartE2EDuration="15.134329002s" podCreationTimestamp="2026-02-02 12:28:51 +0000 UTC" firstStartedPulling="2026-02-02 12:29:00.297975152 +0000 UTC m=+1171.526562005" lastFinishedPulling="2026-02-02 12:29:04.831513086 +0000 UTC m=+1176.060099949" observedRunningTime="2026-02-02 12:29:06.120308991 +0000 UTC m=+1177.348895874" watchObservedRunningTime="2026-02-02 12:29:06.134329002 +0000 UTC m=+1177.362915865" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.219131 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-b2hcc-config-9f76v"] Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.230312 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-b2hcc-config-9f76v"] Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.337857 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-b2hcc-config-9tr2t"] Feb 02 12:29:06 crc kubenswrapper[4846]: E0202 12:29:06.338237 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbe30827-4587-4d5a-a651-df65604399ee" containerName="mariadb-database-create" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338260 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbe30827-4587-4d5a-a651-df65604399ee" containerName="mariadb-database-create" Feb 02 12:29:06 crc kubenswrapper[4846]: E0202 12:29:06.338284 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00f0a2e-5571-4d2b-9be9-700e7186b93a" containerName="ovn-config" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338294 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00f0a2e-5571-4d2b-9be9-700e7186b93a" containerName="ovn-config" Feb 02 12:29:06 crc kubenswrapper[4846]: E0202 12:29:06.338307 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd98e97-bd56-4dde-8592-8836bbc2a0ee" containerName="mariadb-database-create" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338314 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd98e97-bd56-4dde-8592-8836bbc2a0ee" containerName="mariadb-database-create" Feb 02 12:29:06 crc kubenswrapper[4846]: E0202 12:29:06.338326 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a488c453-b188-4d57-a136-0a002cd342bb" containerName="mariadb-account-create-update" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338333 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a488c453-b188-4d57-a136-0a002cd342bb" containerName="mariadb-account-create-update" Feb 02 12:29:06 crc kubenswrapper[4846]: E0202 12:29:06.338347 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d670b5c2-8eea-4413-90df-0f4b72d8acd7" containerName="mariadb-database-create" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338354 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d670b5c2-8eea-4413-90df-0f4b72d8acd7" containerName="mariadb-database-create" Feb 02 12:29:06 crc kubenswrapper[4846]: E0202 12:29:06.338369 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1" containerName="mariadb-account-create-update" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338377 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1" containerName="mariadb-account-create-update" Feb 02 12:29:06 crc kubenswrapper[4846]: E0202 12:29:06.338398 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a78e75-b5be-43f1-ac04-911885a75395" containerName="mariadb-account-create-update" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338406 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a78e75-b5be-43f1-ac04-911885a75395" containerName="mariadb-account-create-update" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338579 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbe30827-4587-4d5a-a651-df65604399ee" containerName="mariadb-database-create" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338599 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a78e75-b5be-43f1-ac04-911885a75395" containerName="mariadb-account-create-update" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338611 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00f0a2e-5571-4d2b-9be9-700e7186b93a" containerName="ovn-config" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338642 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd98e97-bd56-4dde-8592-8836bbc2a0ee" containerName="mariadb-database-create" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338660 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a488c453-b188-4d57-a136-0a002cd342bb" containerName="mariadb-account-create-update" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338674 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1" containerName="mariadb-account-create-update" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.338689 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d670b5c2-8eea-4413-90df-0f4b72d8acd7" containerName="mariadb-database-create" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.339219 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.346093 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.348741 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b2hcc-config-9tr2t"] Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.474441 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmm9s\" (UniqueName: \"kubernetes.io/projected/890cee21-11ca-48ba-88d4-0418d57f01c4-kube-api-access-vmm9s\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.474500 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.474587 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-additional-scripts\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.474795 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-log-ovn\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.474870 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-scripts\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.475069 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run-ovn\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.576017 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmm9s\" (UniqueName: \"kubernetes.io/projected/890cee21-11ca-48ba-88d4-0418d57f01c4-kube-api-access-vmm9s\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.576077 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.576119 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-additional-scripts\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.576189 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-log-ovn\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.576212 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-scripts\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.576264 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run-ovn\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.576470 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.576508 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run-ovn\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.576570 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-log-ovn\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.577606 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-additional-scripts\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.578491 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-scripts\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.598885 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmm9s\" (UniqueName: \"kubernetes.io/projected/890cee21-11ca-48ba-88d4-0418d57f01c4-kube-api-access-vmm9s\") pod \"ovn-controller-b2hcc-config-9tr2t\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:06 crc kubenswrapper[4846]: I0202 12:29:06.655761 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:07 crc kubenswrapper[4846]: I0202 12:29:07.116440 4846 generic.go:334] "Generic (PLEG): container finished" podID="a4d8a08e-0bf2-4830-b1f0-c5d599fc5757" containerID="c62f51d03fd9f435dd3c6722b2d45f51aea94cdc0dc8bea514ed230dba6a1885" exitCode=0 Feb 02 12:29:07 crc kubenswrapper[4846]: I0202 12:29:07.116583 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lwh8p" event={"ID":"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757","Type":"ContainerDied","Data":"c62f51d03fd9f435dd3c6722b2d45f51aea94cdc0dc8bea514ed230dba6a1885"} Feb 02 12:29:07 crc kubenswrapper[4846]: W0202 12:29:07.131173 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod890cee21_11ca_48ba_88d4_0418d57f01c4.slice/crio-9b045f3d05506f86e2358773b35c2fc1ecfb41328d3245d076db9bc3cb858b67 WatchSource:0}: Error finding container 9b045f3d05506f86e2358773b35c2fc1ecfb41328d3245d076db9bc3cb858b67: Status 404 returned error can't find the container with id 9b045f3d05506f86e2358773b35c2fc1ecfb41328d3245d076db9bc3cb858b67 Feb 02 12:29:07 crc kubenswrapper[4846]: I0202 12:29:07.139207 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b2hcc-config-9tr2t"] Feb 02 12:29:07 crc kubenswrapper[4846]: I0202 12:29:07.439143 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a00f0a2e-5571-4d2b-9be9-700e7186b93a" path="/var/lib/kubelet/pods/a00f0a2e-5571-4d2b-9be9-700e7186b93a/volumes" Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.128568 4846 generic.go:334] "Generic (PLEG): container finished" podID="890cee21-11ca-48ba-88d4-0418d57f01c4" containerID="052031d7c9748afe7ec26c0238c176702ba54688df56666d646d445914895bf0" exitCode=0 Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.128698 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc-config-9tr2t" event={"ID":"890cee21-11ca-48ba-88d4-0418d57f01c4","Type":"ContainerDied","Data":"052031d7c9748afe7ec26c0238c176702ba54688df56666d646d445914895bf0"} Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.128787 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc-config-9tr2t" event={"ID":"890cee21-11ca-48ba-88d4-0418d57f01c4","Type":"ContainerStarted","Data":"9b045f3d05506f86e2358773b35c2fc1ecfb41328d3245d076db9bc3cb858b67"} Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.559237 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lwh8p" Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.716343 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-db-sync-config-data\") pod \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.716432 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-config-data\") pod \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.716490 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-combined-ca-bundle\") pod \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.716529 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hckkz\" (UniqueName: \"kubernetes.io/projected/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-kube-api-access-hckkz\") pod \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\" (UID: \"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757\") " Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.723122 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-kube-api-access-hckkz" (OuterVolumeSpecName: "kube-api-access-hckkz") pod "a4d8a08e-0bf2-4830-b1f0-c5d599fc5757" (UID: "a4d8a08e-0bf2-4830-b1f0-c5d599fc5757"). InnerVolumeSpecName "kube-api-access-hckkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.724844 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a4d8a08e-0bf2-4830-b1f0-c5d599fc5757" (UID: "a4d8a08e-0bf2-4830-b1f0-c5d599fc5757"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.747368 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4d8a08e-0bf2-4830-b1f0-c5d599fc5757" (UID: "a4d8a08e-0bf2-4830-b1f0-c5d599fc5757"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.768495 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-config-data" (OuterVolumeSpecName: "config-data") pod "a4d8a08e-0bf2-4830-b1f0-c5d599fc5757" (UID: "a4d8a08e-0bf2-4830-b1f0-c5d599fc5757"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.819116 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.819160 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.819173 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hckkz\" (UniqueName: \"kubernetes.io/projected/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-kube-api-access-hckkz\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:08 crc kubenswrapper[4846]: I0202 12:29:08.819204 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.138166 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-lwh8p" event={"ID":"a4d8a08e-0bf2-4830-b1f0-c5d599fc5757","Type":"ContainerDied","Data":"063dee3c9dc61f2259f99b3fecc01cf936ec4c545c8ecaf0ab9e93128aca1780"} Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.138195 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-lwh8p" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.138201 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="063dee3c9dc61f2259f99b3fecc01cf936ec4c545c8ecaf0ab9e93128aca1780" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.140307 4846 generic.go:334] "Generic (PLEG): container finished" podID="b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757" containerID="66243d036d20a71042e4208e8b9c88b63381e5c9b16d891c7d371cdcefee7420" exitCode=0 Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.140512 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xf7xw" event={"ID":"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757","Type":"ContainerDied","Data":"66243d036d20a71042e4208e8b9c88b63381e5c9b16d891c7d371cdcefee7420"} Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.414986 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.538132 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-log-ovn\") pod \"890cee21-11ca-48ba-88d4-0418d57f01c4\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.538203 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-scripts\") pod \"890cee21-11ca-48ba-88d4-0418d57f01c4\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.538231 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmm9s\" (UniqueName: \"kubernetes.io/projected/890cee21-11ca-48ba-88d4-0418d57f01c4-kube-api-access-vmm9s\") pod \"890cee21-11ca-48ba-88d4-0418d57f01c4\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.538374 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run\") pod \"890cee21-11ca-48ba-88d4-0418d57f01c4\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.538399 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run-ovn\") pod \"890cee21-11ca-48ba-88d4-0418d57f01c4\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.538474 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-additional-scripts\") pod \"890cee21-11ca-48ba-88d4-0418d57f01c4\" (UID: \"890cee21-11ca-48ba-88d4-0418d57f01c4\") " Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.539783 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "890cee21-11ca-48ba-88d4-0418d57f01c4" (UID: "890cee21-11ca-48ba-88d4-0418d57f01c4"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.540096 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "890cee21-11ca-48ba-88d4-0418d57f01c4" (UID: "890cee21-11ca-48ba-88d4-0418d57f01c4"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.540065 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run" (OuterVolumeSpecName: "var-run") pod "890cee21-11ca-48ba-88d4-0418d57f01c4" (UID: "890cee21-11ca-48ba-88d4-0418d57f01c4"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.541314 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "890cee21-11ca-48ba-88d4-0418d57f01c4" (UID: "890cee21-11ca-48ba-88d4-0418d57f01c4"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.542999 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-scripts" (OuterVolumeSpecName: "scripts") pod "890cee21-11ca-48ba-88d4-0418d57f01c4" (UID: "890cee21-11ca-48ba-88d4-0418d57f01c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.557900 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/890cee21-11ca-48ba-88d4-0418d57f01c4-kube-api-access-vmm9s" (OuterVolumeSpecName: "kube-api-access-vmm9s") pod "890cee21-11ca-48ba-88d4-0418d57f01c4" (UID: "890cee21-11ca-48ba-88d4-0418d57f01c4"). InnerVolumeSpecName "kube-api-access-vmm9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.604429 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lw7hb"] Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.604706 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" podUID="99c7f7c1-1b76-4d26-84e7-a49eaa21a006" containerName="dnsmasq-dns" containerID="cri-o://f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18" gracePeriod=10 Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.606764 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.641523 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.641564 4846 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.641594 4846 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.641652 4846 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/890cee21-11ca-48ba-88d4-0418d57f01c4-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.641664 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/890cee21-11ca-48ba-88d4-0418d57f01c4-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.641674 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmm9s\" (UniqueName: \"kubernetes.io/projected/890cee21-11ca-48ba-88d4-0418d57f01c4-kube-api-access-vmm9s\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.653750 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-skcw8"] Feb 02 12:29:09 crc kubenswrapper[4846]: E0202 12:29:09.654211 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="890cee21-11ca-48ba-88d4-0418d57f01c4" containerName="ovn-config" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.654232 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="890cee21-11ca-48ba-88d4-0418d57f01c4" containerName="ovn-config" Feb 02 12:29:09 crc kubenswrapper[4846]: E0202 12:29:09.654265 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d8a08e-0bf2-4830-b1f0-c5d599fc5757" containerName="glance-db-sync" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.654273 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d8a08e-0bf2-4830-b1f0-c5d599fc5757" containerName="glance-db-sync" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.654452 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4d8a08e-0bf2-4830-b1f0-c5d599fc5757" containerName="glance-db-sync" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.654480 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="890cee21-11ca-48ba-88d4-0418d57f01c4" containerName="ovn-config" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.655529 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.677179 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-skcw8"] Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.844927 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-swift-storage-0\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.845543 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-config\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.845609 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-sb\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.845653 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-svc\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.845748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-nb\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.845768 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnrt4\" (UniqueName: \"kubernetes.io/projected/efcf28af-d066-46a0-b547-5cd80cbb0fa6-kube-api-access-vnrt4\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.948145 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-nb\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.948210 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnrt4\" (UniqueName: \"kubernetes.io/projected/efcf28af-d066-46a0-b547-5cd80cbb0fa6-kube-api-access-vnrt4\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.948236 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-swift-storage-0\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.948265 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-config\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.948311 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-sb\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.948335 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-svc\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.949149 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-swift-storage-0\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.949184 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-nb\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.949203 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-svc\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.949528 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-config\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:09 crc kubenswrapper[4846]: I0202 12:29:09.951255 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-sb\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.001997 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnrt4\" (UniqueName: \"kubernetes.io/projected/efcf28af-d066-46a0-b547-5cd80cbb0fa6-kube-api-access-vnrt4\") pod \"dnsmasq-dns-54cc5c87cc-skcw8\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.002568 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.132258 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.155195 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfkq6\" (UniqueName: \"kubernetes.io/projected/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-kube-api-access-sfkq6\") pod \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.155256 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-swift-storage-0\") pod \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.155396 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-sb\") pod \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.155434 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-config\") pod \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.155459 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-nb\") pod \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.155569 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-svc\") pod \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\" (UID: \"99c7f7c1-1b76-4d26-84e7-a49eaa21a006\") " Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.166055 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-kube-api-access-sfkq6" (OuterVolumeSpecName: "kube-api-access-sfkq6") pod "99c7f7c1-1b76-4d26-84e7-a49eaa21a006" (UID: "99c7f7c1-1b76-4d26-84e7-a49eaa21a006"). InnerVolumeSpecName "kube-api-access-sfkq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.198357 4846 generic.go:334] "Generic (PLEG): container finished" podID="99c7f7c1-1b76-4d26-84e7-a49eaa21a006" containerID="f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18" exitCode=0 Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.198424 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" event={"ID":"99c7f7c1-1b76-4d26-84e7-a49eaa21a006","Type":"ContainerDied","Data":"f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18"} Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.198452 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" event={"ID":"99c7f7c1-1b76-4d26-84e7-a49eaa21a006","Type":"ContainerDied","Data":"fc123523fe75b12a27a66f778e3b01d4407e7d406fd37186a25da89ff4aaa2f1"} Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.198468 4846 scope.go:117] "RemoveContainer" containerID="f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.198586 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5799dddbf-lw7hb" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.209757 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc-config-9tr2t" event={"ID":"890cee21-11ca-48ba-88d4-0418d57f01c4","Type":"ContainerDied","Data":"9b045f3d05506f86e2358773b35c2fc1ecfb41328d3245d076db9bc3cb858b67"} Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.209828 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b045f3d05506f86e2358773b35c2fc1ecfb41328d3245d076db9bc3cb858b67" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.209898 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc-config-9tr2t" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.226645 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "99c7f7c1-1b76-4d26-84e7-a49eaa21a006" (UID: "99c7f7c1-1b76-4d26-84e7-a49eaa21a006"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.238822 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "99c7f7c1-1b76-4d26-84e7-a49eaa21a006" (UID: "99c7f7c1-1b76-4d26-84e7-a49eaa21a006"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.247248 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "99c7f7c1-1b76-4d26-84e7-a49eaa21a006" (UID: "99c7f7c1-1b76-4d26-84e7-a49eaa21a006"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.247824 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "99c7f7c1-1b76-4d26-84e7-a49eaa21a006" (UID: "99c7f7c1-1b76-4d26-84e7-a49eaa21a006"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.252389 4846 scope.go:117] "RemoveContainer" containerID="3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.261694 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfkq6\" (UniqueName: \"kubernetes.io/projected/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-kube-api-access-sfkq6\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.261750 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.261763 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.261776 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.261790 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.282716 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-config" (OuterVolumeSpecName: "config") pod "99c7f7c1-1b76-4d26-84e7-a49eaa21a006" (UID: "99c7f7c1-1b76-4d26-84e7-a49eaa21a006"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.292746 4846 scope.go:117] "RemoveContainer" containerID="f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18" Feb 02 12:29:10 crc kubenswrapper[4846]: E0202 12:29:10.294885 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18\": container with ID starting with f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18 not found: ID does not exist" containerID="f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.294941 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18"} err="failed to get container status \"f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18\": rpc error: code = NotFound desc = could not find container \"f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18\": container with ID starting with f58a49e28a0afc0da6a9867737e7a79e662ee40f186961119139588b5a83ad18 not found: ID does not exist" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.294973 4846 scope.go:117] "RemoveContainer" containerID="3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748" Feb 02 12:29:10 crc kubenswrapper[4846]: E0202 12:29:10.296284 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748\": container with ID starting with 3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748 not found: ID does not exist" containerID="3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.296322 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748"} err="failed to get container status \"3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748\": rpc error: code = NotFound desc = could not find container \"3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748\": container with ID starting with 3447d1429f92efedcde7ea10ee362f4892908baf3c51995a6ad69db3c650b748 not found: ID does not exist" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.363527 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c7f7c1-1b76-4d26-84e7-a49eaa21a006-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.520326 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-b2hcc-config-9tr2t"] Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.527516 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-b2hcc-config-9tr2t"] Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.566513 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lw7hb"] Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.576948 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5799dddbf-lw7hb"] Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.599377 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-skcw8"] Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.733360 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.780243 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djxp7\" (UniqueName: \"kubernetes.io/projected/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-kube-api-access-djxp7\") pod \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.780313 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-combined-ca-bundle\") pod \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.780404 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-config-data\") pod \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\" (UID: \"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757\") " Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.808515 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-kube-api-access-djxp7" (OuterVolumeSpecName: "kube-api-access-djxp7") pod "b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757" (UID: "b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757"). InnerVolumeSpecName "kube-api-access-djxp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.836217 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757" (UID: "b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.845156 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-config-data" (OuterVolumeSpecName: "config-data") pod "b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757" (UID: "b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.882308 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djxp7\" (UniqueName: \"kubernetes.io/projected/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-kube-api-access-djxp7\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.882347 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:10 crc kubenswrapper[4846]: I0202 12:29:10.882357 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.221219 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xf7xw" event={"ID":"b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757","Type":"ContainerDied","Data":"cc2f9ab12743b74111196c4c26f621d8c66e5e26cdb67ab799e309f0433783cd"} Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.221598 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc2f9ab12743b74111196c4c26f621d8c66e5e26cdb67ab799e309f0433783cd" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.221383 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xf7xw" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.223551 4846 generic.go:334] "Generic (PLEG): container finished" podID="efcf28af-d066-46a0-b547-5cd80cbb0fa6" containerID="71b2bf04a20293a34531d93bbc57a0e6142283523db7fb9dfbb50ae7af61ee55" exitCode=0 Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.223648 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" event={"ID":"efcf28af-d066-46a0-b547-5cd80cbb0fa6","Type":"ContainerDied","Data":"71b2bf04a20293a34531d93bbc57a0e6142283523db7fb9dfbb50ae7af61ee55"} Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.223713 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" event={"ID":"efcf28af-d066-46a0-b547-5cd80cbb0fa6","Type":"ContainerStarted","Data":"516a249500970cc9ae4b43a339b3085b9f02ff5e00da49b9e7340f7d87a5c945"} Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.363096 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-skcw8"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.395893 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wgp84"] Feb 02 12:29:11 crc kubenswrapper[4846]: E0202 12:29:11.396413 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c7f7c1-1b76-4d26-84e7-a49eaa21a006" containerName="init" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.396434 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c7f7c1-1b76-4d26-84e7-a49eaa21a006" containerName="init" Feb 02 12:29:11 crc kubenswrapper[4846]: E0202 12:29:11.396467 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757" containerName="keystone-db-sync" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.396473 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757" containerName="keystone-db-sync" Feb 02 12:29:11 crc kubenswrapper[4846]: E0202 12:29:11.396487 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c7f7c1-1b76-4d26-84e7-a49eaa21a006" containerName="dnsmasq-dns" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.396495 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c7f7c1-1b76-4d26-84e7-a49eaa21a006" containerName="dnsmasq-dns" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.396714 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="99c7f7c1-1b76-4d26-84e7-a49eaa21a006" containerName="dnsmasq-dns" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.396735 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757" containerName="keystone-db-sync" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.397550 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.401602 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.401755 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9wp9t" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.401912 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.409047 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.416615 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-866975879-w6pjj"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.418011 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.422154 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.459945 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="890cee21-11ca-48ba-88d4-0418d57f01c4" path="/var/lib/kubelet/pods/890cee21-11ca-48ba-88d4-0418d57f01c4/volumes" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.461094 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99c7f7c1-1b76-4d26-84e7-a49eaa21a006" path="/var/lib/kubelet/pods/99c7f7c1-1b76-4d26-84e7-a49eaa21a006/volumes" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.461734 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wgp84"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.467169 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-866975879-w6pjj"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-sb\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496322 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9tzt\" (UniqueName: \"kubernetes.io/projected/d60e3c65-fdf5-411b-970c-7c896d567c9a-kube-api-access-t9tzt\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496374 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-config-data\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496417 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-swift-storage-0\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496441 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-svc\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496471 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-credential-keys\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496532 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-scripts\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496590 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pwnt\" (UniqueName: \"kubernetes.io/projected/c427a83f-c397-4886-bcd9-a7c992c7fba9-kube-api-access-9pwnt\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496644 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-config\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496672 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-fernet-keys\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496736 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-nb\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.496774 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-combined-ca-bundle\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602482 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-swift-storage-0\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602526 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-svc\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602550 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-credential-keys\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-scripts\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602643 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pwnt\" (UniqueName: \"kubernetes.io/projected/c427a83f-c397-4886-bcd9-a7c992c7fba9-kube-api-access-9pwnt\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602663 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-config\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602681 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-fernet-keys\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602718 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-nb\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602740 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-combined-ca-bundle\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602762 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-sb\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602782 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9tzt\" (UniqueName: \"kubernetes.io/projected/d60e3c65-fdf5-411b-970c-7c896d567c9a-kube-api-access-t9tzt\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.602806 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-config-data\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.604834 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-svc\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.605396 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-swift-storage-0\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.606249 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-sb\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.607523 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-nb\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: E0202 12:29:11.611839 4846 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Feb 02 12:29:11 crc kubenswrapper[4846]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/efcf28af-d066-46a0-b547-5cd80cbb0fa6/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 02 12:29:11 crc kubenswrapper[4846]: > podSandboxID="516a249500970cc9ae4b43a339b3085b9f02ff5e00da49b9e7340f7d87a5c945" Feb 02 12:29:11 crc kubenswrapper[4846]: E0202 12:29:11.612012 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:29:11 crc kubenswrapper[4846]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64h68fh95h595h67fh597hfch57ch68fh5ffh6hf4h689h659h569h65bh67bh65dh594h64h5d6hd8h5bfh9fh5c4h676h5cdh56h8bh569h664h645q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vnrt4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-54cc5c87cc-skcw8_openstack(efcf28af-d066-46a0-b547-5cd80cbb0fa6): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/efcf28af-d066-46a0-b547-5cd80cbb0fa6/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 02 12:29:11 crc kubenswrapper[4846]: > logger="UnhandledError" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.612067 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-config\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.613233 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-config-data\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: E0202 12:29:11.613298 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/efcf28af-d066-46a0-b547-5cd80cbb0fa6/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" podUID="efcf28af-d066-46a0-b547-5cd80cbb0fa6" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.616836 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-combined-ca-bundle\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.625460 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-fernet-keys\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.626229 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-scripts\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.627978 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-credential-keys\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.641489 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pwnt\" (UniqueName: \"kubernetes.io/projected/c427a83f-c397-4886-bcd9-a7c992c7fba9-kube-api-access-9pwnt\") pod \"keystone-bootstrap-wgp84\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.653251 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9tzt\" (UniqueName: \"kubernetes.io/projected/d60e3c65-fdf5-411b-970c-7c896d567c9a-kube-api-access-t9tzt\") pod \"dnsmasq-dns-866975879-w6pjj\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.677012 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-2z995"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.678710 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.682696 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.682966 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-plfcv" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.683139 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.707303 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-2z995"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.716246 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.719161 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.724806 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.725132 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.745399 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.809231 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-db-sync-config-data\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.809717 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-scripts\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.809791 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-config-data\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.809868 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.809965 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.810037 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdklw\" (UniqueName: \"kubernetes.io/projected/8b58cb4f-6d15-4d97-871d-724c160ec765-kube-api-access-zdklw\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.810127 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b58cb4f-6d15-4d97-871d-724c160ec765-etc-machine-id\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.810209 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-config-data\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.810297 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-combined-ca-bundle\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.810382 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phlbl\" (UniqueName: \"kubernetes.io/projected/eaece0f9-90e4-4f49-b5e4-6566e68f115a-kube-api-access-phlbl\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.810504 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-scripts\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.810588 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-run-httpd\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.813906 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-log-httpd\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.831682 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-kd7sk"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.833662 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.837244 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.838300 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.840565 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nx5gj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.840841 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.849150 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.856830 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-5vmgj"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.857889 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.876608 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kd7sk"] Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.893909 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.894094 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-vmnbh" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.925770 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phlbl\" (UniqueName: \"kubernetes.io/projected/eaece0f9-90e4-4f49-b5e4-6566e68f115a-kube-api-access-phlbl\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.925834 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-db-sync-config-data\") pod \"barbican-db-sync-5vmgj\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.925861 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-scripts\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.925884 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5qsx\" (UniqueName: \"kubernetes.io/projected/18307d58-f249-4ec0-8556-fd98e4febb66-kube-api-access-s5qsx\") pod \"barbican-db-sync-5vmgj\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.925906 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-run-httpd\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.925921 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-log-httpd\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.925938 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-config\") pod \"neutron-db-sync-kd7sk\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.925967 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-db-sync-config-data\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.925999 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-combined-ca-bundle\") pod \"barbican-db-sync-5vmgj\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.926025 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-combined-ca-bundle\") pod \"neutron-db-sync-kd7sk\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.926044 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-scripts\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.926058 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-config-data\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.926089 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.926112 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.926135 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdklw\" (UniqueName: \"kubernetes.io/projected/8b58cb4f-6d15-4d97-871d-724c160ec765-kube-api-access-zdklw\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.926176 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b58cb4f-6d15-4d97-871d-724c160ec765-etc-machine-id\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.926197 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-config-data\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.926227 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-combined-ca-bundle\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.926243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hm5v\" (UniqueName: \"kubernetes.io/projected/e924adc7-fffe-4682-a601-cac1917e8f54-kube-api-access-9hm5v\") pod \"neutron-db-sync-kd7sk\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.928148 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-run-httpd\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.932993 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b58cb4f-6d15-4d97-871d-724c160ec765-etc-machine-id\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.935073 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-log-httpd\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.935193 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-scripts\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.940934 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-config-data\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.943288 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.944468 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-scripts\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.954575 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phlbl\" (UniqueName: \"kubernetes.io/projected/eaece0f9-90e4-4f49-b5e4-6566e68f115a-kube-api-access-phlbl\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.960269 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-db-sync-config-data\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.961095 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-config-data\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.962384 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdklw\" (UniqueName: \"kubernetes.io/projected/8b58cb4f-6d15-4d97-871d-724c160ec765-kube-api-access-zdklw\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.959392 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-combined-ca-bundle\") pod \"cinder-db-sync-2z995\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.980437 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " pod="openstack/ceilometer-0" Feb 02 12:29:11 crc kubenswrapper[4846]: I0202 12:29:11.980506 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5vmgj"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.025994 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-858jm"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.027309 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.029196 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-combined-ca-bundle\") pod \"barbican-db-sync-5vmgj\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.029234 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-combined-ca-bundle\") pod \"neutron-db-sync-kd7sk\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.029316 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hm5v\" (UniqueName: \"kubernetes.io/projected/e924adc7-fffe-4682-a601-cac1917e8f54-kube-api-access-9hm5v\") pod \"neutron-db-sync-kd7sk\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.029348 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-db-sync-config-data\") pod \"barbican-db-sync-5vmgj\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.029369 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5qsx\" (UniqueName: \"kubernetes.io/projected/18307d58-f249-4ec0-8556-fd98e4febb66-kube-api-access-s5qsx\") pod \"barbican-db-sync-5vmgj\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.029389 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-config\") pod \"neutron-db-sync-kd7sk\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.033041 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.034657 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866975879-w6pjj"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.034899 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.034941 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-tx4l4" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.038184 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-combined-ca-bundle\") pod \"barbican-db-sync-5vmgj\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.038743 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-config\") pod \"neutron-db-sync-kd7sk\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.046996 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-db-sync-config-data\") pod \"barbican-db-sync-5vmgj\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.050983 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-858jm"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.052289 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-combined-ca-bundle\") pod \"neutron-db-sync-kd7sk\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.088506 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.100968 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.110497 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-5zrpv"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.112808 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hm5v\" (UniqueName: \"kubernetes.io/projected/e924adc7-fffe-4682-a601-cac1917e8f54-kube-api-access-9hm5v\") pod \"neutron-db-sync-kd7sk\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.114509 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.135650 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-scripts\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.136779 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-config-data\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.136798 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-logs\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.136843 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-combined-ca-bundle\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.136863 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqlxr\" (UniqueName: \"kubernetes.io/projected/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-kube-api-access-dqlxr\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.152687 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5qsx\" (UniqueName: \"kubernetes.io/projected/18307d58-f249-4ec0-8556-fd98e4febb66-kube-api-access-s5qsx\") pod \"barbican-db-sync-5vmgj\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.161645 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.162495 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-5zrpv"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.236852 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238072 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-sb\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238127 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-scripts\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238147 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-config\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238193 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2d9z\" (UniqueName: \"kubernetes.io/projected/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-kube-api-access-g2d9z\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238251 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-swift-storage-0\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238354 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-nb\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238378 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-config-data\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238395 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-logs\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238461 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-combined-ca-bundle\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238482 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqlxr\" (UniqueName: \"kubernetes.io/projected/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-kube-api-access-dqlxr\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.238514 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-svc\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.239938 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-logs\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.244537 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-scripts\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.246250 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-combined-ca-bundle\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.248428 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-config-data\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.267442 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqlxr\" (UniqueName: \"kubernetes.io/projected/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-kube-api-access-dqlxr\") pod \"placement-db-sync-858jm\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.340293 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-config\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.340371 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2d9z\" (UniqueName: \"kubernetes.io/projected/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-kube-api-access-g2d9z\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.340402 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-swift-storage-0\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.340456 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-nb\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.340524 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-svc\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.340595 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-sb\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.342439 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-swift-storage-0\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.342455 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-config\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.343579 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-svc\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.344979 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-nb\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.344997 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-sb\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.367148 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2d9z\" (UniqueName: \"kubernetes.io/projected/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-kube-api-access-g2d9z\") pod \"dnsmasq-dns-f58898b5f-5zrpv\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.371136 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-858jm" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.452196 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.532417 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.534399 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.541132 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-c2sk4" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.541239 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.541284 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.543387 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.609688 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.611131 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.616582 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.628874 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.646265 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.647064 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.647222 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.647317 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.647397 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n66d\" (UniqueName: \"kubernetes.io/projected/00a49cfd-e78e-44c0-a26c-116ec72c84c8-kube-api-access-8n66d\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.649261 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kbnc\" (UniqueName: \"kubernetes.io/projected/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-kube-api-access-2kbnc\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.649474 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.649598 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-config-data\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.649779 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.649894 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.650020 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.650163 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.650289 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-scripts\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.650411 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-logs\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.651949 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wgp84"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.751668 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.751729 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-config-data\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.751765 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.751811 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.751856 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.751883 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.751932 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-scripts\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.751970 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-logs\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.751994 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.752016 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.752087 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.752125 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.752150 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n66d\" (UniqueName: \"kubernetes.io/projected/00a49cfd-e78e-44c0-a26c-116ec72c84c8-kube-api-access-8n66d\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.752178 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kbnc\" (UniqueName: \"kubernetes.io/projected/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-kube-api-access-2kbnc\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.753501 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.762958 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-scripts\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.765392 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-logs\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.765933 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.766839 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.767512 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.768293 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.771016 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.780268 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-config-data\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.780447 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.786237 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n66d\" (UniqueName: \"kubernetes.io/projected/00a49cfd-e78e-44c0-a26c-116ec72c84c8-kube-api-access-8n66d\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.793369 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-2z995"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.794542 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kbnc\" (UniqueName: \"kubernetes.io/projected/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-kube-api-access-2kbnc\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.794903 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.803899 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.808916 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866975879-w6pjj"] Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.822224 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.837100 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.907887 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:29:12 crc kubenswrapper[4846]: I0202 12:29:12.945803 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.002406 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kd7sk"] Feb 02 12:29:13 crc kubenswrapper[4846]: W0202 12:29:13.028883 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode924adc7_fffe_4682_a601_cac1917e8f54.slice/crio-15c8e863f6f4492ca509df9d7bcfd91200fc72a6997959e2cbfb9fb50594252f WatchSource:0}: Error finding container 15c8e863f6f4492ca509df9d7bcfd91200fc72a6997959e2cbfb9fb50594252f: Status 404 returned error can't find the container with id 15c8e863f6f4492ca509df9d7bcfd91200fc72a6997959e2cbfb9fb50594252f Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.054779 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:29:13 crc kubenswrapper[4846]: W0202 12:29:13.219880 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffdcb1cd_5944_4d12_8da0_9ccc2b5f9bc3.slice/crio-e835e235aaaddaa6f65f3e9decabd4976d3dd94084152f836fb7dfd61ee0a37b WatchSource:0}: Error finding container e835e235aaaddaa6f65f3e9decabd4976d3dd94084152f836fb7dfd61ee0a37b: Status 404 returned error can't find the container with id e835e235aaaddaa6f65f3e9decabd4976d3dd94084152f836fb7dfd61ee0a37b Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.227833 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-5zrpv"] Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.238966 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5vmgj"] Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.248021 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-858jm"] Feb 02 12:29:13 crc kubenswrapper[4846]: W0202 12:29:13.249850 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode02dd0b8_6bd8_40b2_8a18_64d298c72a67.slice/crio-4fa5450ef3e8e9522271cc126cc1fcf00b43a3a35f39da799acac7233e310bfc WatchSource:0}: Error finding container 4fa5450ef3e8e9522271cc126cc1fcf00b43a3a35f39da799acac7233e310bfc: Status 404 returned error can't find the container with id 4fa5450ef3e8e9522271cc126cc1fcf00b43a3a35f39da799acac7233e310bfc Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.253866 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" event={"ID":"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3","Type":"ContainerStarted","Data":"e835e235aaaddaa6f65f3e9decabd4976d3dd94084152f836fb7dfd61ee0a37b"} Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.255527 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866975879-w6pjj" event={"ID":"d60e3c65-fdf5-411b-970c-7c896d567c9a","Type":"ContainerStarted","Data":"13029f162c79d9fae2f4a8057ec65c9f54d32c005653a98972c65409caf74682"} Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.263432 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kd7sk" event={"ID":"e924adc7-fffe-4682-a601-cac1917e8f54","Type":"ContainerStarted","Data":"15c8e863f6f4492ca509df9d7bcfd91200fc72a6997959e2cbfb9fb50594252f"} Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.271909 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaece0f9-90e4-4f49-b5e4-6566e68f115a","Type":"ContainerStarted","Data":"87a9e5bce5fb483da5d5792d05755698945894c86f07a50466b033f549e6cf1e"} Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.277716 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.278222 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5vmgj" event={"ID":"18307d58-f249-4ec0-8556-fd98e4febb66","Type":"ContainerStarted","Data":"f339e7c1fd39e322f50cd38287d51f6177c816f19874355c73eed6af1433c525"} Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.285024 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wgp84" event={"ID":"c427a83f-c397-4886-bcd9-a7c992c7fba9","Type":"ContainerStarted","Data":"6755de517414e72bc7274a2791bbb44228da7566761219b74c6d5baca0a1ce22"} Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.286097 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2z995" event={"ID":"8b58cb4f-6d15-4d97-871d-724c160ec765","Type":"ContainerStarted","Data":"058e3abed1c7fa20ff7ae22cf4b12aca6926aaa714b57c36b4a25d932b2e26ab"} Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.367917 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnrt4\" (UniqueName: \"kubernetes.io/projected/efcf28af-d066-46a0-b547-5cd80cbb0fa6-kube-api-access-vnrt4\") pod \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.368754 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-swift-storage-0\") pod \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.368785 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-sb\") pod \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.368825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-nb\") pod \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.368934 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-config\") pod \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.369279 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-svc\") pod \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\" (UID: \"efcf28af-d066-46a0-b547-5cd80cbb0fa6\") " Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.377155 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efcf28af-d066-46a0-b547-5cd80cbb0fa6-kube-api-access-vnrt4" (OuterVolumeSpecName: "kube-api-access-vnrt4") pod "efcf28af-d066-46a0-b547-5cd80cbb0fa6" (UID: "efcf28af-d066-46a0-b547-5cd80cbb0fa6"). InnerVolumeSpecName "kube-api-access-vnrt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.437795 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "efcf28af-d066-46a0-b547-5cd80cbb0fa6" (UID: "efcf28af-d066-46a0-b547-5cd80cbb0fa6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.440424 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "efcf28af-d066-46a0-b547-5cd80cbb0fa6" (UID: "efcf28af-d066-46a0-b547-5cd80cbb0fa6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.474207 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnrt4\" (UniqueName: \"kubernetes.io/projected/efcf28af-d066-46a0-b547-5cd80cbb0fa6-kube-api-access-vnrt4\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.474240 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.474253 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.496249 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "efcf28af-d066-46a0-b547-5cd80cbb0fa6" (UID: "efcf28af-d066-46a0-b547-5cd80cbb0fa6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.520311 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-config" (OuterVolumeSpecName: "config") pod "efcf28af-d066-46a0-b547-5cd80cbb0fa6" (UID: "efcf28af-d066-46a0-b547-5cd80cbb0fa6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.536276 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "efcf28af-d066-46a0-b547-5cd80cbb0fa6" (UID: "efcf28af-d066-46a0-b547-5cd80cbb0fa6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.576299 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.576337 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:13 crc kubenswrapper[4846]: I0202 12:29:13.576386 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efcf28af-d066-46a0-b547-5cd80cbb0fa6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.311199 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-858jm" event={"ID":"e02dd0b8-6bd8-40b2-8a18-64d298c72a67","Type":"ContainerStarted","Data":"4fa5450ef3e8e9522271cc126cc1fcf00b43a3a35f39da799acac7233e310bfc"} Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.315180 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kd7sk" event={"ID":"e924adc7-fffe-4682-a601-cac1917e8f54","Type":"ContainerStarted","Data":"990eec9ad0d43277d54b187042320a08193aaca93bf2eb669cefbe31ab8568b3"} Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.319174 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wgp84" event={"ID":"c427a83f-c397-4886-bcd9-a7c992c7fba9","Type":"ContainerStarted","Data":"49a43bd07d83a9a7f59d576c12458101178e1ae9330d85850931baa6d139ed2f"} Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.348654 4846 generic.go:334] "Generic (PLEG): container finished" podID="ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" containerID="6beb6342e776aa866deb61277b624fe882c39aaa430b0686aa5d6ccf94528526" exitCode=0 Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.348736 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" event={"ID":"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3","Type":"ContainerDied","Data":"6beb6342e776aa866deb61277b624fe882c39aaa430b0686aa5d6ccf94528526"} Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.379965 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-kd7sk" podStartSLOduration=3.379945092 podStartE2EDuration="3.379945092s" podCreationTimestamp="2026-02-02 12:29:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:14.35910088 +0000 UTC m=+1185.587687743" watchObservedRunningTime="2026-02-02 12:29:14.379945092 +0000 UTC m=+1185.608531955" Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.403002 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" event={"ID":"efcf28af-d066-46a0-b547-5cd80cbb0fa6","Type":"ContainerDied","Data":"516a249500970cc9ae4b43a339b3085b9f02ff5e00da49b9e7340f7d87a5c945"} Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.403066 4846 scope.go:117] "RemoveContainer" containerID="71b2bf04a20293a34531d93bbc57a0e6142283523db7fb9dfbb50ae7af61ee55" Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.403250 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cc5c87cc-skcw8" Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.439536 4846 generic.go:334] "Generic (PLEG): container finished" podID="d60e3c65-fdf5-411b-970c-7c896d567c9a" containerID="79d7acbb015357d9b2c551a970ddd67e68e5d10d810684b2f3b6884b26b7dec4" exitCode=0 Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.439771 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866975879-w6pjj" event={"ID":"d60e3c65-fdf5-411b-970c-7c896d567c9a","Type":"ContainerDied","Data":"79d7acbb015357d9b2c551a970ddd67e68e5d10d810684b2f3b6884b26b7dec4"} Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.482154 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.537821 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wgp84" podStartSLOduration=3.537799246 podStartE2EDuration="3.537799246s" podCreationTimestamp="2026-02-02 12:29:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:14.496718777 +0000 UTC m=+1185.725305650" watchObservedRunningTime="2026-02-02 12:29:14.537799246 +0000 UTC m=+1185.766386109" Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.714420 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.755488 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.777526 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.804411 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:29:14 crc kubenswrapper[4846]: W0202 12:29:14.827526 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f2eb9d6_2305_46ff_93da_4e688a72ceb7.slice/crio-9e8e8e8c7817767e5eccd3a5fae3c605175e7ed55761331117011843c6685ca1 WatchSource:0}: Error finding container 9e8e8e8c7817767e5eccd3a5fae3c605175e7ed55761331117011843c6685ca1: Status 404 returned error can't find the container with id 9e8e8e8c7817767e5eccd3a5fae3c605175e7ed55761331117011843c6685ca1 Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.848953 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-skcw8"] Feb 02 12:29:14 crc kubenswrapper[4846]: I0202 12:29:14.875851 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54cc5c87cc-skcw8"] Feb 02 12:29:14 crc kubenswrapper[4846]: E0202 12:29:14.988902 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefcf28af_d066_46a0_b547_5cd80cbb0fa6.slice/crio-516a249500970cc9ae4b43a339b3085b9f02ff5e00da49b9e7340f7d87a5c945\": RecentStats: unable to find data in memory cache]" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.152685 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.346717 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-sb\") pod \"d60e3c65-fdf5-411b-970c-7c896d567c9a\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.346873 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-nb\") pod \"d60e3c65-fdf5-411b-970c-7c896d567c9a\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.346943 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9tzt\" (UniqueName: \"kubernetes.io/projected/d60e3c65-fdf5-411b-970c-7c896d567c9a-kube-api-access-t9tzt\") pod \"d60e3c65-fdf5-411b-970c-7c896d567c9a\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.347026 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-svc\") pod \"d60e3c65-fdf5-411b-970c-7c896d567c9a\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.347071 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-config\") pod \"d60e3c65-fdf5-411b-970c-7c896d567c9a\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.347093 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-swift-storage-0\") pod \"d60e3c65-fdf5-411b-970c-7c896d567c9a\" (UID: \"d60e3c65-fdf5-411b-970c-7c896d567c9a\") " Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.362397 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d60e3c65-fdf5-411b-970c-7c896d567c9a-kube-api-access-t9tzt" (OuterVolumeSpecName: "kube-api-access-t9tzt") pod "d60e3c65-fdf5-411b-970c-7c896d567c9a" (UID: "d60e3c65-fdf5-411b-970c-7c896d567c9a"). InnerVolumeSpecName "kube-api-access-t9tzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.379118 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d60e3c65-fdf5-411b-970c-7c896d567c9a" (UID: "d60e3c65-fdf5-411b-970c-7c896d567c9a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.389135 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d60e3c65-fdf5-411b-970c-7c896d567c9a" (UID: "d60e3c65-fdf5-411b-970c-7c896d567c9a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.389453 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d60e3c65-fdf5-411b-970c-7c896d567c9a" (UID: "d60e3c65-fdf5-411b-970c-7c896d567c9a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.402451 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d60e3c65-fdf5-411b-970c-7c896d567c9a" (UID: "d60e3c65-fdf5-411b-970c-7c896d567c9a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.409774 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-config" (OuterVolumeSpecName: "config") pod "d60e3c65-fdf5-411b-970c-7c896d567c9a" (UID: "d60e3c65-fdf5-411b-970c-7c896d567c9a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.449419 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.449493 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9tzt\" (UniqueName: \"kubernetes.io/projected/d60e3c65-fdf5-411b-970c-7c896d567c9a-kube-api-access-t9tzt\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.449508 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.449518 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.449530 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.449540 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d60e3c65-fdf5-411b-970c-7c896d567c9a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.459061 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efcf28af-d066-46a0-b547-5cd80cbb0fa6" path="/var/lib/kubelet/pods/efcf28af-d066-46a0-b547-5cd80cbb0fa6/volumes" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.462558 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" event={"ID":"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3","Type":"ContainerStarted","Data":"676abdb2a9bef4899561705a6f61a832b0d5b2934cf37d0d5b2f5f8287ea8d40"} Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.462670 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.485121 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" podStartSLOduration=4.485102566 podStartE2EDuration="4.485102566s" podCreationTimestamp="2026-02-02 12:29:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:15.48288579 +0000 UTC m=+1186.711472663" watchObservedRunningTime="2026-02-02 12:29:15.485102566 +0000 UTC m=+1186.713689439" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.497510 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866975879-w6pjj" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.497727 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866975879-w6pjj" event={"ID":"d60e3c65-fdf5-411b-970c-7c896d567c9a","Type":"ContainerDied","Data":"13029f162c79d9fae2f4a8057ec65c9f54d32c005653a98972c65409caf74682"} Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.497800 4846 scope.go:117] "RemoveContainer" containerID="79d7acbb015357d9b2c551a970ddd67e68e5d10d810684b2f3b6884b26b7dec4" Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.508951 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f2eb9d6-2305-46ff-93da-4e688a72ceb7","Type":"ContainerStarted","Data":"9e8e8e8c7817767e5eccd3a5fae3c605175e7ed55761331117011843c6685ca1"} Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.515292 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a49cfd-e78e-44c0-a26c-116ec72c84c8","Type":"ContainerStarted","Data":"343f4a2f392e2f4e19e9aed9bee3c85ddb017e6705d291213c05daf2d3c93ec6"} Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.605557 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866975879-w6pjj"] Feb 02 12:29:15 crc kubenswrapper[4846]: I0202 12:29:15.615603 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-866975879-w6pjj"] Feb 02 12:29:16 crc kubenswrapper[4846]: I0202 12:29:16.529551 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a49cfd-e78e-44c0-a26c-116ec72c84c8","Type":"ContainerStarted","Data":"a86140d8729857f89e7fbef9961b2e9f9b729073991986b6ddcace48b0e374fc"} Feb 02 12:29:16 crc kubenswrapper[4846]: I0202 12:29:16.534180 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f2eb9d6-2305-46ff-93da-4e688a72ceb7","Type":"ContainerStarted","Data":"6843e6f1486847657fd8e1d4ee31744e65ab164d4e1091da10feebdec38930d6"} Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.437317 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d60e3c65-fdf5-411b-970c-7c896d567c9a" path="/var/lib/kubelet/pods/d60e3c65-fdf5-411b-970c-7c896d567c9a/volumes" Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.549496 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f2eb9d6-2305-46ff-93da-4e688a72ceb7","Type":"ContainerStarted","Data":"df18efa3e1fe04de874837096c58135814e4a90fa95c6b0934e1d05343f2bad1"} Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.549556 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" containerName="glance-log" containerID="cri-o://6843e6f1486847657fd8e1d4ee31744e65ab164d4e1091da10feebdec38930d6" gracePeriod=30 Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.549679 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" containerName="glance-httpd" containerID="cri-o://df18efa3e1fe04de874837096c58135814e4a90fa95c6b0934e1d05343f2bad1" gracePeriod=30 Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.551911 4846 generic.go:334] "Generic (PLEG): container finished" podID="c427a83f-c397-4886-bcd9-a7c992c7fba9" containerID="49a43bd07d83a9a7f59d576c12458101178e1ae9330d85850931baa6d139ed2f" exitCode=0 Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.551980 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wgp84" event={"ID":"c427a83f-c397-4886-bcd9-a7c992c7fba9","Type":"ContainerDied","Data":"49a43bd07d83a9a7f59d576c12458101178e1ae9330d85850931baa6d139ed2f"} Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.554577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a49cfd-e78e-44c0-a26c-116ec72c84c8","Type":"ContainerStarted","Data":"ffa25c7fef4304c1a8eab5eaed582b347eb9f18d0ea1144e4528868849ae6d30"} Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.554893 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" containerName="glance-log" containerID="cri-o://a86140d8729857f89e7fbef9961b2e9f9b729073991986b6ddcace48b0e374fc" gracePeriod=30 Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.555083 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" containerName="glance-httpd" containerID="cri-o://ffa25c7fef4304c1a8eab5eaed582b347eb9f18d0ea1144e4528868849ae6d30" gracePeriod=30 Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.614142 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.614120938 podStartE2EDuration="6.614120938s" podCreationTimestamp="2026-02-02 12:29:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:17.581247644 +0000 UTC m=+1188.809834517" watchObservedRunningTime="2026-02-02 12:29:17.614120938 +0000 UTC m=+1188.842707821" Feb 02 12:29:17 crc kubenswrapper[4846]: I0202 12:29:17.638704 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.638679042 podStartE2EDuration="6.638679042s" podCreationTimestamp="2026-02-02 12:29:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:17.623152894 +0000 UTC m=+1188.851739757" watchObservedRunningTime="2026-02-02 12:29:17.638679042 +0000 UTC m=+1188.867265905" Feb 02 12:29:18 crc kubenswrapper[4846]: I0202 12:29:18.574141 4846 generic.go:334] "Generic (PLEG): container finished" podID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" containerID="df18efa3e1fe04de874837096c58135814e4a90fa95c6b0934e1d05343f2bad1" exitCode=0 Feb 02 12:29:18 crc kubenswrapper[4846]: I0202 12:29:18.574413 4846 generic.go:334] "Generic (PLEG): container finished" podID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" containerID="6843e6f1486847657fd8e1d4ee31744e65ab164d4e1091da10feebdec38930d6" exitCode=143 Feb 02 12:29:18 crc kubenswrapper[4846]: I0202 12:29:18.574453 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f2eb9d6-2305-46ff-93da-4e688a72ceb7","Type":"ContainerDied","Data":"df18efa3e1fe04de874837096c58135814e4a90fa95c6b0934e1d05343f2bad1"} Feb 02 12:29:18 crc kubenswrapper[4846]: I0202 12:29:18.574478 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f2eb9d6-2305-46ff-93da-4e688a72ceb7","Type":"ContainerDied","Data":"6843e6f1486847657fd8e1d4ee31744e65ab164d4e1091da10feebdec38930d6"} Feb 02 12:29:18 crc kubenswrapper[4846]: I0202 12:29:18.576210 4846 generic.go:334] "Generic (PLEG): container finished" podID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" containerID="ffa25c7fef4304c1a8eab5eaed582b347eb9f18d0ea1144e4528868849ae6d30" exitCode=0 Feb 02 12:29:18 crc kubenswrapper[4846]: I0202 12:29:18.576237 4846 generic.go:334] "Generic (PLEG): container finished" podID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" containerID="a86140d8729857f89e7fbef9961b2e9f9b729073991986b6ddcace48b0e374fc" exitCode=143 Feb 02 12:29:18 crc kubenswrapper[4846]: I0202 12:29:18.576417 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a49cfd-e78e-44c0-a26c-116ec72c84c8","Type":"ContainerDied","Data":"ffa25c7fef4304c1a8eab5eaed582b347eb9f18d0ea1144e4528868849ae6d30"} Feb 02 12:29:18 crc kubenswrapper[4846]: I0202 12:29:18.576449 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a49cfd-e78e-44c0-a26c-116ec72c84c8","Type":"ContainerDied","Data":"a86140d8729857f89e7fbef9961b2e9f9b729073991986b6ddcace48b0e374fc"} Feb 02 12:29:22 crc kubenswrapper[4846]: I0202 12:29:22.454906 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:22 crc kubenswrapper[4846]: I0202 12:29:22.523014 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-hzrlk"] Feb 02 12:29:22 crc kubenswrapper[4846]: I0202 12:29:22.523502 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" podUID="240e8c55-ad59-4bee-a92e-93876302d7f5" containerName="dnsmasq-dns" containerID="cri-o://6e21400ac8e3c4e79243ecc8b3b8272649285daca0a5f4b7d6f4cc9592ac49c6" gracePeriod=10 Feb 02 12:29:23 crc kubenswrapper[4846]: I0202 12:29:23.626040 4846 generic.go:334] "Generic (PLEG): container finished" podID="240e8c55-ad59-4bee-a92e-93876302d7f5" containerID="6e21400ac8e3c4e79243ecc8b3b8272649285daca0a5f4b7d6f4cc9592ac49c6" exitCode=0 Feb 02 12:29:23 crc kubenswrapper[4846]: I0202 12:29:23.626106 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" event={"ID":"240e8c55-ad59-4bee-a92e-93876302d7f5","Type":"ContainerDied","Data":"6e21400ac8e3c4e79243ecc8b3b8272649285daca0a5f4b7d6f4cc9592ac49c6"} Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.030214 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.130649 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-config-data\") pod \"c427a83f-c397-4886-bcd9-a7c992c7fba9\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.131394 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-combined-ca-bundle\") pod \"c427a83f-c397-4886-bcd9-a7c992c7fba9\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.131541 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-scripts\") pod \"c427a83f-c397-4886-bcd9-a7c992c7fba9\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.131631 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-fernet-keys\") pod \"c427a83f-c397-4886-bcd9-a7c992c7fba9\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.131677 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-credential-keys\") pod \"c427a83f-c397-4886-bcd9-a7c992c7fba9\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.131825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pwnt\" (UniqueName: \"kubernetes.io/projected/c427a83f-c397-4886-bcd9-a7c992c7fba9-kube-api-access-9pwnt\") pod \"c427a83f-c397-4886-bcd9-a7c992c7fba9\" (UID: \"c427a83f-c397-4886-bcd9-a7c992c7fba9\") " Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.137433 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-scripts" (OuterVolumeSpecName: "scripts") pod "c427a83f-c397-4886-bcd9-a7c992c7fba9" (UID: "c427a83f-c397-4886-bcd9-a7c992c7fba9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.137467 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c427a83f-c397-4886-bcd9-a7c992c7fba9" (UID: "c427a83f-c397-4886-bcd9-a7c992c7fba9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.146854 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c427a83f-c397-4886-bcd9-a7c992c7fba9" (UID: "c427a83f-c397-4886-bcd9-a7c992c7fba9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.146921 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c427a83f-c397-4886-bcd9-a7c992c7fba9-kube-api-access-9pwnt" (OuterVolumeSpecName: "kube-api-access-9pwnt") pod "c427a83f-c397-4886-bcd9-a7c992c7fba9" (UID: "c427a83f-c397-4886-bcd9-a7c992c7fba9"). InnerVolumeSpecName "kube-api-access-9pwnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.159892 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c427a83f-c397-4886-bcd9-a7c992c7fba9" (UID: "c427a83f-c397-4886-bcd9-a7c992c7fba9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.173131 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-config-data" (OuterVolumeSpecName: "config-data") pod "c427a83f-c397-4886-bcd9-a7c992c7fba9" (UID: "c427a83f-c397-4886-bcd9-a7c992c7fba9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.233501 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.233535 4846 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.233544 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pwnt\" (UniqueName: \"kubernetes.io/projected/c427a83f-c397-4886-bcd9-a7c992c7fba9-kube-api-access-9pwnt\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.233554 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.233563 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.233571 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c427a83f-c397-4886-bcd9-a7c992c7fba9-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.635527 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wgp84" event={"ID":"c427a83f-c397-4886-bcd9-a7c992c7fba9","Type":"ContainerDied","Data":"6755de517414e72bc7274a2791bbb44228da7566761219b74c6d5baca0a1ce22"} Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.635564 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6755de517414e72bc7274a2791bbb44228da7566761219b74c6d5baca0a1ce22" Feb 02 12:29:24 crc kubenswrapper[4846]: I0202 12:29:24.635614 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wgp84" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.120814 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wgp84"] Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.126760 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wgp84"] Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.245225 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-m2dtn"] Feb 02 12:29:25 crc kubenswrapper[4846]: E0202 12:29:25.245728 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efcf28af-d066-46a0-b547-5cd80cbb0fa6" containerName="init" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.245749 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="efcf28af-d066-46a0-b547-5cd80cbb0fa6" containerName="init" Feb 02 12:29:25 crc kubenswrapper[4846]: E0202 12:29:25.245762 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c427a83f-c397-4886-bcd9-a7c992c7fba9" containerName="keystone-bootstrap" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.245769 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c427a83f-c397-4886-bcd9-a7c992c7fba9" containerName="keystone-bootstrap" Feb 02 12:29:25 crc kubenswrapper[4846]: E0202 12:29:25.245800 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d60e3c65-fdf5-411b-970c-7c896d567c9a" containerName="init" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.245806 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d60e3c65-fdf5-411b-970c-7c896d567c9a" containerName="init" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.245971 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c427a83f-c397-4886-bcd9-a7c992c7fba9" containerName="keystone-bootstrap" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.245982 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d60e3c65-fdf5-411b-970c-7c896d567c9a" containerName="init" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.245990 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="efcf28af-d066-46a0-b547-5cd80cbb0fa6" containerName="init" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.246678 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.250082 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9wp9t" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.250105 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.250821 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.250869 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-combined-ca-bundle\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.250939 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-credential-keys\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.250981 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-config-data\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.251023 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-fernet-keys\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.251047 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hpvm\" (UniqueName: \"kubernetes.io/projected/5812b4fc-8c66-418f-8886-0ff26018b905-kube-api-access-8hpvm\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.251058 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.251141 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-scripts\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.251567 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.279159 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-m2dtn"] Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.351674 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hpvm\" (UniqueName: \"kubernetes.io/projected/5812b4fc-8c66-418f-8886-0ff26018b905-kube-api-access-8hpvm\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.351725 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-fernet-keys\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.351789 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-scripts\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.351836 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-combined-ca-bundle\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.351857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-credential-keys\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.351879 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-config-data\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.358651 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-config-data\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.359134 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-credential-keys\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.359146 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-combined-ca-bundle\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.359786 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-scripts\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.367990 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hpvm\" (UniqueName: \"kubernetes.io/projected/5812b4fc-8c66-418f-8886-0ff26018b905-kube-api-access-8hpvm\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.376794 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-fernet-keys\") pod \"keystone-bootstrap-m2dtn\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.434886 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c427a83f-c397-4886-bcd9-a7c992c7fba9" path="/var/lib/kubelet/pods/c427a83f-c397-4886-bcd9-a7c992c7fba9/volumes" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.574211 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:25 crc kubenswrapper[4846]: I0202 12:29:25.750307 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" podUID="240e8c55-ad59-4bee-a92e-93876302d7f5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Feb 02 12:29:30 crc kubenswrapper[4846]: I0202 12:29:30.750203 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" podUID="240e8c55-ad59-4bee-a92e-93876302d7f5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Feb 02 12:29:30 crc kubenswrapper[4846]: I0202 12:29:30.966678 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:30 crc kubenswrapper[4846]: I0202 12:29:30.974957 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.067592 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-httpd-run\") pod \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.067949 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-combined-ca-bundle\") pod \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.067968 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.067987 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-logs\") pod \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.068011 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-scripts\") pod \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.068040 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.068096 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-scripts\") pod \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.068123 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-logs\") pod \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.068138 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-config-data\") pod \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.068185 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kbnc\" (UniqueName: \"kubernetes.io/projected/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-kube-api-access-2kbnc\") pod \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.068240 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-config-data\") pod \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.068264 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-httpd-run\") pod \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.068311 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-combined-ca-bundle\") pod \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\" (UID: \"8f2eb9d6-2305-46ff-93da-4e688a72ceb7\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.068329 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n66d\" (UniqueName: \"kubernetes.io/projected/00a49cfd-e78e-44c0-a26c-116ec72c84c8-kube-api-access-8n66d\") pod \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\" (UID: \"00a49cfd-e78e-44c0-a26c-116ec72c84c8\") " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.071029 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "00a49cfd-e78e-44c0-a26c-116ec72c84c8" (UID: "00a49cfd-e78e-44c0-a26c-116ec72c84c8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.071150 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-logs" (OuterVolumeSpecName: "logs") pod "8f2eb9d6-2305-46ff-93da-4e688a72ceb7" (UID: "8f2eb9d6-2305-46ff-93da-4e688a72ceb7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.071815 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-logs" (OuterVolumeSpecName: "logs") pod "00a49cfd-e78e-44c0-a26c-116ec72c84c8" (UID: "00a49cfd-e78e-44c0-a26c-116ec72c84c8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.072178 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8f2eb9d6-2305-46ff-93da-4e688a72ceb7" (UID: "8f2eb9d6-2305-46ff-93da-4e688a72ceb7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.075450 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "8f2eb9d6-2305-46ff-93da-4e688a72ceb7" (UID: "8f2eb9d6-2305-46ff-93da-4e688a72ceb7"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.075488 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-scripts" (OuterVolumeSpecName: "scripts") pod "00a49cfd-e78e-44c0-a26c-116ec72c84c8" (UID: "00a49cfd-e78e-44c0-a26c-116ec72c84c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.076380 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "00a49cfd-e78e-44c0-a26c-116ec72c84c8" (UID: "00a49cfd-e78e-44c0-a26c-116ec72c84c8"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.077804 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a49cfd-e78e-44c0-a26c-116ec72c84c8-kube-api-access-8n66d" (OuterVolumeSpecName: "kube-api-access-8n66d") pod "00a49cfd-e78e-44c0-a26c-116ec72c84c8" (UID: "00a49cfd-e78e-44c0-a26c-116ec72c84c8"). InnerVolumeSpecName "kube-api-access-8n66d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.091182 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-scripts" (OuterVolumeSpecName: "scripts") pod "8f2eb9d6-2305-46ff-93da-4e688a72ceb7" (UID: "8f2eb9d6-2305-46ff-93da-4e688a72ceb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.091375 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-kube-api-access-2kbnc" (OuterVolumeSpecName: "kube-api-access-2kbnc") pod "8f2eb9d6-2305-46ff-93da-4e688a72ceb7" (UID: "8f2eb9d6-2305-46ff-93da-4e688a72ceb7"). InnerVolumeSpecName "kube-api-access-2kbnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.103751 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f2eb9d6-2305-46ff-93da-4e688a72ceb7" (UID: "8f2eb9d6-2305-46ff-93da-4e688a72ceb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.109091 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00a49cfd-e78e-44c0-a26c-116ec72c84c8" (UID: "00a49cfd-e78e-44c0-a26c-116ec72c84c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.120192 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-config-data" (OuterVolumeSpecName: "config-data") pod "8f2eb9d6-2305-46ff-93da-4e688a72ceb7" (UID: "8f2eb9d6-2305-46ff-93da-4e688a72ceb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.126375 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-config-data" (OuterVolumeSpecName: "config-data") pod "00a49cfd-e78e-44c0-a26c-116ec72c84c8" (UID: "00a49cfd-e78e-44c0-a26c-116ec72c84c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170525 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170556 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170566 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170577 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n66d\" (UniqueName: \"kubernetes.io/projected/00a49cfd-e78e-44c0-a26c-116ec72c84c8-kube-api-access-8n66d\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170586 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170594 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170640 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170650 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00a49cfd-e78e-44c0-a26c-116ec72c84c8-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170658 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170671 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170678 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170686 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170695 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a49cfd-e78e-44c0-a26c-116ec72c84c8-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.170704 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kbnc\" (UniqueName: \"kubernetes.io/projected/8f2eb9d6-2305-46ff-93da-4e688a72ceb7-kube-api-access-2kbnc\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.187183 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.188813 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.271817 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.271851 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:31 crc kubenswrapper[4846]: E0202 12:29:31.457568 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:a5f8855b2ed00a661ac827cc3908e540ed2327354ac5a1d39491f4507237b4ec" Feb 02 12:29:31 crc kubenswrapper[4846]: E0202 12:29:31.457720 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:a5f8855b2ed00a661ac827cc3908e540ed2327354ac5a1d39491f4507237b4ec,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s5qsx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-5vmgj_openstack(18307d58-f249-4ec0-8556-fd98e4febb66): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 12:29:31 crc kubenswrapper[4846]: E0202 12:29:31.458950 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-5vmgj" podUID="18307d58-f249-4ec0-8556-fd98e4febb66" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.707420 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.707457 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f2eb9d6-2305-46ff-93da-4e688a72ceb7","Type":"ContainerDied","Data":"9e8e8e8c7817767e5eccd3a5fae3c605175e7ed55761331117011843c6685ca1"} Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.707518 4846 scope.go:117] "RemoveContainer" containerID="df18efa3e1fe04de874837096c58135814e4a90fa95c6b0934e1d05343f2bad1" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.712268 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.712263 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"00a49cfd-e78e-44c0-a26c-116ec72c84c8","Type":"ContainerDied","Data":"343f4a2f392e2f4e19e9aed9bee3c85ddb017e6705d291213c05daf2d3c93ec6"} Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.735564 4846 generic.go:334] "Generic (PLEG): container finished" podID="e924adc7-fffe-4682-a601-cac1917e8f54" containerID="990eec9ad0d43277d54b187042320a08193aaca93bf2eb669cefbe31ab8568b3" exitCode=0 Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.737052 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kd7sk" event={"ID":"e924adc7-fffe-4682-a601-cac1917e8f54","Type":"ContainerDied","Data":"990eec9ad0d43277d54b187042320a08193aaca93bf2eb669cefbe31ab8568b3"} Feb 02 12:29:31 crc kubenswrapper[4846]: E0202 12:29:31.738164 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:a5f8855b2ed00a661ac827cc3908e540ed2327354ac5a1d39491f4507237b4ec\\\"\"" pod="openstack/barbican-db-sync-5vmgj" podUID="18307d58-f249-4ec0-8556-fd98e4febb66" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.775694 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.820814 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.838865 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.849107 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.858208 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:29:31 crc kubenswrapper[4846]: E0202 12:29:31.858532 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" containerName="glance-httpd" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.858555 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" containerName="glance-httpd" Feb 02 12:29:31 crc kubenswrapper[4846]: E0202 12:29:31.858577 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" containerName="glance-log" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.858586 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" containerName="glance-log" Feb 02 12:29:31 crc kubenswrapper[4846]: E0202 12:29:31.858594 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" containerName="glance-log" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.858602 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" containerName="glance-log" Feb 02 12:29:31 crc kubenswrapper[4846]: E0202 12:29:31.858640 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" containerName="glance-httpd" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.858648 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" containerName="glance-httpd" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.858860 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" containerName="glance-log" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.858886 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" containerName="glance-httpd" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.858894 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" containerName="glance-log" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.858903 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" containerName="glance-httpd" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.859785 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.862956 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-c2sk4" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.863113 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.863390 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.863543 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.868170 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.879481 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.881542 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.883958 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.884156 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.888676 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982035 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982086 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982213 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4zxn\" (UniqueName: \"kubernetes.io/projected/d51c1b58-d44f-44cd-a513-af4fb272c0a7-kube-api-access-n4zxn\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982296 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982359 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-logs\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982408 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982456 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982483 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-scripts\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982517 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982571 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982635 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqjv9\" (UniqueName: \"kubernetes.io/projected/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-kube-api-access-pqjv9\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982676 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-logs\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982802 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982863 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982920 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:31 crc kubenswrapper[4846]: I0202 12:29:31.982952 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-config-data\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.084690 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.084853 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.084932 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085000 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-config-data\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085044 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085115 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085153 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4zxn\" (UniqueName: \"kubernetes.io/projected/d51c1b58-d44f-44cd-a513-af4fb272c0a7-kube-api-access-n4zxn\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085204 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085238 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-logs\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085270 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085294 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085312 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-scripts\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085332 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085360 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085376 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqjv9\" (UniqueName: \"kubernetes.io/projected/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-kube-api-access-pqjv9\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.085399 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-logs\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.086367 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.086915 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-logs\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.086946 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.087403 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.087794 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.090542 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-logs\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.090686 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.091134 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-scripts\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.092364 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.092584 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-config-data\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.092965 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.095444 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.102424 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.102419 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.106513 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqjv9\" (UniqueName: \"kubernetes.io/projected/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-kube-api-access-pqjv9\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.115585 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4zxn\" (UniqueName: \"kubernetes.io/projected/d51c1b58-d44f-44cd-a513-af4fb272c0a7-kube-api-access-n4zxn\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.125970 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.126006 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.176185 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.200976 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:29:32 crc kubenswrapper[4846]: E0202 12:29:32.740617 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:3fa6e687aa002b92fedbfe2c1ccaa2906b399c58d17bf9ecece2c4cd69a0210b" Feb 02 12:29:32 crc kubenswrapper[4846]: E0202 12:29:32.740773 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:3fa6e687aa002b92fedbfe2c1ccaa2906b399c58d17bf9ecece2c4cd69a0210b,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zdklw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-2z995_openstack(8b58cb4f-6d15-4d97-871d-724c160ec765): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 12:29:32 crc kubenswrapper[4846]: E0202 12:29:32.741850 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-2z995" podUID="8b58cb4f-6d15-4d97-871d-724c160ec765" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.753852 4846 scope.go:117] "RemoveContainer" containerID="6843e6f1486847657fd8e1d4ee31744e65ab164d4e1091da10feebdec38930d6" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.755086 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" event={"ID":"240e8c55-ad59-4bee-a92e-93876302d7f5","Type":"ContainerDied","Data":"41f357ff8f68cee95cec3690836cfc9d309b40430d83a1ef97613e78e69f6740"} Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.755148 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41f357ff8f68cee95cec3690836cfc9d309b40430d83a1ef97613e78e69f6740" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.883474 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.918366 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvg2z\" (UniqueName: \"kubernetes.io/projected/240e8c55-ad59-4bee-a92e-93876302d7f5-kube-api-access-hvg2z\") pod \"240e8c55-ad59-4bee-a92e-93876302d7f5\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.918414 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-config\") pod \"240e8c55-ad59-4bee-a92e-93876302d7f5\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.918506 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-dns-svc\") pod \"240e8c55-ad59-4bee-a92e-93876302d7f5\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.918678 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-nb\") pod \"240e8c55-ad59-4bee-a92e-93876302d7f5\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.918760 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-sb\") pod \"240e8c55-ad59-4bee-a92e-93876302d7f5\" (UID: \"240e8c55-ad59-4bee-a92e-93876302d7f5\") " Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.928531 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/240e8c55-ad59-4bee-a92e-93876302d7f5-kube-api-access-hvg2z" (OuterVolumeSpecName: "kube-api-access-hvg2z") pod "240e8c55-ad59-4bee-a92e-93876302d7f5" (UID: "240e8c55-ad59-4bee-a92e-93876302d7f5"). InnerVolumeSpecName "kube-api-access-hvg2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.933999 4846 scope.go:117] "RemoveContainer" containerID="ffa25c7fef4304c1a8eab5eaed582b347eb9f18d0ea1144e4528868849ae6d30" Feb 02 12:29:32 crc kubenswrapper[4846]: I0202 12:29:32.987172 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "240e8c55-ad59-4bee-a92e-93876302d7f5" (UID: "240e8c55-ad59-4bee-a92e-93876302d7f5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.002554 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-config" (OuterVolumeSpecName: "config") pod "240e8c55-ad59-4bee-a92e-93876302d7f5" (UID: "240e8c55-ad59-4bee-a92e-93876302d7f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.006602 4846 scope.go:117] "RemoveContainer" containerID="a86140d8729857f89e7fbef9961b2e9f9b729073991986b6ddcace48b0e374fc" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.020554 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.021113 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvg2z\" (UniqueName: \"kubernetes.io/projected/240e8c55-ad59-4bee-a92e-93876302d7f5-kube-api-access-hvg2z\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.021130 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.029432 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "240e8c55-ad59-4bee-a92e-93876302d7f5" (UID: "240e8c55-ad59-4bee-a92e-93876302d7f5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.035455 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "240e8c55-ad59-4bee-a92e-93876302d7f5" (UID: "240e8c55-ad59-4bee-a92e-93876302d7f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.083789 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.121960 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-config\") pod \"e924adc7-fffe-4682-a601-cac1917e8f54\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.122074 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-combined-ca-bundle\") pod \"e924adc7-fffe-4682-a601-cac1917e8f54\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.122103 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hm5v\" (UniqueName: \"kubernetes.io/projected/e924adc7-fffe-4682-a601-cac1917e8f54-kube-api-access-9hm5v\") pod \"e924adc7-fffe-4682-a601-cac1917e8f54\" (UID: \"e924adc7-fffe-4682-a601-cac1917e8f54\") " Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.122565 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.122582 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/240e8c55-ad59-4bee-a92e-93876302d7f5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.133738 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e924adc7-fffe-4682-a601-cac1917e8f54-kube-api-access-9hm5v" (OuterVolumeSpecName: "kube-api-access-9hm5v") pod "e924adc7-fffe-4682-a601-cac1917e8f54" (UID: "e924adc7-fffe-4682-a601-cac1917e8f54"). InnerVolumeSpecName "kube-api-access-9hm5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.153873 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e924adc7-fffe-4682-a601-cac1917e8f54" (UID: "e924adc7-fffe-4682-a601-cac1917e8f54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.175215 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-config" (OuterVolumeSpecName: "config") pod "e924adc7-fffe-4682-a601-cac1917e8f54" (UID: "e924adc7-fffe-4682-a601-cac1917e8f54"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.224532 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.224561 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e924adc7-fffe-4682-a601-cac1917e8f54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.224570 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hm5v\" (UniqueName: \"kubernetes.io/projected/e924adc7-fffe-4682-a601-cac1917e8f54-kube-api-access-9hm5v\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.269598 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-m2dtn"] Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.278373 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.340546 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:29:33 crc kubenswrapper[4846]: W0202 12:29:33.345677 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd51c1b58_d44f_44cd_a513_af4fb272c0a7.slice/crio-d087ebbc17840f60e2cfa79c52c93f604e603a562b25bcde4928405b0199c93a WatchSource:0}: Error finding container d087ebbc17840f60e2cfa79c52c93f604e603a562b25bcde4928405b0199c93a: Status 404 returned error can't find the container with id d087ebbc17840f60e2cfa79c52c93f604e603a562b25bcde4928405b0199c93a Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.435850 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a49cfd-e78e-44c0-a26c-116ec72c84c8" path="/var/lib/kubelet/pods/00a49cfd-e78e-44c0-a26c-116ec72c84c8/volumes" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.437301 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f2eb9d6-2305-46ff-93da-4e688a72ceb7" path="/var/lib/kubelet/pods/8f2eb9d6-2305-46ff-93da-4e688a72ceb7/volumes" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.497581 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:29:33 crc kubenswrapper[4846]: W0202 12:29:33.502829 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3be8e6d_863b_49ba_b177_5ee2bb76ba51.slice/crio-7a0837b7c2c6d34b2e66db2be02cd58c4b57ec9bf10d0675de223e0bbf85adc7 WatchSource:0}: Error finding container 7a0837b7c2c6d34b2e66db2be02cd58c4b57ec9bf10d0675de223e0bbf85adc7: Status 404 returned error can't find the container with id 7a0837b7c2c6d34b2e66db2be02cd58c4b57ec9bf10d0675de223e0bbf85adc7 Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.771310 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-858jm" event={"ID":"e02dd0b8-6bd8-40b2-8a18-64d298c72a67","Type":"ContainerStarted","Data":"2acbbb7c816156b529232e467d5fe13563d39a52756e14b37ed7aa312ee11088"} Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.788961 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaece0f9-90e4-4f49-b5e4-6566e68f115a","Type":"ContainerStarted","Data":"06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4"} Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.791676 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-858jm" podStartSLOduration=3.335676962 podStartE2EDuration="22.791656981s" podCreationTimestamp="2026-02-02 12:29:11 +0000 UTC" firstStartedPulling="2026-02-02 12:29:13.253283819 +0000 UTC m=+1184.481870682" lastFinishedPulling="2026-02-02 12:29:32.709263838 +0000 UTC m=+1203.937850701" observedRunningTime="2026-02-02 12:29:33.78838386 +0000 UTC m=+1205.016970743" watchObservedRunningTime="2026-02-02 12:29:33.791656981 +0000 UTC m=+1205.020243864" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.794838 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3be8e6d-863b-49ba-b177-5ee2bb76ba51","Type":"ContainerStarted","Data":"7a0837b7c2c6d34b2e66db2be02cd58c4b57ec9bf10d0675de223e0bbf85adc7"} Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.797282 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d51c1b58-d44f-44cd-a513-af4fb272c0a7","Type":"ContainerStarted","Data":"d087ebbc17840f60e2cfa79c52c93f604e603a562b25bcde4928405b0199c93a"} Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.800113 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m2dtn" event={"ID":"5812b4fc-8c66-418f-8886-0ff26018b905","Type":"ContainerStarted","Data":"61857b150929bc24955da85821ee7c4e71280ebcd4d2731ae05aa0b38a82d290"} Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.800183 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m2dtn" event={"ID":"5812b4fc-8c66-418f-8886-0ff26018b905","Type":"ContainerStarted","Data":"702aa68a164f5867839ab6a9b6bd9b4cc98c3b3f803fc69a69da399385cab4d3"} Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.804084 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kd7sk" event={"ID":"e924adc7-fffe-4682-a601-cac1917e8f54","Type":"ContainerDied","Data":"15c8e863f6f4492ca509df9d7bcfd91200fc72a6997959e2cbfb9fb50594252f"} Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.804137 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15c8e863f6f4492ca509df9d7bcfd91200fc72a6997959e2cbfb9fb50594252f" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.804234 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kd7sk" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.804882 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b577f8c-hzrlk" Feb 02 12:29:33 crc kubenswrapper[4846]: E0202 12:29:33.806999 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:3fa6e687aa002b92fedbfe2c1ccaa2906b399c58d17bf9ecece2c4cd69a0210b\\\"\"" pod="openstack/cinder-db-sync-2z995" podUID="8b58cb4f-6d15-4d97-871d-724c160ec765" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.816995 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-m2dtn" podStartSLOduration=8.816970615 podStartE2EDuration="8.816970615s" podCreationTimestamp="2026-02-02 12:29:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:33.81516393 +0000 UTC m=+1205.043750793" watchObservedRunningTime="2026-02-02 12:29:33.816970615 +0000 UTC m=+1205.045557478" Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.884270 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-hzrlk"] Feb 02 12:29:33 crc kubenswrapper[4846]: I0202 12:29:33.892743 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66b577f8c-hzrlk"] Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.073753 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-9tl7z"] Feb 02 12:29:34 crc kubenswrapper[4846]: E0202 12:29:34.074492 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="240e8c55-ad59-4bee-a92e-93876302d7f5" containerName="init" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.074507 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="240e8c55-ad59-4bee-a92e-93876302d7f5" containerName="init" Feb 02 12:29:34 crc kubenswrapper[4846]: E0202 12:29:34.074529 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e924adc7-fffe-4682-a601-cac1917e8f54" containerName="neutron-db-sync" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.074538 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e924adc7-fffe-4682-a601-cac1917e8f54" containerName="neutron-db-sync" Feb 02 12:29:34 crc kubenswrapper[4846]: E0202 12:29:34.074575 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="240e8c55-ad59-4bee-a92e-93876302d7f5" containerName="dnsmasq-dns" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.074584 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="240e8c55-ad59-4bee-a92e-93876302d7f5" containerName="dnsmasq-dns" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.074832 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e924adc7-fffe-4682-a601-cac1917e8f54" containerName="neutron-db-sync" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.074862 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="240e8c55-ad59-4bee-a92e-93876302d7f5" containerName="dnsmasq-dns" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.075970 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.082922 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-9tl7z"] Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.166383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-config\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.166457 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbjxg\" (UniqueName: \"kubernetes.io/projected/832895eb-b40b-427c-9485-166c86ef9d83-kube-api-access-rbjxg\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.166477 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-sb\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.166499 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-swift-storage-0\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.170788 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-svc\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.170862 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-nb\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.239950 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5564fdb9d4-xt6f7"] Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.241493 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.250136 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.250337 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nx5gj" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.250636 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.250922 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.261275 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5564fdb9d4-xt6f7"] Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.274231 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbjxg\" (UniqueName: \"kubernetes.io/projected/832895eb-b40b-427c-9485-166c86ef9d83-kube-api-access-rbjxg\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.274325 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-sb\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.275827 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-swift-storage-0\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.276105 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-svc\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.276153 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-nb\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.276232 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-config\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.277548 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-config\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.278592 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-sb\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.279765 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-nb\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.280446 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-svc\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.280559 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-swift-storage-0\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.308278 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbjxg\" (UniqueName: \"kubernetes.io/projected/832895eb-b40b-427c-9485-166c86ef9d83-kube-api-access-rbjxg\") pod \"dnsmasq-dns-86dd6887f5-9tl7z\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.379315 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-httpd-config\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.379384 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-combined-ca-bundle\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.379426 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvm6k\" (UniqueName: \"kubernetes.io/projected/447fc9d5-7789-43d6-8d11-d5ef5b00c371-kube-api-access-tvm6k\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.379513 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-ovndb-tls-certs\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.379561 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-config\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.481180 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-ovndb-tls-certs\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.481286 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-config\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.481380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-httpd-config\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.481443 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-combined-ca-bundle\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.481512 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvm6k\" (UniqueName: \"kubernetes.io/projected/447fc9d5-7789-43d6-8d11-d5ef5b00c371-kube-api-access-tvm6k\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.486362 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-combined-ca-bundle\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.489394 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-httpd-config\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.492227 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-ovndb-tls-certs\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.492653 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-config\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.504168 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvm6k\" (UniqueName: \"kubernetes.io/projected/447fc9d5-7789-43d6-8d11-d5ef5b00c371-kube-api-access-tvm6k\") pod \"neutron-5564fdb9d4-xt6f7\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.520776 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.573173 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.825014 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3be8e6d-863b-49ba-b177-5ee2bb76ba51","Type":"ContainerStarted","Data":"0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952"} Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.829008 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d51c1b58-d44f-44cd-a513-af4fb272c0a7","Type":"ContainerStarted","Data":"78297be1a506fe0c9e7ba8b6348cafd64ffab90d1e75cb6ad2177171c185713b"} Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.829072 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d51c1b58-d44f-44cd-a513-af4fb272c0a7","Type":"ContainerStarted","Data":"daf627fdb7650538b5227b20b9c3661cc4f32bff6d3c41d2a44219cc248b6742"} Feb 02 12:29:34 crc kubenswrapper[4846]: I0202 12:29:34.854504 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.854483625 podStartE2EDuration="3.854483625s" podCreationTimestamp="2026-02-02 12:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:34.851707235 +0000 UTC m=+1206.080294098" watchObservedRunningTime="2026-02-02 12:29:34.854483625 +0000 UTC m=+1206.083070488" Feb 02 12:29:35 crc kubenswrapper[4846]: I0202 12:29:35.455566 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="240e8c55-ad59-4bee-a92e-93876302d7f5" path="/var/lib/kubelet/pods/240e8c55-ad59-4bee-a92e-93876302d7f5/volumes" Feb 02 12:29:35 crc kubenswrapper[4846]: I0202 12:29:35.598842 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-9tl7z"] Feb 02 12:29:35 crc kubenswrapper[4846]: I0202 12:29:35.612125 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5564fdb9d4-xt6f7"] Feb 02 12:29:35 crc kubenswrapper[4846]: I0202 12:29:35.843718 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaece0f9-90e4-4f49-b5e4-6566e68f115a","Type":"ContainerStarted","Data":"65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db"} Feb 02 12:29:35 crc kubenswrapper[4846]: I0202 12:29:35.844976 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5564fdb9d4-xt6f7" event={"ID":"447fc9d5-7789-43d6-8d11-d5ef5b00c371","Type":"ContainerStarted","Data":"134b847080e6eb8962ce3f55225f92a0323d0bac5e6436cc287b32e457441961"} Feb 02 12:29:35 crc kubenswrapper[4846]: I0202 12:29:35.845997 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" event={"ID":"832895eb-b40b-427c-9485-166c86ef9d83","Type":"ContainerStarted","Data":"ed6716686f243ecf222e6cdb3c58848fa90b90e0f9602a12136b05341bedb39b"} Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.652173 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-85c76b7df7-9b4xg"] Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.653827 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.657003 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.658395 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.663658 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85c76b7df7-9b4xg"] Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.727905 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-config\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.727960 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-internal-tls-certs\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.727989 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-ovndb-tls-certs\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.728016 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-public-tls-certs\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.728051 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-combined-ca-bundle\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.728075 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86rcl\" (UniqueName: \"kubernetes.io/projected/0374610f-a1fd-4650-8ad8-cc4455c06611-kube-api-access-86rcl\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.728103 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-httpd-config\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.829214 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-internal-tls-certs\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.829275 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-ovndb-tls-certs\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.829313 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-public-tls-certs\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.829366 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-combined-ca-bundle\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.829399 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86rcl\" (UniqueName: \"kubernetes.io/projected/0374610f-a1fd-4650-8ad8-cc4455c06611-kube-api-access-86rcl\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.829436 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-httpd-config\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.829527 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-config\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.837580 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-combined-ca-bundle\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.838278 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-public-tls-certs\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.839679 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-ovndb-tls-certs\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.840494 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-config\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.844156 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-httpd-config\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.860319 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-internal-tls-certs\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.865370 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86rcl\" (UniqueName: \"kubernetes.io/projected/0374610f-a1fd-4650-8ad8-cc4455c06611-kube-api-access-86rcl\") pod \"neutron-85c76b7df7-9b4xg\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.866868 4846 generic.go:334] "Generic (PLEG): container finished" podID="e02dd0b8-6bd8-40b2-8a18-64d298c72a67" containerID="2acbbb7c816156b529232e467d5fe13563d39a52756e14b37ed7aa312ee11088" exitCode=0 Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.866934 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-858jm" event={"ID":"e02dd0b8-6bd8-40b2-8a18-64d298c72a67","Type":"ContainerDied","Data":"2acbbb7c816156b529232e467d5fe13563d39a52756e14b37ed7aa312ee11088"} Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.870885 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3be8e6d-863b-49ba-b177-5ee2bb76ba51","Type":"ContainerStarted","Data":"a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90"} Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.881703 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5564fdb9d4-xt6f7" event={"ID":"447fc9d5-7789-43d6-8d11-d5ef5b00c371","Type":"ContainerStarted","Data":"6af2e184bc53c534dc08e92be427ce1c5bdc36efdacf88e01825552ab6c352fb"} Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.881752 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5564fdb9d4-xt6f7" event={"ID":"447fc9d5-7789-43d6-8d11-d5ef5b00c371","Type":"ContainerStarted","Data":"a7c5dc480cc9c268bfef70458e359db736dad2aa52a7110af88fdcd9bdce0c4a"} Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.881971 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.884179 4846 generic.go:334] "Generic (PLEG): container finished" podID="832895eb-b40b-427c-9485-166c86ef9d83" containerID="b06a7d3f575d6dadfe158b28c17ac82876b750ab3fb08efd49494832290fec5c" exitCode=0 Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.884222 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" event={"ID":"832895eb-b40b-427c-9485-166c86ef9d83","Type":"ContainerDied","Data":"b06a7d3f575d6dadfe158b28c17ac82876b750ab3fb08efd49494832290fec5c"} Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.945600 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.945580207 podStartE2EDuration="5.945580207s" podCreationTimestamp="2026-02-02 12:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:36.926964521 +0000 UTC m=+1208.155551394" watchObservedRunningTime="2026-02-02 12:29:36.945580207 +0000 UTC m=+1208.174167070" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.969205 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5564fdb9d4-xt6f7" podStartSLOduration=2.969188098 podStartE2EDuration="2.969188098s" podCreationTimestamp="2026-02-02 12:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:36.961935317 +0000 UTC m=+1208.190522180" watchObservedRunningTime="2026-02-02 12:29:36.969188098 +0000 UTC m=+1208.197774961" Feb 02 12:29:36 crc kubenswrapper[4846]: I0202 12:29:36.977759 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:37 crc kubenswrapper[4846]: I0202 12:29:37.631679 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85c76b7df7-9b4xg"] Feb 02 12:29:37 crc kubenswrapper[4846]: W0202 12:29:37.641993 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0374610f_a1fd_4650_8ad8_cc4455c06611.slice/crio-e8699d00e2a7950dfaf6b63f50aa81c359403d573b6117eacfab24ad2a6ff406 WatchSource:0}: Error finding container e8699d00e2a7950dfaf6b63f50aa81c359403d573b6117eacfab24ad2a6ff406: Status 404 returned error can't find the container with id e8699d00e2a7950dfaf6b63f50aa81c359403d573b6117eacfab24ad2a6ff406 Feb 02 12:29:37 crc kubenswrapper[4846]: I0202 12:29:37.912868 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85c76b7df7-9b4xg" event={"ID":"0374610f-a1fd-4650-8ad8-cc4455c06611","Type":"ContainerStarted","Data":"e8699d00e2a7950dfaf6b63f50aa81c359403d573b6117eacfab24ad2a6ff406"} Feb 02 12:29:37 crc kubenswrapper[4846]: I0202 12:29:37.922887 4846 generic.go:334] "Generic (PLEG): container finished" podID="5812b4fc-8c66-418f-8886-0ff26018b905" containerID="61857b150929bc24955da85821ee7c4e71280ebcd4d2731ae05aa0b38a82d290" exitCode=0 Feb 02 12:29:37 crc kubenswrapper[4846]: I0202 12:29:37.923024 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m2dtn" event={"ID":"5812b4fc-8c66-418f-8886-0ff26018b905","Type":"ContainerDied","Data":"61857b150929bc24955da85821ee7c4e71280ebcd4d2731ae05aa0b38a82d290"} Feb 02 12:29:37 crc kubenswrapper[4846]: I0202 12:29:37.932145 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" event={"ID":"832895eb-b40b-427c-9485-166c86ef9d83","Type":"ContainerStarted","Data":"df2bd6d33b3d7847a66799b3d337ef4de2e21f5a107b247cd084706377c989f2"} Feb 02 12:29:37 crc kubenswrapper[4846]: I0202 12:29:37.979221 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" podStartSLOduration=3.97920236 podStartE2EDuration="3.97920236s" podCreationTimestamp="2026-02-02 12:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:37.964589493 +0000 UTC m=+1209.193176366" watchObservedRunningTime="2026-02-02 12:29:37.97920236 +0000 UTC m=+1209.207789223" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.330597 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-858jm" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.365732 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-logs\") pod \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.365812 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqlxr\" (UniqueName: \"kubernetes.io/projected/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-kube-api-access-dqlxr\") pod \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.365839 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-config-data\") pod \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.365929 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-scripts\") pod \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.365966 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-combined-ca-bundle\") pod \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\" (UID: \"e02dd0b8-6bd8-40b2-8a18-64d298c72a67\") " Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.366149 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-logs" (OuterVolumeSpecName: "logs") pod "e02dd0b8-6bd8-40b2-8a18-64d298c72a67" (UID: "e02dd0b8-6bd8-40b2-8a18-64d298c72a67"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.366382 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.372866 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-scripts" (OuterVolumeSpecName: "scripts") pod "e02dd0b8-6bd8-40b2-8a18-64d298c72a67" (UID: "e02dd0b8-6bd8-40b2-8a18-64d298c72a67"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.373153 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-kube-api-access-dqlxr" (OuterVolumeSpecName: "kube-api-access-dqlxr") pod "e02dd0b8-6bd8-40b2-8a18-64d298c72a67" (UID: "e02dd0b8-6bd8-40b2-8a18-64d298c72a67"). InnerVolumeSpecName "kube-api-access-dqlxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.390248 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e02dd0b8-6bd8-40b2-8a18-64d298c72a67" (UID: "e02dd0b8-6bd8-40b2-8a18-64d298c72a67"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.391135 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-config-data" (OuterVolumeSpecName: "config-data") pod "e02dd0b8-6bd8-40b2-8a18-64d298c72a67" (UID: "e02dd0b8-6bd8-40b2-8a18-64d298c72a67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.469473 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqlxr\" (UniqueName: \"kubernetes.io/projected/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-kube-api-access-dqlxr\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.469499 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.469508 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.469519 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02dd0b8-6bd8-40b2-8a18-64d298c72a67-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.942428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-858jm" event={"ID":"e02dd0b8-6bd8-40b2-8a18-64d298c72a67","Type":"ContainerDied","Data":"4fa5450ef3e8e9522271cc126cc1fcf00b43a3a35f39da799acac7233e310bfc"} Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.942791 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fa5450ef3e8e9522271cc126cc1fcf00b43a3a35f39da799acac7233e310bfc" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.942886 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-858jm" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.942896 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.983499 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7b686584c4-zpsks"] Feb 02 12:29:38 crc kubenswrapper[4846]: E0202 12:29:38.984036 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e02dd0b8-6bd8-40b2-8a18-64d298c72a67" containerName="placement-db-sync" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.984057 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e02dd0b8-6bd8-40b2-8a18-64d298c72a67" containerName="placement-db-sync" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.984224 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e02dd0b8-6bd8-40b2-8a18-64d298c72a67" containerName="placement-db-sync" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.985399 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.992808 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.993055 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.993190 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.993299 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-tx4l4" Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.993326 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7b686584c4-zpsks"] Feb 02 12:29:38 crc kubenswrapper[4846]: I0202 12:29:38.993402 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.186142 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-public-tls-certs\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.186493 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc8x9\" (UniqueName: \"kubernetes.io/projected/a8dce541-0c00-476e-8efe-fea46f5fd26b-kube-api-access-tc8x9\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.186522 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-config-data\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.186554 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-internal-tls-certs\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.186573 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-combined-ca-bundle\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.186647 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-scripts\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.186679 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8dce541-0c00-476e-8efe-fea46f5fd26b-logs\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.289233 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-public-tls-certs\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.289330 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc8x9\" (UniqueName: \"kubernetes.io/projected/a8dce541-0c00-476e-8efe-fea46f5fd26b-kube-api-access-tc8x9\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.289356 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-config-data\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.289388 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-internal-tls-certs\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.289405 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-combined-ca-bundle\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.289443 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-scripts\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.289464 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8dce541-0c00-476e-8efe-fea46f5fd26b-logs\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.289857 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8dce541-0c00-476e-8efe-fea46f5fd26b-logs\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.296071 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-combined-ca-bundle\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.298279 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-internal-tls-certs\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.298912 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-public-tls-certs\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.300203 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-scripts\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.301094 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-config-data\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.307729 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc8x9\" (UniqueName: \"kubernetes.io/projected/a8dce541-0c00-476e-8efe-fea46f5fd26b-kube-api-access-tc8x9\") pod \"placement-7b686584c4-zpsks\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.336641 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:39 crc kubenswrapper[4846]: I0202 12:29:39.954142 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85c76b7df7-9b4xg" event={"ID":"0374610f-a1fd-4650-8ad8-cc4455c06611","Type":"ContainerStarted","Data":"d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6"} Feb 02 12:29:41 crc kubenswrapper[4846]: I0202 12:29:41.986814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m2dtn" event={"ID":"5812b4fc-8c66-418f-8886-0ff26018b905","Type":"ContainerDied","Data":"702aa68a164f5867839ab6a9b6bd9b4cc98c3b3f803fc69a69da399385cab4d3"} Feb 02 12:29:41 crc kubenswrapper[4846]: I0202 12:29:41.987233 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="702aa68a164f5867839ab6a9b6bd9b4cc98c3b3f803fc69a69da399385cab4d3" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.173918 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.176808 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.176860 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.201140 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.201195 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.222318 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.237374 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.237526 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.258707 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.343901 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-credential-keys\") pod \"5812b4fc-8c66-418f-8886-0ff26018b905\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.343961 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hpvm\" (UniqueName: \"kubernetes.io/projected/5812b4fc-8c66-418f-8886-0ff26018b905-kube-api-access-8hpvm\") pod \"5812b4fc-8c66-418f-8886-0ff26018b905\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.344079 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-config-data\") pod \"5812b4fc-8c66-418f-8886-0ff26018b905\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.344184 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-scripts\") pod \"5812b4fc-8c66-418f-8886-0ff26018b905\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.344205 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-fernet-keys\") pod \"5812b4fc-8c66-418f-8886-0ff26018b905\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.344244 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-combined-ca-bundle\") pod \"5812b4fc-8c66-418f-8886-0ff26018b905\" (UID: \"5812b4fc-8c66-418f-8886-0ff26018b905\") " Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.349262 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5812b4fc-8c66-418f-8886-0ff26018b905" (UID: "5812b4fc-8c66-418f-8886-0ff26018b905"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.350385 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-scripts" (OuterVolumeSpecName: "scripts") pod "5812b4fc-8c66-418f-8886-0ff26018b905" (UID: "5812b4fc-8c66-418f-8886-0ff26018b905"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.350753 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5812b4fc-8c66-418f-8886-0ff26018b905-kube-api-access-8hpvm" (OuterVolumeSpecName: "kube-api-access-8hpvm") pod "5812b4fc-8c66-418f-8886-0ff26018b905" (UID: "5812b4fc-8c66-418f-8886-0ff26018b905"). InnerVolumeSpecName "kube-api-access-8hpvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.351774 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5812b4fc-8c66-418f-8886-0ff26018b905" (UID: "5812b4fc-8c66-418f-8886-0ff26018b905"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.368725 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5812b4fc-8c66-418f-8886-0ff26018b905" (UID: "5812b4fc-8c66-418f-8886-0ff26018b905"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.381770 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-config-data" (OuterVolumeSpecName: "config-data") pod "5812b4fc-8c66-418f-8886-0ff26018b905" (UID: "5812b4fc-8c66-418f-8886-0ff26018b905"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.416426 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7b686584c4-zpsks"] Feb 02 12:29:42 crc kubenswrapper[4846]: W0202 12:29:42.417070 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8dce541_0c00_476e_8efe_fea46f5fd26b.slice/crio-1febc541411b1d75707a3e5c2074281449c1ae411202f330f81cbeb589253295 WatchSource:0}: Error finding container 1febc541411b1d75707a3e5c2074281449c1ae411202f330f81cbeb589253295: Status 404 returned error can't find the container with id 1febc541411b1d75707a3e5c2074281449c1ae411202f330f81cbeb589253295 Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.446098 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.446137 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.446148 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.446156 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.446164 4846 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5812b4fc-8c66-418f-8886-0ff26018b905-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.446172 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hpvm\" (UniqueName: \"kubernetes.io/projected/5812b4fc-8c66-418f-8886-0ff26018b905-kube-api-access-8hpvm\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.996505 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85c76b7df7-9b4xg" event={"ID":"0374610f-a1fd-4650-8ad8-cc4455c06611","Type":"ContainerStarted","Data":"39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca"} Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.996800 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.999705 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b686584c4-zpsks" event={"ID":"a8dce541-0c00-476e-8efe-fea46f5fd26b","Type":"ContainerStarted","Data":"493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb"} Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.999749 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b686584c4-zpsks" event={"ID":"a8dce541-0c00-476e-8efe-fea46f5fd26b","Type":"ContainerStarted","Data":"07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376"} Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.999763 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b686584c4-zpsks" event={"ID":"a8dce541-0c00-476e-8efe-fea46f5fd26b","Type":"ContainerStarted","Data":"1febc541411b1d75707a3e5c2074281449c1ae411202f330f81cbeb589253295"} Feb 02 12:29:42 crc kubenswrapper[4846]: I0202 12:29:42.999952 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.000008 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.002007 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaece0f9-90e4-4f49-b5e4-6566e68f115a","Type":"ContainerStarted","Data":"04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340"} Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.002165 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m2dtn" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.002867 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.003120 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.003141 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.003152 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.029584 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-85c76b7df7-9b4xg" podStartSLOduration=7.02956599 podStartE2EDuration="7.02956599s" podCreationTimestamp="2026-02-02 12:29:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:43.016123823 +0000 UTC m=+1214.244710706" watchObservedRunningTime="2026-02-02 12:29:43.02956599 +0000 UTC m=+1214.258152853" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.049134 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7b686584c4-zpsks" podStartSLOduration=5.0491107 podStartE2EDuration="5.0491107s" podCreationTimestamp="2026-02-02 12:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:43.043032177 +0000 UTC m=+1214.271619060" watchObservedRunningTime="2026-02-02 12:29:43.0491107 +0000 UTC m=+1214.277697563" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.291079 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6fb558fdfb-5z9p5"] Feb 02 12:29:43 crc kubenswrapper[4846]: E0202 12:29:43.291565 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5812b4fc-8c66-418f-8886-0ff26018b905" containerName="keystone-bootstrap" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.291590 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5812b4fc-8c66-418f-8886-0ff26018b905" containerName="keystone-bootstrap" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.292747 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5812b4fc-8c66-418f-8886-0ff26018b905" containerName="keystone-bootstrap" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.293554 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.297524 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.298333 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9wp9t" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.298364 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.299432 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.300064 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.300129 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.314413 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6fb558fdfb-5z9p5"] Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.466020 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-combined-ca-bundle\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.466082 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-public-tls-certs\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.466127 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-scripts\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.466146 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-fernet-keys\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.466241 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-credential-keys\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.466323 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-internal-tls-certs\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.466428 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-config-data\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.466477 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cglmv\" (UniqueName: \"kubernetes.io/projected/c858d5df-042d-4211-898b-1ad4ef1fea47-kube-api-access-cglmv\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.568369 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-combined-ca-bundle\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.568442 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-public-tls-certs\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.568545 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-scripts\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.568594 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-fernet-keys\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.568633 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-credential-keys\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.568681 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-internal-tls-certs\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.568809 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-config-data\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.568861 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cglmv\" (UniqueName: \"kubernetes.io/projected/c858d5df-042d-4211-898b-1ad4ef1fea47-kube-api-access-cglmv\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.578037 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-credential-keys\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.580353 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-config-data\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.581140 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-public-tls-certs\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.582452 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-internal-tls-certs\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.583043 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-combined-ca-bundle\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.586070 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-scripts\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.590893 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-fernet-keys\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.592402 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cglmv\" (UniqueName: \"kubernetes.io/projected/c858d5df-042d-4211-898b-1ad4ef1fea47-kube-api-access-cglmv\") pod \"keystone-6fb558fdfb-5z9p5\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:43 crc kubenswrapper[4846]: I0202 12:29:43.610395 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:44 crc kubenswrapper[4846]: I0202 12:29:44.118029 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6fb558fdfb-5z9p5"] Feb 02 12:29:44 crc kubenswrapper[4846]: W0202 12:29:44.127521 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc858d5df_042d_4211_898b_1ad4ef1fea47.slice/crio-86f259fd640dbbb237ee8cd278959e04a228664bd7b954c37f50421a3305dd96 WatchSource:0}: Error finding container 86f259fd640dbbb237ee8cd278959e04a228664bd7b954c37f50421a3305dd96: Status 404 returned error can't find the container with id 86f259fd640dbbb237ee8cd278959e04a228664bd7b954c37f50421a3305dd96 Feb 02 12:29:44 crc kubenswrapper[4846]: I0202 12:29:44.522774 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:29:44 crc kubenswrapper[4846]: I0202 12:29:44.586900 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-5zrpv"] Feb 02 12:29:44 crc kubenswrapper[4846]: I0202 12:29:44.587158 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" podUID="ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" containerName="dnsmasq-dns" containerID="cri-o://676abdb2a9bef4899561705a6f61a832b0d5b2934cf37d0d5b2f5f8287ea8d40" gracePeriod=10 Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.060444 4846 generic.go:334] "Generic (PLEG): container finished" podID="ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" containerID="676abdb2a9bef4899561705a6f61a832b0d5b2934cf37d0d5b2f5f8287ea8d40" exitCode=0 Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.060909 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" event={"ID":"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3","Type":"ContainerDied","Data":"676abdb2a9bef4899561705a6f61a832b0d5b2934cf37d0d5b2f5f8287ea8d40"} Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.067983 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.068013 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.069785 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6fb558fdfb-5z9p5" event={"ID":"c858d5df-042d-4211-898b-1ad4ef1fea47","Type":"ContainerStarted","Data":"c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced"} Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.069818 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6fb558fdfb-5z9p5" event={"ID":"c858d5df-042d-4211-898b-1ad4ef1fea47","Type":"ContainerStarted","Data":"86f259fd640dbbb237ee8cd278959e04a228664bd7b954c37f50421a3305dd96"} Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.069851 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.100859 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6fb558fdfb-5z9p5" podStartSLOduration=2.100838425 podStartE2EDuration="2.100838425s" podCreationTimestamp="2026-02-02 12:29:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:45.095022019 +0000 UTC m=+1216.323608892" watchObservedRunningTime="2026-02-02 12:29:45.100838425 +0000 UTC m=+1216.329425298" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.238944 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.299131 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-config\") pod \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.299235 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2d9z\" (UniqueName: \"kubernetes.io/projected/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-kube-api-access-g2d9z\") pod \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.299294 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-swift-storage-0\") pod \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.299319 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-svc\") pod \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.299391 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-sb\") pod \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.299525 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-nb\") pod \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\" (UID: \"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3\") " Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.310949 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-kube-api-access-g2d9z" (OuterVolumeSpecName: "kube-api-access-g2d9z") pod "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" (UID: "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3"). InnerVolumeSpecName "kube-api-access-g2d9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.344336 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.363695 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" (UID: "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.367571 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" (UID: "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.368008 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" (UID: "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.372171 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-config" (OuterVolumeSpecName: "config") pod "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" (UID: "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.401584 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.401642 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.401656 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.401673 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.401685 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2d9z\" (UniqueName: \"kubernetes.io/projected/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-kube-api-access-g2d9z\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.405672 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" (UID: "ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.504728 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.637366 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.637472 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.721001 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 02 12:29:45 crc kubenswrapper[4846]: I0202 12:29:45.847806 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 02 12:29:46 crc kubenswrapper[4846]: I0202 12:29:46.081874 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" event={"ID":"ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3","Type":"ContainerDied","Data":"e835e235aaaddaa6f65f3e9decabd4976d3dd94084152f836fb7dfd61ee0a37b"} Feb 02 12:29:46 crc kubenswrapper[4846]: I0202 12:29:46.082750 4846 scope.go:117] "RemoveContainer" containerID="676abdb2a9bef4899561705a6f61a832b0d5b2934cf37d0d5b2f5f8287ea8d40" Feb 02 12:29:46 crc kubenswrapper[4846]: I0202 12:29:46.082196 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f58898b5f-5zrpv" Feb 02 12:29:46 crc kubenswrapper[4846]: I0202 12:29:46.119011 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-5zrpv"] Feb 02 12:29:46 crc kubenswrapper[4846]: I0202 12:29:46.121416 4846 scope.go:117] "RemoveContainer" containerID="6beb6342e776aa866deb61277b624fe882c39aaa430b0686aa5d6ccf94528526" Feb 02 12:29:46 crc kubenswrapper[4846]: I0202 12:29:46.133052 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f58898b5f-5zrpv"] Feb 02 12:29:47 crc kubenswrapper[4846]: I0202 12:29:47.092911 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2z995" event={"ID":"8b58cb4f-6d15-4d97-871d-724c160ec765","Type":"ContainerStarted","Data":"6d07b794031104d16f41e77ce48912977df34ffc3a88d5151c8b49bbc089573c"} Feb 02 12:29:47 crc kubenswrapper[4846]: I0202 12:29:47.118610 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-2z995" podStartSLOduration=2.935623732 podStartE2EDuration="36.118590209s" podCreationTimestamp="2026-02-02 12:29:11 +0000 UTC" firstStartedPulling="2026-02-02 12:29:12.801285487 +0000 UTC m=+1184.029872350" lastFinishedPulling="2026-02-02 12:29:45.984251964 +0000 UTC m=+1217.212838827" observedRunningTime="2026-02-02 12:29:47.107667485 +0000 UTC m=+1218.336254358" watchObservedRunningTime="2026-02-02 12:29:47.118590209 +0000 UTC m=+1218.347177072" Feb 02 12:29:47 crc kubenswrapper[4846]: I0202 12:29:47.454237 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" path="/var/lib/kubelet/pods/ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3/volumes" Feb 02 12:29:48 crc kubenswrapper[4846]: I0202 12:29:48.105044 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5vmgj" event={"ID":"18307d58-f249-4ec0-8556-fd98e4febb66","Type":"ContainerStarted","Data":"a8b3459915c6a0efed9d2ec47db4ee7ad0d37fbe1e2d2df99bb10ca0e374e45f"} Feb 02 12:29:48 crc kubenswrapper[4846]: I0202 12:29:48.133595 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-5vmgj" podStartSLOduration=3.477159538 podStartE2EDuration="37.133574324s" podCreationTimestamp="2026-02-02 12:29:11 +0000 UTC" firstStartedPulling="2026-02-02 12:29:13.24132777 +0000 UTC m=+1184.469914633" lastFinishedPulling="2026-02-02 12:29:46.897742556 +0000 UTC m=+1218.126329419" observedRunningTime="2026-02-02 12:29:48.121912992 +0000 UTC m=+1219.350499875" watchObservedRunningTime="2026-02-02 12:29:48.133574324 +0000 UTC m=+1219.362161187" Feb 02 12:29:51 crc kubenswrapper[4846]: I0202 12:29:51.133418 4846 generic.go:334] "Generic (PLEG): container finished" podID="18307d58-f249-4ec0-8556-fd98e4febb66" containerID="a8b3459915c6a0efed9d2ec47db4ee7ad0d37fbe1e2d2df99bb10ca0e374e45f" exitCode=0 Feb 02 12:29:51 crc kubenswrapper[4846]: I0202 12:29:51.133472 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5vmgj" event={"ID":"18307d58-f249-4ec0-8556-fd98e4febb66","Type":"ContainerDied","Data":"a8b3459915c6a0efed9d2ec47db4ee7ad0d37fbe1e2d2df99bb10ca0e374e45f"} Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.143025 4846 generic.go:334] "Generic (PLEG): container finished" podID="8b58cb4f-6d15-4d97-871d-724c160ec765" containerID="6d07b794031104d16f41e77ce48912977df34ffc3a88d5151c8b49bbc089573c" exitCode=0 Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.143142 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2z995" event={"ID":"8b58cb4f-6d15-4d97-871d-724c160ec765","Type":"ContainerDied","Data":"6d07b794031104d16f41e77ce48912977df34ffc3a88d5151c8b49bbc089573c"} Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.481550 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.632549 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-combined-ca-bundle\") pod \"18307d58-f249-4ec0-8556-fd98e4febb66\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.632675 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5qsx\" (UniqueName: \"kubernetes.io/projected/18307d58-f249-4ec0-8556-fd98e4febb66-kube-api-access-s5qsx\") pod \"18307d58-f249-4ec0-8556-fd98e4febb66\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.632833 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-db-sync-config-data\") pod \"18307d58-f249-4ec0-8556-fd98e4febb66\" (UID: \"18307d58-f249-4ec0-8556-fd98e4febb66\") " Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.636741 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "18307d58-f249-4ec0-8556-fd98e4febb66" (UID: "18307d58-f249-4ec0-8556-fd98e4febb66"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.636787 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18307d58-f249-4ec0-8556-fd98e4febb66-kube-api-access-s5qsx" (OuterVolumeSpecName: "kube-api-access-s5qsx") pod "18307d58-f249-4ec0-8556-fd98e4febb66" (UID: "18307d58-f249-4ec0-8556-fd98e4febb66"). InnerVolumeSpecName "kube-api-access-s5qsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.654757 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18307d58-f249-4ec0-8556-fd98e4febb66" (UID: "18307d58-f249-4ec0-8556-fd98e4febb66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.734731 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.734792 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18307d58-f249-4ec0-8556-fd98e4febb66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:52 crc kubenswrapper[4846]: I0202 12:29:52.734812 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5qsx\" (UniqueName: \"kubernetes.io/projected/18307d58-f249-4ec0-8556-fd98e4febb66-kube-api-access-s5qsx\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.157053 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5vmgj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.160262 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5vmgj" event={"ID":"18307d58-f249-4ec0-8556-fd98e4febb66","Type":"ContainerDied","Data":"f339e7c1fd39e322f50cd38287d51f6177c816f19874355c73eed6af1433c525"} Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.160586 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f339e7c1fd39e322f50cd38287d51f6177c816f19874355c73eed6af1433c525" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.168819 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaece0f9-90e4-4f49-b5e4-6566e68f115a","Type":"ContainerStarted","Data":"fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2"} Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.168858 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="ceilometer-central-agent" containerID="cri-o://06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4" gracePeriod=30 Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.168944 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="sg-core" containerID="cri-o://04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340" gracePeriod=30 Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.168982 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="proxy-httpd" containerID="cri-o://fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2" gracePeriod=30 Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.169040 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="ceilometer-notification-agent" containerID="cri-o://65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db" gracePeriod=30 Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.169183 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.352515 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.482764348 podStartE2EDuration="42.352492236s" podCreationTimestamp="2026-02-02 12:29:11 +0000 UTC" firstStartedPulling="2026-02-02 12:29:13.06726457 +0000 UTC m=+1184.295851433" lastFinishedPulling="2026-02-02 12:29:51.936992458 +0000 UTC m=+1223.165579321" observedRunningTime="2026-02-02 12:29:53.204776057 +0000 UTC m=+1224.433362920" watchObservedRunningTime="2026-02-02 12:29:53.352492236 +0000 UTC m=+1224.581079099" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.358877 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5d9447b957-4tn8l"] Feb 02 12:29:53 crc kubenswrapper[4846]: E0202 12:29:53.359322 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" containerName="dnsmasq-dns" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.359339 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" containerName="dnsmasq-dns" Feb 02 12:29:53 crc kubenswrapper[4846]: E0202 12:29:53.359355 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" containerName="init" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.359362 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" containerName="init" Feb 02 12:29:53 crc kubenswrapper[4846]: E0202 12:29:53.359385 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18307d58-f249-4ec0-8556-fd98e4febb66" containerName="barbican-db-sync" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.359392 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="18307d58-f249-4ec0-8556-fd98e4febb66" containerName="barbican-db-sync" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.359650 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="18307d58-f249-4ec0-8556-fd98e4febb66" containerName="barbican-db-sync" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.359682 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffdcb1cd-5944-4d12-8da0-9ccc2b5f9bc3" containerName="dnsmasq-dns" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.360939 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.377315 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.377433 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.377643 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-vmnbh" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.378677 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5d9447b957-4tn8l"] Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.411831 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-df8f98698-ddglj"] Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.413667 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.423992 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.449128 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da322cd4-f770-42d7-af2f-e4ed438da7dd-logs\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.449193 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m2bw\" (UniqueName: \"kubernetes.io/projected/da322cd4-f770-42d7-af2f-e4ed438da7dd-kube-api-access-9m2bw\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.449228 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.449271 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-combined-ca-bundle\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.449324 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data-custom\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.449754 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-df8f98698-ddglj"] Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.483611 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54c675984c-k6jsw"] Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.485034 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.499965 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c675984c-k6jsw"] Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.560779 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data-custom\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.560831 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data-custom\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.560960 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-logs\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.560993 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w9ln\" (UniqueName: \"kubernetes.io/projected/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-kube-api-access-4w9ln\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.561018 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da322cd4-f770-42d7-af2f-e4ed438da7dd-logs\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.561042 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-combined-ca-bundle\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.561078 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m2bw\" (UniqueName: \"kubernetes.io/projected/da322cd4-f770-42d7-af2f-e4ed438da7dd-kube-api-access-9m2bw\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.561127 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.561158 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.561223 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-combined-ca-bundle\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.562159 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da322cd4-f770-42d7-af2f-e4ed438da7dd-logs\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.571082 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-combined-ca-bundle\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.583830 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data-custom\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.585731 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m2bw\" (UniqueName: \"kubernetes.io/projected/da322cd4-f770-42d7-af2f-e4ed438da7dd-kube-api-access-9m2bw\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.590538 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data\") pod \"barbican-worker-5d9447b957-4tn8l\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.632051 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6f59fdf6f6-rsrcc"] Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.633446 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.635968 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.644818 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f59fdf6f6-rsrcc"] Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663356 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-sb\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663415 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-nb\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663431 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-svc\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663467 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data-custom\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663520 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-logs\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663545 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w9ln\" (UniqueName: \"kubernetes.io/projected/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-kube-api-access-4w9ln\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663563 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-combined-ca-bundle\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663602 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpd95\" (UniqueName: \"kubernetes.io/projected/079a218d-def0-4f54-9012-9185b473772d-kube-api-access-rpd95\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663634 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-config\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663653 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.663669 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-swift-storage-0\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.665055 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-logs\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.669905 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.670965 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data-custom\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.671644 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-combined-ca-bundle\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.699186 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w9ln\" (UniqueName: \"kubernetes.io/projected/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-kube-api-access-4w9ln\") pod \"barbican-keystone-listener-df8f98698-ddglj\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.717569 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.758801 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765459 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-svc\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765506 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-nb\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765556 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765603 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx95l\" (UniqueName: \"kubernetes.io/projected/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-kube-api-access-tx95l\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765647 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data-custom\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765715 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-logs\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765759 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpd95\" (UniqueName: \"kubernetes.io/projected/079a218d-def0-4f54-9012-9185b473772d-kube-api-access-rpd95\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765784 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-config\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765814 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-swift-storage-0\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765838 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-combined-ca-bundle\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.765897 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-sb\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.766777 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.766983 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-sb\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.767688 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-svc\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.767839 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-config\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.770865 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-swift-storage-0\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.771118 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-nb\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.794036 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpd95\" (UniqueName: \"kubernetes.io/projected/079a218d-def0-4f54-9012-9185b473772d-kube-api-access-rpd95\") pod \"dnsmasq-dns-54c675984c-k6jsw\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.818635 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.867397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-scripts\") pod \"8b58cb4f-6d15-4d97-871d-724c160ec765\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.867432 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-combined-ca-bundle\") pod \"8b58cb4f-6d15-4d97-871d-724c160ec765\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.867505 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b58cb4f-6d15-4d97-871d-724c160ec765-etc-machine-id\") pod \"8b58cb4f-6d15-4d97-871d-724c160ec765\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.867530 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-db-sync-config-data\") pod \"8b58cb4f-6d15-4d97-871d-724c160ec765\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.867663 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdklw\" (UniqueName: \"kubernetes.io/projected/8b58cb4f-6d15-4d97-871d-724c160ec765-kube-api-access-zdklw\") pod \"8b58cb4f-6d15-4d97-871d-724c160ec765\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.867722 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-config-data\") pod \"8b58cb4f-6d15-4d97-871d-724c160ec765\" (UID: \"8b58cb4f-6d15-4d97-871d-724c160ec765\") " Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.867859 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b58cb4f-6d15-4d97-871d-724c160ec765-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8b58cb4f-6d15-4d97-871d-724c160ec765" (UID: "8b58cb4f-6d15-4d97-871d-724c160ec765"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.867929 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-logs\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.867978 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-combined-ca-bundle\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.868070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.868100 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx95l\" (UniqueName: \"kubernetes.io/projected/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-kube-api-access-tx95l\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.868119 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data-custom\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.868187 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8b58cb4f-6d15-4d97-871d-724c160ec765-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.868614 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-logs\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.872359 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-scripts" (OuterVolumeSpecName: "scripts") pod "8b58cb4f-6d15-4d97-871d-724c160ec765" (UID: "8b58cb4f-6d15-4d97-871d-724c160ec765"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.873762 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-combined-ca-bundle\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.875996 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8b58cb4f-6d15-4d97-871d-724c160ec765" (UID: "8b58cb4f-6d15-4d97-871d-724c160ec765"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.877227 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b58cb4f-6d15-4d97-871d-724c160ec765-kube-api-access-zdklw" (OuterVolumeSpecName: "kube-api-access-zdklw") pod "8b58cb4f-6d15-4d97-871d-724c160ec765" (UID: "8b58cb4f-6d15-4d97-871d-724c160ec765"). InnerVolumeSpecName "kube-api-access-zdklw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.879056 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.882932 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data-custom\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.890642 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx95l\" (UniqueName: \"kubernetes.io/projected/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-kube-api-access-tx95l\") pod \"barbican-api-6f59fdf6f6-rsrcc\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.935134 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-config-data" (OuterVolumeSpecName: "config-data") pod "8b58cb4f-6d15-4d97-871d-724c160ec765" (UID: "8b58cb4f-6d15-4d97-871d-724c160ec765"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.937564 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b58cb4f-6d15-4d97-871d-724c160ec765" (UID: "8b58cb4f-6d15-4d97-871d-724c160ec765"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.969663 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.969703 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.969717 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.969728 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdklw\" (UniqueName: \"kubernetes.io/projected/8b58cb4f-6d15-4d97-871d-724c160ec765-kube-api-access-zdklw\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.969743 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b58cb4f-6d15-4d97-871d-724c160ec765-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:53 crc kubenswrapper[4846]: I0202 12:29:53.997326 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.183640 4846 generic.go:334] "Generic (PLEG): container finished" podID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerID="fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2" exitCode=0 Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.183674 4846 generic.go:334] "Generic (PLEG): container finished" podID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerID="04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340" exitCode=2 Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.183682 4846 generic.go:334] "Generic (PLEG): container finished" podID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerID="06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4" exitCode=0 Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.183736 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaece0f9-90e4-4f49-b5e4-6566e68f115a","Type":"ContainerDied","Data":"fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2"} Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.183789 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaece0f9-90e4-4f49-b5e4-6566e68f115a","Type":"ContainerDied","Data":"04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340"} Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.183807 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaece0f9-90e4-4f49-b5e4-6566e68f115a","Type":"ContainerDied","Data":"06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4"} Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.186535 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2z995" event={"ID":"8b58cb4f-6d15-4d97-871d-724c160ec765","Type":"ContainerDied","Data":"058e3abed1c7fa20ff7ae22cf4b12aca6926aaa714b57c36b4a25d932b2e26ab"} Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.186582 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="058e3abed1c7fa20ff7ae22cf4b12aca6926aaa714b57c36b4a25d932b2e26ab" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.186671 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2z995" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.253889 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5d9447b957-4tn8l"] Feb 02 12:29:54 crc kubenswrapper[4846]: W0202 12:29:54.262001 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda322cd4_f770_42d7_af2f_e4ed438da7dd.slice/crio-70d06d2b49e642d48d21ff3ba2fbf33bfa031c755653a0cdd03546e05a3866d0 WatchSource:0}: Error finding container 70d06d2b49e642d48d21ff3ba2fbf33bfa031c755653a0cdd03546e05a3866d0: Status 404 returned error can't find the container with id 70d06d2b49e642d48d21ff3ba2fbf33bfa031c755653a0cdd03546e05a3866d0 Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.380276 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-df8f98698-ddglj"] Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.413257 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c675984c-k6jsw"] Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.486649 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:29:54 crc kubenswrapper[4846]: E0202 12:29:54.487055 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b58cb4f-6d15-4d97-871d-724c160ec765" containerName="cinder-db-sync" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.487068 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b58cb4f-6d15-4d97-871d-724c160ec765" containerName="cinder-db-sync" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.487226 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b58cb4f-6d15-4d97-871d-724c160ec765" containerName="cinder-db-sync" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.490302 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.496950 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-plfcv" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.497229 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.497545 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.497979 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.507878 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.587761 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv8qz\" (UniqueName: \"kubernetes.io/projected/87c361f2-a89b-475e-b485-55438adf0c71-kube-api-access-lv8qz\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.587822 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.587870 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.587914 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87c361f2-a89b-475e-b485-55438adf0c71-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.587971 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-scripts\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.588033 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.596790 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c675984c-k6jsw"] Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.611408 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f59fdf6f6-rsrcc"] Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.620458 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-849fd69845-gdnm8"] Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.623098 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.691323 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87c361f2-a89b-475e-b485-55438adf0c71-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.691589 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-config\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.691682 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-scripts\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.691750 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-swift-storage-0\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.691843 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-svc\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.691915 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.692013 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5jss\" (UniqueName: \"kubernetes.io/projected/910f6926-083b-489e-b89c-7f8a46a1966e-kube-api-access-h5jss\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.692089 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv8qz\" (UniqueName: \"kubernetes.io/projected/87c361f2-a89b-475e-b485-55438adf0c71-kube-api-access-lv8qz\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.692156 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-sb\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.692232 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.692318 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-nb\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.692394 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.697095 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87c361f2-a89b-475e-b485-55438adf0c71-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.704476 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.704976 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.706883 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-scripts\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.707249 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-849fd69845-gdnm8"] Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.728857 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.729192 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv8qz\" (UniqueName: \"kubernetes.io/projected/87c361f2-a89b-475e-b485-55438adf0c71-kube-api-access-lv8qz\") pod \"cinder-scheduler-0\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.777356 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.783273 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.784839 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.788389 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.795513 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-nb\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.795646 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-config\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.795670 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-swift-storage-0\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.795713 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-svc\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.795756 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5jss\" (UniqueName: \"kubernetes.io/projected/910f6926-083b-489e-b89c-7f8a46a1966e-kube-api-access-h5jss\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.795773 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-sb\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.797031 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-sb\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.798666 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-svc\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.798850 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-swift-storage-0\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.799422 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-nb\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.799902 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-config\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.827899 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.830862 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5jss\" (UniqueName: \"kubernetes.io/projected/910f6926-083b-489e-b89c-7f8a46a1966e-kube-api-access-h5jss\") pod \"dnsmasq-dns-849fd69845-gdnm8\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.897657 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-scripts\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.897755 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.897782 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7bf5\" (UniqueName: \"kubernetes.io/projected/3991f275-c433-48e1-a384-2c454ccb05f2-kube-api-access-q7bf5\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.897865 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.897982 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3991f275-c433-48e1-a384-2c454ccb05f2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.898072 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data-custom\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.898126 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3991f275-c433-48e1-a384-2c454ccb05f2-logs\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:54 crc kubenswrapper[4846]: I0202 12:29:54.957077 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:54.999501 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3991f275-c433-48e1-a384-2c454ccb05f2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:54.999596 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data-custom\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:54.999629 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3991f275-c433-48e1-a384-2c454ccb05f2-logs\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:54.999662 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-scripts\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:54.999679 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3991f275-c433-48e1-a384-2c454ccb05f2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:54.999691 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:54.999764 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7bf5\" (UniqueName: \"kubernetes.io/projected/3991f275-c433-48e1-a384-2c454ccb05f2-kube-api-access-q7bf5\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:54.999848 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.000540 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3991f275-c433-48e1-a384-2c454ccb05f2-logs\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.005329 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data-custom\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.006490 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.024503 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.024591 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-scripts\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.036114 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7bf5\" (UniqueName: \"kubernetes.io/projected/3991f275-c433-48e1-a384-2c454ccb05f2-kube-api-access-q7bf5\") pod \"cinder-api-0\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.104989 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.201537 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" event={"ID":"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3","Type":"ContainerStarted","Data":"d9e9e45c9a9509e1b26760b56d70a5a4d9c8a2fe419074443d10faf93fcdd5d2"} Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.207561 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9447b957-4tn8l" event={"ID":"da322cd4-f770-42d7-af2f-e4ed438da7dd","Type":"ContainerStarted","Data":"70d06d2b49e642d48d21ff3ba2fbf33bfa031c755653a0cdd03546e05a3866d0"} Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.215253 4846 generic.go:334] "Generic (PLEG): container finished" podID="079a218d-def0-4f54-9012-9185b473772d" containerID="f7c2e75253f0150d5dd924abe0c199fdebde96df0102ac1969b5fa2267591a85" exitCode=0 Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.215349 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c675984c-k6jsw" event={"ID":"079a218d-def0-4f54-9012-9185b473772d","Type":"ContainerDied","Data":"f7c2e75253f0150d5dd924abe0c199fdebde96df0102ac1969b5fa2267591a85"} Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.215379 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c675984c-k6jsw" event={"ID":"079a218d-def0-4f54-9012-9185b473772d","Type":"ContainerStarted","Data":"ab916b2f8f93a0a87a14d372ad256ac6d007a3b1524d846a0aa8e401bdeb9127"} Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.229160 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" event={"ID":"a4500640-81f0-49dd-a0f4-e9b5a4e211b0","Type":"ContainerStarted","Data":"616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5"} Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.229217 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" event={"ID":"a4500640-81f0-49dd-a0f4-e9b5a4e211b0","Type":"ContainerStarted","Data":"e1397205faac1ccd81752df80c4a7f4c972080add1cc8d8c3af334e9b4d74a0a"} Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.317700 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:29:55 crc kubenswrapper[4846]: W0202 12:29:55.323273 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87c361f2_a89b_475e_b485_55438adf0c71.slice/crio-0ac8881f488d257a91d1cf9481cb05fc16d234cc06fe709015505bebaf77d57d WatchSource:0}: Error finding container 0ac8881f488d257a91d1cf9481cb05fc16d234cc06fe709015505bebaf77d57d: Status 404 returned error can't find the container with id 0ac8881f488d257a91d1cf9481cb05fc16d234cc06fe709015505bebaf77d57d Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.472027 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-849fd69845-gdnm8"] Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.642000 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:29:55 crc kubenswrapper[4846]: I0202 12:29:55.889115 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.022278 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-nb\") pod \"079a218d-def0-4f54-9012-9185b473772d\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.022308 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-svc\") pod \"079a218d-def0-4f54-9012-9185b473772d\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.022363 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpd95\" (UniqueName: \"kubernetes.io/projected/079a218d-def0-4f54-9012-9185b473772d-kube-api-access-rpd95\") pod \"079a218d-def0-4f54-9012-9185b473772d\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.022404 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-swift-storage-0\") pod \"079a218d-def0-4f54-9012-9185b473772d\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.022429 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-config\") pod \"079a218d-def0-4f54-9012-9185b473772d\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.022466 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-sb\") pod \"079a218d-def0-4f54-9012-9185b473772d\" (UID: \"079a218d-def0-4f54-9012-9185b473772d\") " Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.044239 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/079a218d-def0-4f54-9012-9185b473772d-kube-api-access-rpd95" (OuterVolumeSpecName: "kube-api-access-rpd95") pod "079a218d-def0-4f54-9012-9185b473772d" (UID: "079a218d-def0-4f54-9012-9185b473772d"). InnerVolumeSpecName "kube-api-access-rpd95". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.052698 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "079a218d-def0-4f54-9012-9185b473772d" (UID: "079a218d-def0-4f54-9012-9185b473772d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.055517 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-config" (OuterVolumeSpecName: "config") pod "079a218d-def0-4f54-9012-9185b473772d" (UID: "079a218d-def0-4f54-9012-9185b473772d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.065721 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "079a218d-def0-4f54-9012-9185b473772d" (UID: "079a218d-def0-4f54-9012-9185b473772d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.068734 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "079a218d-def0-4f54-9012-9185b473772d" (UID: "079a218d-def0-4f54-9012-9185b473772d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.085801 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "079a218d-def0-4f54-9012-9185b473772d" (UID: "079a218d-def0-4f54-9012-9185b473772d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.125309 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.125353 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.125367 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpd95\" (UniqueName: \"kubernetes.io/projected/079a218d-def0-4f54-9012-9185b473772d-kube-api-access-rpd95\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.125381 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.125392 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.125403 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/079a218d-def0-4f54-9012-9185b473772d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.239508 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3991f275-c433-48e1-a384-2c454ccb05f2","Type":"ContainerStarted","Data":"c1e26bd7ee64db45433192fc76798b171773f8c99b375fb18ea6eac0e3612771"} Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.241293 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c675984c-k6jsw" event={"ID":"079a218d-def0-4f54-9012-9185b473772d","Type":"ContainerDied","Data":"ab916b2f8f93a0a87a14d372ad256ac6d007a3b1524d846a0aa8e401bdeb9127"} Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.241589 4846 scope.go:117] "RemoveContainer" containerID="f7c2e75253f0150d5dd924abe0c199fdebde96df0102ac1969b5fa2267591a85" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.241330 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c675984c-k6jsw" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.255334 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" event={"ID":"a4500640-81f0-49dd-a0f4-e9b5a4e211b0","Type":"ContainerStarted","Data":"662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae"} Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.255496 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.255525 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.259190 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" event={"ID":"910f6926-083b-489e-b89c-7f8a46a1966e","Type":"ContainerStarted","Data":"3ac2f45ab6d6663923341dbe2d10412efb0c4a33702b91b1797ab6a418c5934e"} Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.260907 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87c361f2-a89b-475e-b485-55438adf0c71","Type":"ContainerStarted","Data":"0ac8881f488d257a91d1cf9481cb05fc16d234cc06fe709015505bebaf77d57d"} Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.304662 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" podStartSLOduration=3.304189437 podStartE2EDuration="3.304189437s" podCreationTimestamp="2026-02-02 12:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:56.272873851 +0000 UTC m=+1227.501489715" watchObservedRunningTime="2026-02-02 12:29:56.304189437 +0000 UTC m=+1227.532776320" Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.330167 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c675984c-k6jsw"] Feb 02 12:29:56 crc kubenswrapper[4846]: I0202 12:29:56.344343 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54c675984c-k6jsw"] Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.272532 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3991f275-c433-48e1-a384-2c454ccb05f2","Type":"ContainerStarted","Data":"38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d"} Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.279076 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" event={"ID":"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3","Type":"ContainerStarted","Data":"319bed97e3756574fc101f5461b63927d20d9217f036ffcd758a126ac4c88d35"} Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.279131 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" event={"ID":"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3","Type":"ContainerStarted","Data":"199fdc9812e4067e4c230bdb3772de61e8c1276ecc5f01f740fc1af2fa9d74c4"} Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.283302 4846 generic.go:334] "Generic (PLEG): container finished" podID="910f6926-083b-489e-b89c-7f8a46a1966e" containerID="bc9070cb63fe7e9b5243f71356b106983b303c752a2ace197a2a53a26bcccd20" exitCode=0 Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.283552 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" event={"ID":"910f6926-083b-489e-b89c-7f8a46a1966e","Type":"ContainerDied","Data":"bc9070cb63fe7e9b5243f71356b106983b303c752a2ace197a2a53a26bcccd20"} Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.287570 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9447b957-4tn8l" event={"ID":"da322cd4-f770-42d7-af2f-e4ed438da7dd","Type":"ContainerStarted","Data":"77c5ca7cfef457fcaadda875671e3c5d3ab8c7606c3409257026722b9bd9fa0f"} Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.287657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9447b957-4tn8l" event={"ID":"da322cd4-f770-42d7-af2f-e4ed438da7dd","Type":"ContainerStarted","Data":"c6820cbfbd66e9d71a05716d17fadd1549cc9619a0b848aee68552415ffa96a1"} Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.295847 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87c361f2-a89b-475e-b485-55438adf0c71","Type":"ContainerStarted","Data":"c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e"} Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.311560 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" podStartSLOduration=2.347522781 podStartE2EDuration="4.31153895s" podCreationTimestamp="2026-02-02 12:29:53 +0000 UTC" firstStartedPulling="2026-02-02 12:29:54.404484799 +0000 UTC m=+1225.633071662" lastFinishedPulling="2026-02-02 12:29:56.368500968 +0000 UTC m=+1227.597087831" observedRunningTime="2026-02-02 12:29:57.295661843 +0000 UTC m=+1228.524248726" watchObservedRunningTime="2026-02-02 12:29:57.31153895 +0000 UTC m=+1228.540125833" Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.339390 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5d9447b957-4tn8l" podStartSLOduration=2.244161722 podStartE2EDuration="4.339365197s" podCreationTimestamp="2026-02-02 12:29:53 +0000 UTC" firstStartedPulling="2026-02-02 12:29:54.26403951 +0000 UTC m=+1225.492626373" lastFinishedPulling="2026-02-02 12:29:56.359242995 +0000 UTC m=+1227.587829848" observedRunningTime="2026-02-02 12:29:57.316311939 +0000 UTC m=+1228.544898802" watchObservedRunningTime="2026-02-02 12:29:57.339365197 +0000 UTC m=+1228.567952070" Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.463423 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="079a218d-def0-4f54-9012-9185b473772d" path="/var/lib/kubelet/pods/079a218d-def0-4f54-9012-9185b473772d/volumes" Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.464016 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:29:57 crc kubenswrapper[4846]: I0202 12:29:57.944597 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.064528 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-combined-ca-bundle\") pod \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.064947 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-log-httpd\") pod \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.065079 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-sg-core-conf-yaml\") pod \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.065104 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-scripts\") pod \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.065130 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-config-data\") pod \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.065274 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phlbl\" (UniqueName: \"kubernetes.io/projected/eaece0f9-90e4-4f49-b5e4-6566e68f115a-kube-api-access-phlbl\") pod \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.065330 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-run-httpd\") pod \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\" (UID: \"eaece0f9-90e4-4f49-b5e4-6566e68f115a\") " Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.065754 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eaece0f9-90e4-4f49-b5e4-6566e68f115a" (UID: "eaece0f9-90e4-4f49-b5e4-6566e68f115a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.066086 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eaece0f9-90e4-4f49-b5e4-6566e68f115a" (UID: "eaece0f9-90e4-4f49-b5e4-6566e68f115a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.070539 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-scripts" (OuterVolumeSpecName: "scripts") pod "eaece0f9-90e4-4f49-b5e4-6566e68f115a" (UID: "eaece0f9-90e4-4f49-b5e4-6566e68f115a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.072465 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaece0f9-90e4-4f49-b5e4-6566e68f115a-kube-api-access-phlbl" (OuterVolumeSpecName: "kube-api-access-phlbl") pod "eaece0f9-90e4-4f49-b5e4-6566e68f115a" (UID: "eaece0f9-90e4-4f49-b5e4-6566e68f115a"). InnerVolumeSpecName "kube-api-access-phlbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.099849 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eaece0f9-90e4-4f49-b5e4-6566e68f115a" (UID: "eaece0f9-90e4-4f49-b5e4-6566e68f115a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.167852 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaece0f9-90e4-4f49-b5e4-6566e68f115a" (UID: "eaece0f9-90e4-4f49-b5e4-6566e68f115a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.168436 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phlbl\" (UniqueName: \"kubernetes.io/projected/eaece0f9-90e4-4f49-b5e4-6566e68f115a-kube-api-access-phlbl\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.168465 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.168476 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.168485 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaece0f9-90e4-4f49-b5e4-6566e68f115a-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.168493 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.168502 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.205105 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-config-data" (OuterVolumeSpecName: "config-data") pod "eaece0f9-90e4-4f49-b5e4-6566e68f115a" (UID: "eaece0f9-90e4-4f49-b5e4-6566e68f115a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.270093 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaece0f9-90e4-4f49-b5e4-6566e68f115a-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.305993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3991f275-c433-48e1-a384-2c454ccb05f2","Type":"ContainerStarted","Data":"366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476"} Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.306174 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3991f275-c433-48e1-a384-2c454ccb05f2" containerName="cinder-api-log" containerID="cri-o://38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d" gracePeriod=30 Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.306496 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.306845 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3991f275-c433-48e1-a384-2c454ccb05f2" containerName="cinder-api" containerID="cri-o://366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476" gracePeriod=30 Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.311169 4846 generic.go:334] "Generic (PLEG): container finished" podID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerID="65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db" exitCode=0 Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.311248 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaece0f9-90e4-4f49-b5e4-6566e68f115a","Type":"ContainerDied","Data":"65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db"} Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.311281 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaece0f9-90e4-4f49-b5e4-6566e68f115a","Type":"ContainerDied","Data":"87a9e5bce5fb483da5d5792d05755698945894c86f07a50466b033f549e6cf1e"} Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.311303 4846 scope.go:117] "RemoveContainer" containerID="fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.311347 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.314991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" event={"ID":"910f6926-083b-489e-b89c-7f8a46a1966e","Type":"ContainerStarted","Data":"8100cf88b62535f5fde3d24535b341bccc58bbc5c406f5b2996fb84c5d013d72"} Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.316127 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.323598 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87c361f2-a89b-475e-b485-55438adf0c71","Type":"ContainerStarted","Data":"d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e"} Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.339264 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.339241264 podStartE2EDuration="4.339241264s" podCreationTimestamp="2026-02-02 12:29:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:58.322759951 +0000 UTC m=+1229.551346834" watchObservedRunningTime="2026-02-02 12:29:58.339241264 +0000 UTC m=+1229.567828127" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.342614 4846 scope.go:117] "RemoveContainer" containerID="04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.349914 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" podStartSLOduration=4.349894471 podStartE2EDuration="4.349894471s" podCreationTimestamp="2026-02-02 12:29:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:29:58.348057894 +0000 UTC m=+1229.576644767" watchObservedRunningTime="2026-02-02 12:29:58.349894471 +0000 UTC m=+1229.578481334" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.385843 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.393817 4846 scope.go:117] "RemoveContainer" containerID="65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.399096 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.411283 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:29:58 crc kubenswrapper[4846]: E0202 12:29:58.411790 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="proxy-httpd" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.411817 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="proxy-httpd" Feb 02 12:29:58 crc kubenswrapper[4846]: E0202 12:29:58.411841 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="ceilometer-central-agent" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.411849 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="ceilometer-central-agent" Feb 02 12:29:58 crc kubenswrapper[4846]: E0202 12:29:58.411861 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="ceilometer-notification-agent" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.411868 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="ceilometer-notification-agent" Feb 02 12:29:58 crc kubenswrapper[4846]: E0202 12:29:58.411876 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="sg-core" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.411882 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="sg-core" Feb 02 12:29:58 crc kubenswrapper[4846]: E0202 12:29:58.411913 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="079a218d-def0-4f54-9012-9185b473772d" containerName="init" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.411921 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="079a218d-def0-4f54-9012-9185b473772d" containerName="init" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.412130 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="ceilometer-notification-agent" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.412158 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="ceilometer-central-agent" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.412175 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="079a218d-def0-4f54-9012-9185b473772d" containerName="init" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.412192 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="sg-core" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.412203 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" containerName="proxy-httpd" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.413726 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.368831685 podStartE2EDuration="4.413712709s" podCreationTimestamp="2026-02-02 12:29:54 +0000 UTC" firstStartedPulling="2026-02-02 12:29:55.327779317 +0000 UTC m=+1226.556366180" lastFinishedPulling="2026-02-02 12:29:56.372660341 +0000 UTC m=+1227.601247204" observedRunningTime="2026-02-02 12:29:58.397170384 +0000 UTC m=+1229.625757247" watchObservedRunningTime="2026-02-02 12:29:58.413712709 +0000 UTC m=+1229.642299582" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.414020 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.433065 4846 scope.go:117] "RemoveContainer" containerID="06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.433747 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.437086 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.475668 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-config-data\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.475811 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb6ts\" (UniqueName: \"kubernetes.io/projected/2f74434a-8ddb-40fd-8ecb-637c3a63c919-kube-api-access-gb6ts\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.475837 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.475871 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.475893 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-run-httpd\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.475937 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-log-httpd\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.476135 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-scripts\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.486189 4846 scope.go:117] "RemoveContainer" containerID="fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2" Feb 02 12:29:58 crc kubenswrapper[4846]: E0202 12:29:58.489564 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2\": container with ID starting with fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2 not found: ID does not exist" containerID="fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.489717 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2"} err="failed to get container status \"fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2\": rpc error: code = NotFound desc = could not find container \"fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2\": container with ID starting with fe739958c5ca0890f3d2bfd13ff45c6de8f4b0b58e795b0ad04bf1c73dc70fc2 not found: ID does not exist" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.489761 4846 scope.go:117] "RemoveContainer" containerID="04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340" Feb 02 12:29:58 crc kubenswrapper[4846]: E0202 12:29:58.490093 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340\": container with ID starting with 04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340 not found: ID does not exist" containerID="04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.490124 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340"} err="failed to get container status \"04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340\": rpc error: code = NotFound desc = could not find container \"04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340\": container with ID starting with 04580f3aaeb269168c2f3e7c36172077373b6baa126adf62d79d74fc6e7ae340 not found: ID does not exist" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.490143 4846 scope.go:117] "RemoveContainer" containerID="65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db" Feb 02 12:29:58 crc kubenswrapper[4846]: E0202 12:29:58.490398 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db\": container with ID starting with 65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db not found: ID does not exist" containerID="65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.490424 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db"} err="failed to get container status \"65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db\": rpc error: code = NotFound desc = could not find container \"65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db\": container with ID starting with 65ac3dd9bdf01cc19119e74f3c08b67ce141e2b4c0c2bd86d8682da3ea65e9db not found: ID does not exist" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.490444 4846 scope.go:117] "RemoveContainer" containerID="06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4" Feb 02 12:29:58 crc kubenswrapper[4846]: E0202 12:29:58.490799 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4\": container with ID starting with 06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4 not found: ID does not exist" containerID="06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.490827 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4"} err="failed to get container status \"06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4\": rpc error: code = NotFound desc = could not find container \"06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4\": container with ID starting with 06ff8806c621c4fcf39296af5a9e7fffecdaea1e1306c27d098efa4c7ce0d1e4 not found: ID does not exist" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.500513 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.577990 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-config-data\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.578123 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb6ts\" (UniqueName: \"kubernetes.io/projected/2f74434a-8ddb-40fd-8ecb-637c3a63c919-kube-api-access-gb6ts\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.578163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.578205 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.578235 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-run-httpd\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.578294 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-log-httpd\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.578337 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-scripts\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.579276 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-log-httpd\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.580025 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-run-httpd\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.582783 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-scripts\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.583397 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.589516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.597239 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb6ts\" (UniqueName: \"kubernetes.io/projected/2f74434a-8ddb-40fd-8ecb-637c3a63c919-kube-api-access-gb6ts\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.597729 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-config-data\") pod \"ceilometer-0\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " pod="openstack/ceilometer-0" Feb 02 12:29:58 crc kubenswrapper[4846]: I0202 12:29:58.762643 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.219909 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.334849 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.341668 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f74434a-8ddb-40fd-8ecb-637c3a63c919","Type":"ContainerStarted","Data":"531541ff093aac860ab0078d5ef28a6104f5f755676ac72d600ac38cd351dea1"} Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.345515 4846 generic.go:334] "Generic (PLEG): container finished" podID="3991f275-c433-48e1-a384-2c454ccb05f2" containerID="366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476" exitCode=0 Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.345570 4846 generic.go:334] "Generic (PLEG): container finished" podID="3991f275-c433-48e1-a384-2c454ccb05f2" containerID="38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d" exitCode=143 Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.346504 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.346693 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3991f275-c433-48e1-a384-2c454ccb05f2","Type":"ContainerDied","Data":"366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476"} Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.346719 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3991f275-c433-48e1-a384-2c454ccb05f2","Type":"ContainerDied","Data":"38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d"} Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.346731 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3991f275-c433-48e1-a384-2c454ccb05f2","Type":"ContainerDied","Data":"c1e26bd7ee64db45433192fc76798b171773f8c99b375fb18ea6eac0e3612771"} Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.346756 4846 scope.go:117] "RemoveContainer" containerID="366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.392134 4846 scope.go:117] "RemoveContainer" containerID="38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.393232 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data\") pod \"3991f275-c433-48e1-a384-2c454ccb05f2\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.393282 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3991f275-c433-48e1-a384-2c454ccb05f2-etc-machine-id\") pod \"3991f275-c433-48e1-a384-2c454ccb05f2\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.393342 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7bf5\" (UniqueName: \"kubernetes.io/projected/3991f275-c433-48e1-a384-2c454ccb05f2-kube-api-access-q7bf5\") pod \"3991f275-c433-48e1-a384-2c454ccb05f2\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.393378 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-combined-ca-bundle\") pod \"3991f275-c433-48e1-a384-2c454ccb05f2\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.393439 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3991f275-c433-48e1-a384-2c454ccb05f2-logs\") pod \"3991f275-c433-48e1-a384-2c454ccb05f2\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.393542 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data-custom\") pod \"3991f275-c433-48e1-a384-2c454ccb05f2\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.393605 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-scripts\") pod \"3991f275-c433-48e1-a384-2c454ccb05f2\" (UID: \"3991f275-c433-48e1-a384-2c454ccb05f2\") " Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.394791 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3991f275-c433-48e1-a384-2c454ccb05f2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3991f275-c433-48e1-a384-2c454ccb05f2" (UID: "3991f275-c433-48e1-a384-2c454ccb05f2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.395196 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3991f275-c433-48e1-a384-2c454ccb05f2-logs" (OuterVolumeSpecName: "logs") pod "3991f275-c433-48e1-a384-2c454ccb05f2" (UID: "3991f275-c433-48e1-a384-2c454ccb05f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.400848 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3991f275-c433-48e1-a384-2c454ccb05f2" (UID: "3991f275-c433-48e1-a384-2c454ccb05f2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.416349 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3991f275-c433-48e1-a384-2c454ccb05f2-kube-api-access-q7bf5" (OuterVolumeSpecName: "kube-api-access-q7bf5") pod "3991f275-c433-48e1-a384-2c454ccb05f2" (UID: "3991f275-c433-48e1-a384-2c454ccb05f2"). InnerVolumeSpecName "kube-api-access-q7bf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.417527 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-scripts" (OuterVolumeSpecName: "scripts") pod "3991f275-c433-48e1-a384-2c454ccb05f2" (UID: "3991f275-c433-48e1-a384-2c454ccb05f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.423745 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3991f275-c433-48e1-a384-2c454ccb05f2" (UID: "3991f275-c433-48e1-a384-2c454ccb05f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.438258 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaece0f9-90e4-4f49-b5e4-6566e68f115a" path="/var/lib/kubelet/pods/eaece0f9-90e4-4f49-b5e4-6566e68f115a/volumes" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.465179 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data" (OuterVolumeSpecName: "config-data") pod "3991f275-c433-48e1-a384-2c454ccb05f2" (UID: "3991f275-c433-48e1-a384-2c454ccb05f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.495449 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3991f275-c433-48e1-a384-2c454ccb05f2-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.495485 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.495498 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.495508 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.495519 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3991f275-c433-48e1-a384-2c454ccb05f2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.495529 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7bf5\" (UniqueName: \"kubernetes.io/projected/3991f275-c433-48e1-a384-2c454ccb05f2-kube-api-access-q7bf5\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.495541 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3991f275-c433-48e1-a384-2c454ccb05f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.520449 4846 scope.go:117] "RemoveContainer" containerID="366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476" Feb 02 12:29:59 crc kubenswrapper[4846]: E0202 12:29:59.521387 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476\": container with ID starting with 366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476 not found: ID does not exist" containerID="366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.521521 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476"} err="failed to get container status \"366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476\": rpc error: code = NotFound desc = could not find container \"366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476\": container with ID starting with 366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476 not found: ID does not exist" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.521650 4846 scope.go:117] "RemoveContainer" containerID="38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d" Feb 02 12:29:59 crc kubenswrapper[4846]: E0202 12:29:59.522038 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d\": container with ID starting with 38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d not found: ID does not exist" containerID="38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.522146 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d"} err="failed to get container status \"38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d\": rpc error: code = NotFound desc = could not find container \"38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d\": container with ID starting with 38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d not found: ID does not exist" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.522237 4846 scope.go:117] "RemoveContainer" containerID="366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.522543 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476"} err="failed to get container status \"366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476\": rpc error: code = NotFound desc = could not find container \"366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476\": container with ID starting with 366779a5cf20a34265677d1299347e791a1bac43fddc49a35b49892739f6f476 not found: ID does not exist" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.522575 4846 scope.go:117] "RemoveContainer" containerID="38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.522781 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d"} err="failed to get container status \"38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d\": rpc error: code = NotFound desc = could not find container \"38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d\": container with ID starting with 38f0316c940353f795a209f64e5a8cefbfabb25a4a065ef0c578546af910e34d not found: ID does not exist" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.679485 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.691069 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.705811 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:29:59 crc kubenswrapper[4846]: E0202 12:29:59.706191 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3991f275-c433-48e1-a384-2c454ccb05f2" containerName="cinder-api-log" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.706233 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3991f275-c433-48e1-a384-2c454ccb05f2" containerName="cinder-api-log" Feb 02 12:29:59 crc kubenswrapper[4846]: E0202 12:29:59.706254 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3991f275-c433-48e1-a384-2c454ccb05f2" containerName="cinder-api" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.706262 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3991f275-c433-48e1-a384-2c454ccb05f2" containerName="cinder-api" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.706504 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3991f275-c433-48e1-a384-2c454ccb05f2" containerName="cinder-api-log" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.706530 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3991f275-c433-48e1-a384-2c454ccb05f2" containerName="cinder-api" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.707656 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.710712 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.711434 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.711755 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.735097 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.803017 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.803075 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/532236a1-a04d-4b99-9131-b1d16a3d5ed4-logs\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.803098 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.803259 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpgqz\" (UniqueName: \"kubernetes.io/projected/532236a1-a04d-4b99-9131-b1d16a3d5ed4-kube-api-access-bpgqz\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.803395 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/532236a1-a04d-4b99-9131-b1d16a3d5ed4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.803491 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.803634 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.803697 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data-custom\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.803760 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-scripts\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.829233 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.904927 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-scripts\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.904983 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.905028 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/532236a1-a04d-4b99-9131-b1d16a3d5ed4-logs\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.905051 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.905099 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpgqz\" (UniqueName: \"kubernetes.io/projected/532236a1-a04d-4b99-9131-b1d16a3d5ed4-kube-api-access-bpgqz\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.905148 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/532236a1-a04d-4b99-9131-b1d16a3d5ed4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.905190 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.905243 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.905266 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data-custom\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.905278 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/532236a1-a04d-4b99-9131-b1d16a3d5ed4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.906021 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/532236a1-a04d-4b99-9131-b1d16a3d5ed4-logs\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.908657 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-scripts\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.910492 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.911121 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.912748 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.919092 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.920030 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data-custom\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:29:59 crc kubenswrapper[4846]: I0202 12:29:59.923533 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpgqz\" (UniqueName: \"kubernetes.io/projected/532236a1-a04d-4b99-9131-b1d16a3d5ed4-kube-api-access-bpgqz\") pod \"cinder-api-0\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " pod="openstack/cinder-api-0" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.026233 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.151504 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs"] Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.153256 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.157759 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.161580 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.170110 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs"] Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.210922 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1c1ec69-7585-45e7-812b-df08a8ede792-config-volume\") pod \"collect-profiles-29500590-vpxfs\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.210984 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pq5n\" (UniqueName: \"kubernetes.io/projected/e1c1ec69-7585-45e7-812b-df08a8ede792-kube-api-access-9pq5n\") pod \"collect-profiles-29500590-vpxfs\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.213997 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1c1ec69-7585-45e7-812b-df08a8ede792-secret-volume\") pod \"collect-profiles-29500590-vpxfs\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.242960 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-f6447845b-zlnf5"] Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.244533 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.248858 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.249028 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.283950 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f6447845b-zlnf5"] Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.318613 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pq5n\" (UniqueName: \"kubernetes.io/projected/e1c1ec69-7585-45e7-812b-df08a8ede792-kube-api-access-9pq5n\") pod \"collect-profiles-29500590-vpxfs\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.318797 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data-custom\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.318853 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-combined-ca-bundle\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.318908 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1c1ec69-7585-45e7-812b-df08a8ede792-secret-volume\") pod \"collect-profiles-29500590-vpxfs\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.318928 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmmpv\" (UniqueName: \"kubernetes.io/projected/69522beb-f7d5-4694-ada8-0d17d7d3f16b-kube-api-access-hmmpv\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.318962 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-internal-tls-certs\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.318994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-public-tls-certs\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.319022 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.319086 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1c1ec69-7585-45e7-812b-df08a8ede792-config-volume\") pod \"collect-profiles-29500590-vpxfs\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.319109 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69522beb-f7d5-4694-ada8-0d17d7d3f16b-logs\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.321729 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1c1ec69-7585-45e7-812b-df08a8ede792-config-volume\") pod \"collect-profiles-29500590-vpxfs\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.334895 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pq5n\" (UniqueName: \"kubernetes.io/projected/e1c1ec69-7585-45e7-812b-df08a8ede792-kube-api-access-9pq5n\") pod \"collect-profiles-29500590-vpxfs\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.337566 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1c1ec69-7585-45e7-812b-df08a8ede792-secret-volume\") pod \"collect-profiles-29500590-vpxfs\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.372376 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f74434a-8ddb-40fd-8ecb-637c3a63c919","Type":"ContainerStarted","Data":"951f36c17c01a2fa6e3011d452524eeca1d82b0bd91bd8c6634e1795396413b0"} Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.421342 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data-custom\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.421425 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-combined-ca-bundle\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.421503 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmmpv\" (UniqueName: \"kubernetes.io/projected/69522beb-f7d5-4694-ada8-0d17d7d3f16b-kube-api-access-hmmpv\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.421543 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-internal-tls-certs\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.421639 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-public-tls-certs\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.421683 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.421765 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69522beb-f7d5-4694-ada8-0d17d7d3f16b-logs\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.423454 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69522beb-f7d5-4694-ada8-0d17d7d3f16b-logs\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.428298 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-internal-tls-certs\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.428423 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-public-tls-certs\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.429236 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.429797 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-combined-ca-bundle\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.431196 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data-custom\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.444551 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmmpv\" (UniqueName: \"kubernetes.io/projected/69522beb-f7d5-4694-ada8-0d17d7d3f16b-kube-api-access-hmmpv\") pod \"barbican-api-f6447845b-zlnf5\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.528885 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.566004 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:00 crc kubenswrapper[4846]: I0202 12:30:00.624727 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:30:00 crc kubenswrapper[4846]: W0202 12:30:00.653904 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod532236a1_a04d_4b99_9131_b1d16a3d5ed4.slice/crio-6cd6eb31b01b4a0397af1e0fddba60ae869a8fe15c7977a5da77283e4cba9946 WatchSource:0}: Error finding container 6cd6eb31b01b4a0397af1e0fddba60ae869a8fe15c7977a5da77283e4cba9946: Status 404 returned error can't find the container with id 6cd6eb31b01b4a0397af1e0fddba60ae869a8fe15c7977a5da77283e4cba9946 Feb 02 12:30:01 crc kubenswrapper[4846]: I0202 12:30:01.033667 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs"] Feb 02 12:30:01 crc kubenswrapper[4846]: W0202 12:30:01.058709 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1c1ec69_7585_45e7_812b_df08a8ede792.slice/crio-926cdcef48bce76185177fec2b7b1727cb332b6082f4c1acf76b296164977a3e WatchSource:0}: Error finding container 926cdcef48bce76185177fec2b7b1727cb332b6082f4c1acf76b296164977a3e: Status 404 returned error can't find the container with id 926cdcef48bce76185177fec2b7b1727cb332b6082f4c1acf76b296164977a3e Feb 02 12:30:01 crc kubenswrapper[4846]: I0202 12:30:01.125407 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f6447845b-zlnf5"] Feb 02 12:30:01 crc kubenswrapper[4846]: I0202 12:30:01.385846 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f74434a-8ddb-40fd-8ecb-637c3a63c919","Type":"ContainerStarted","Data":"21ccdd76b8af8a5c289f14aa6a68389288abd0458ad44f6858c41a788230a56b"} Feb 02 12:30:01 crc kubenswrapper[4846]: I0202 12:30:01.388857 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6447845b-zlnf5" event={"ID":"69522beb-f7d5-4694-ada8-0d17d7d3f16b","Type":"ContainerStarted","Data":"4c0d152aa8838d6ae8e216419456abc7969ff129785bcc1f4be20cac7b2e770f"} Feb 02 12:30:01 crc kubenswrapper[4846]: I0202 12:30:01.390897 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"532236a1-a04d-4b99-9131-b1d16a3d5ed4","Type":"ContainerStarted","Data":"6cd6eb31b01b4a0397af1e0fddba60ae869a8fe15c7977a5da77283e4cba9946"} Feb 02 12:30:01 crc kubenswrapper[4846]: I0202 12:30:01.392581 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" event={"ID":"e1c1ec69-7585-45e7-812b-df08a8ede792","Type":"ContainerStarted","Data":"926cdcef48bce76185177fec2b7b1727cb332b6082f4c1acf76b296164977a3e"} Feb 02 12:30:01 crc kubenswrapper[4846]: I0202 12:30:01.422452 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" podStartSLOduration=1.4224322169999999 podStartE2EDuration="1.422432217s" podCreationTimestamp="2026-02-02 12:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:30:01.416919749 +0000 UTC m=+1232.645506642" watchObservedRunningTime="2026-02-02 12:30:01.422432217 +0000 UTC m=+1232.651019080" Feb 02 12:30:01 crc kubenswrapper[4846]: I0202 12:30:01.443989 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3991f275-c433-48e1-a384-2c454ccb05f2" path="/var/lib/kubelet/pods/3991f275-c433-48e1-a384-2c454ccb05f2/volumes" Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.425924 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"532236a1-a04d-4b99-9131-b1d16a3d5ed4","Type":"ContainerStarted","Data":"afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2"} Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.427485 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.427601 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"532236a1-a04d-4b99-9131-b1d16a3d5ed4","Type":"ContainerStarted","Data":"428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646"} Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.431134 4846 generic.go:334] "Generic (PLEG): container finished" podID="e1c1ec69-7585-45e7-812b-df08a8ede792" containerID="b2355786fff0258ff67af38a657cb1150394658b2101b78593ceac673efc5f8f" exitCode=0 Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.431204 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" event={"ID":"e1c1ec69-7585-45e7-812b-df08a8ede792","Type":"ContainerDied","Data":"b2355786fff0258ff67af38a657cb1150394658b2101b78593ceac673efc5f8f"} Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.440462 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f74434a-8ddb-40fd-8ecb-637c3a63c919","Type":"ContainerStarted","Data":"4420fc9c386fffb528982e84b27202fab05e19c602905ca638a1c18fb64455d7"} Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.443812 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6447845b-zlnf5" event={"ID":"69522beb-f7d5-4694-ada8-0d17d7d3f16b","Type":"ContainerStarted","Data":"b358fa2e3b3d3040b293911322aa36237a5340d9d50ce3922695a3ff266a9b44"} Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.443866 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6447845b-zlnf5" event={"ID":"69522beb-f7d5-4694-ada8-0d17d7d3f16b","Type":"ContainerStarted","Data":"f357c7a0e61fbd03ef490e5638f0b35530ac99e278d44d9f752d0369a28df54b"} Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.444110 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.444142 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.458211 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.458185692 podStartE2EDuration="3.458185692s" podCreationTimestamp="2026-02-02 12:29:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:30:02.446029878 +0000 UTC m=+1233.674616741" watchObservedRunningTime="2026-02-02 12:30:02.458185692 +0000 UTC m=+1233.686772565" Feb 02 12:30:02 crc kubenswrapper[4846]: I0202 12:30:02.497235 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-f6447845b-zlnf5" podStartSLOduration=2.49721801 podStartE2EDuration="2.49721801s" podCreationTimestamp="2026-02-02 12:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:30:02.489536397 +0000 UTC m=+1233.718123260" watchObservedRunningTime="2026-02-02 12:30:02.49721801 +0000 UTC m=+1233.725804873" Feb 02 12:30:03 crc kubenswrapper[4846]: I0202 12:30:03.877152 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:03.999766 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1c1ec69-7585-45e7-812b-df08a8ede792-config-volume\") pod \"e1c1ec69-7585-45e7-812b-df08a8ede792\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.000199 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pq5n\" (UniqueName: \"kubernetes.io/projected/e1c1ec69-7585-45e7-812b-df08a8ede792-kube-api-access-9pq5n\") pod \"e1c1ec69-7585-45e7-812b-df08a8ede792\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.000308 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1c1ec69-7585-45e7-812b-df08a8ede792-secret-volume\") pod \"e1c1ec69-7585-45e7-812b-df08a8ede792\" (UID: \"e1c1ec69-7585-45e7-812b-df08a8ede792\") " Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.000567 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1c1ec69-7585-45e7-812b-df08a8ede792-config-volume" (OuterVolumeSpecName: "config-volume") pod "e1c1ec69-7585-45e7-812b-df08a8ede792" (UID: "e1c1ec69-7585-45e7-812b-df08a8ede792"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.000843 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1c1ec69-7585-45e7-812b-df08a8ede792-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.006211 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1c1ec69-7585-45e7-812b-df08a8ede792-kube-api-access-9pq5n" (OuterVolumeSpecName: "kube-api-access-9pq5n") pod "e1c1ec69-7585-45e7-812b-df08a8ede792" (UID: "e1c1ec69-7585-45e7-812b-df08a8ede792"). InnerVolumeSpecName "kube-api-access-9pq5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.020266 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c1ec69-7585-45e7-812b-df08a8ede792-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e1c1ec69-7585-45e7-812b-df08a8ede792" (UID: "e1c1ec69-7585-45e7-812b-df08a8ede792"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.102818 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pq5n\" (UniqueName: \"kubernetes.io/projected/e1c1ec69-7585-45e7-812b-df08a8ede792-kube-api-access-9pq5n\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.102860 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1c1ec69-7585-45e7-812b-df08a8ede792-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.474577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" event={"ID":"e1c1ec69-7585-45e7-812b-df08a8ede792","Type":"ContainerDied","Data":"926cdcef48bce76185177fec2b7b1727cb332b6082f4c1acf76b296164977a3e"} Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.474662 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="926cdcef48bce76185177fec2b7b1727cb332b6082f4c1acf76b296164977a3e" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.474731 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.582611 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.808056 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85c76b7df7-9b4xg"] Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.810918 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85c76b7df7-9b4xg" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerName="neutron-api" containerID="cri-o://d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6" gracePeriod=30 Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.811911 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85c76b7df7-9b4xg" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerName="neutron-httpd" containerID="cri-o://39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca" gracePeriod=30 Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.831883 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-85c76b7df7-9b4xg" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.153:9696/\": EOF" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.842327 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6497c67f7-24wxw"] Feb 02 12:30:04 crc kubenswrapper[4846]: E0202 12:30:04.842873 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c1ec69-7585-45e7-812b-df08a8ede792" containerName="collect-profiles" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.842892 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c1ec69-7585-45e7-812b-df08a8ede792" containerName="collect-profiles" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.843120 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c1ec69-7585-45e7-812b-df08a8ede792" containerName="collect-profiles" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.844307 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.855830 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6497c67f7-24wxw"] Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.924298 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-combined-ca-bundle\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.924356 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-config\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.924683 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-public-tls-certs\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.924812 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-ovndb-tls-certs\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.924868 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql92s\" (UniqueName: \"kubernetes.io/projected/5a1c1394-3713-41ee-948e-3800e5d51020-kube-api-access-ql92s\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.924987 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-internal-tls-certs\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.925056 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-httpd-config\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:04 crc kubenswrapper[4846]: I0202 12:30:04.958793 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.038982 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-public-tls-certs\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.039076 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-ovndb-tls-certs\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.039133 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql92s\" (UniqueName: \"kubernetes.io/projected/5a1c1394-3713-41ee-948e-3800e5d51020-kube-api-access-ql92s\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.039180 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-internal-tls-certs\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.039406 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-httpd-config\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.039438 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-combined-ca-bundle\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.039463 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-config\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.055806 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-config\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.056220 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-combined-ca-bundle\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.056816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-internal-tls-certs\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.057350 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-httpd-config\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.061286 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-ovndb-tls-certs\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.062230 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-public-tls-certs\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.083726 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql92s\" (UniqueName: \"kubernetes.io/projected/5a1c1394-3713-41ee-948e-3800e5d51020-kube-api-access-ql92s\") pod \"neutron-6497c67f7-24wxw\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.107993 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-9tl7z"] Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.108366 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" podUID="832895eb-b40b-427c-9485-166c86ef9d83" containerName="dnsmasq-dns" containerID="cri-o://df2bd6d33b3d7847a66799b3d337ef4de2e21f5a107b247cd084706377c989f2" gracePeriod=10 Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.189042 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.489940 4846 generic.go:334] "Generic (PLEG): container finished" podID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerID="39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca" exitCode=0 Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.490259 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85c76b7df7-9b4xg" event={"ID":"0374610f-a1fd-4650-8ad8-cc4455c06611","Type":"ContainerDied","Data":"39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca"} Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.491446 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.494338 4846 generic.go:334] "Generic (PLEG): container finished" podID="832895eb-b40b-427c-9485-166c86ef9d83" containerID="df2bd6d33b3d7847a66799b3d337ef4de2e21f5a107b247cd084706377c989f2" exitCode=0 Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.494374 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" event={"ID":"832895eb-b40b-427c-9485-166c86ef9d83","Type":"ContainerDied","Data":"df2bd6d33b3d7847a66799b3d337ef4de2e21f5a107b247cd084706377c989f2"} Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.543591 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.720615 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.824151 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.859554 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6497c67f7-24wxw"] Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.965158 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbjxg\" (UniqueName: \"kubernetes.io/projected/832895eb-b40b-427c-9485-166c86ef9d83-kube-api-access-rbjxg\") pod \"832895eb-b40b-427c-9485-166c86ef9d83\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.965301 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-svc\") pod \"832895eb-b40b-427c-9485-166c86ef9d83\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.965335 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-nb\") pod \"832895eb-b40b-427c-9485-166c86ef9d83\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.965441 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-config\") pod \"832895eb-b40b-427c-9485-166c86ef9d83\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.965487 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-swift-storage-0\") pod \"832895eb-b40b-427c-9485-166c86ef9d83\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.965529 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-sb\") pod \"832895eb-b40b-427c-9485-166c86ef9d83\" (UID: \"832895eb-b40b-427c-9485-166c86ef9d83\") " Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.978504 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/832895eb-b40b-427c-9485-166c86ef9d83-kube-api-access-rbjxg" (OuterVolumeSpecName: "kube-api-access-rbjxg") pod "832895eb-b40b-427c-9485-166c86ef9d83" (UID: "832895eb-b40b-427c-9485-166c86ef9d83"). InnerVolumeSpecName "kube-api-access-rbjxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:05 crc kubenswrapper[4846]: I0202 12:30:05.990429 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.069730 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbjxg\" (UniqueName: \"kubernetes.io/projected/832895eb-b40b-427c-9485-166c86ef9d83-kube-api-access-rbjxg\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.122483 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-config" (OuterVolumeSpecName: "config") pod "832895eb-b40b-427c-9485-166c86ef9d83" (UID: "832895eb-b40b-427c-9485-166c86ef9d83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.151690 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "832895eb-b40b-427c-9485-166c86ef9d83" (UID: "832895eb-b40b-427c-9485-166c86ef9d83"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.172142 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.172182 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.208933 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "832895eb-b40b-427c-9485-166c86ef9d83" (UID: "832895eb-b40b-427c-9485-166c86ef9d83"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.213179 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "832895eb-b40b-427c-9485-166c86ef9d83" (UID: "832895eb-b40b-427c-9485-166c86ef9d83"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.217921 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "832895eb-b40b-427c-9485-166c86ef9d83" (UID: "832895eb-b40b-427c-9485-166c86ef9d83"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.273651 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.273687 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.273700 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/832895eb-b40b-427c-9485-166c86ef9d83-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.505368 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" event={"ID":"832895eb-b40b-427c-9485-166c86ef9d83","Type":"ContainerDied","Data":"ed6716686f243ecf222e6cdb3c58848fa90b90e0f9602a12136b05341bedb39b"} Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.505421 4846 scope.go:117] "RemoveContainer" containerID="df2bd6d33b3d7847a66799b3d337ef4de2e21f5a107b247cd084706377c989f2" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.505530 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dd6887f5-9tl7z" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.513660 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6497c67f7-24wxw" event={"ID":"5a1c1394-3713-41ee-948e-3800e5d51020","Type":"ContainerStarted","Data":"f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292"} Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.513714 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6497c67f7-24wxw" event={"ID":"5a1c1394-3713-41ee-948e-3800e5d51020","Type":"ContainerStarted","Data":"38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd"} Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.513726 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6497c67f7-24wxw" event={"ID":"5a1c1394-3713-41ee-948e-3800e5d51020","Type":"ContainerStarted","Data":"64b75209ff5acfa45f93f1ea20834dcdc3631d0b716bc99475e3d63d64e1ad21"} Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.513943 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="87c361f2-a89b-475e-b485-55438adf0c71" containerName="cinder-scheduler" containerID="cri-o://c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e" gracePeriod=30 Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.513960 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="87c361f2-a89b-475e-b485-55438adf0c71" containerName="probe" containerID="cri-o://d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e" gracePeriod=30 Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.550530 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6497c67f7-24wxw" podStartSLOduration=2.5505117569999998 podStartE2EDuration="2.550511757s" podCreationTimestamp="2026-02-02 12:30:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:30:06.542650402 +0000 UTC m=+1237.771237265" watchObservedRunningTime="2026-02-02 12:30:06.550511757 +0000 UTC m=+1237.779098660" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.552217 4846 scope.go:117] "RemoveContainer" containerID="b06a7d3f575d6dadfe158b28c17ac82876b750ab3fb08efd49494832290fec5c" Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.585451 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-9tl7z"] Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.593006 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86dd6887f5-9tl7z"] Feb 02 12:30:06 crc kubenswrapper[4846]: I0202 12:30:06.979256 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-85c76b7df7-9b4xg" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.153:9696/\": dial tcp 10.217.0.153:9696: connect: connection refused" Feb 02 12:30:07 crc kubenswrapper[4846]: I0202 12:30:07.436804 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="832895eb-b40b-427c-9485-166c86ef9d83" path="/var/lib/kubelet/pods/832895eb-b40b-427c-9485-166c86ef9d83/volumes" Feb 02 12:30:07 crc kubenswrapper[4846]: I0202 12:30:07.523979 4846 generic.go:334] "Generic (PLEG): container finished" podID="87c361f2-a89b-475e-b485-55438adf0c71" containerID="d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e" exitCode=0 Feb 02 12:30:07 crc kubenswrapper[4846]: I0202 12:30:07.524047 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87c361f2-a89b-475e-b485-55438adf0c71","Type":"ContainerDied","Data":"d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e"} Feb 02 12:30:07 crc kubenswrapper[4846]: I0202 12:30:07.527452 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:08 crc kubenswrapper[4846]: I0202 12:30:08.538709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f74434a-8ddb-40fd-8ecb-637c3a63c919","Type":"ContainerStarted","Data":"7dc372b2f0c4719343735ced9862d29617e80edaf89ef2898292799a898c3e83"} Feb 02 12:30:08 crc kubenswrapper[4846]: I0202 12:30:08.564869 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9400171510000002 podStartE2EDuration="10.564851027s" podCreationTimestamp="2026-02-02 12:29:58 +0000 UTC" firstStartedPulling="2026-02-02 12:29:59.225119775 +0000 UTC m=+1230.453706638" lastFinishedPulling="2026-02-02 12:30:07.849953651 +0000 UTC m=+1239.078540514" observedRunningTime="2026-02-02 12:30:08.559017362 +0000 UTC m=+1239.787604225" watchObservedRunningTime="2026-02-02 12:30:08.564851027 +0000 UTC m=+1239.793437890" Feb 02 12:30:08 crc kubenswrapper[4846]: I0202 12:30:08.945286 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.022266 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data-custom\") pod \"87c361f2-a89b-475e-b485-55438adf0c71\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.022359 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-combined-ca-bundle\") pod \"87c361f2-a89b-475e-b485-55438adf0c71\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.022391 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv8qz\" (UniqueName: \"kubernetes.io/projected/87c361f2-a89b-475e-b485-55438adf0c71-kube-api-access-lv8qz\") pod \"87c361f2-a89b-475e-b485-55438adf0c71\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.022456 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-scripts\") pod \"87c361f2-a89b-475e-b485-55438adf0c71\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.022479 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data\") pod \"87c361f2-a89b-475e-b485-55438adf0c71\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.022588 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87c361f2-a89b-475e-b485-55438adf0c71-etc-machine-id\") pod \"87c361f2-a89b-475e-b485-55438adf0c71\" (UID: \"87c361f2-a89b-475e-b485-55438adf0c71\") " Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.022978 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87c361f2-a89b-475e-b485-55438adf0c71-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "87c361f2-a89b-475e-b485-55438adf0c71" (UID: "87c361f2-a89b-475e-b485-55438adf0c71"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.028426 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87c361f2-a89b-475e-b485-55438adf0c71-kube-api-access-lv8qz" (OuterVolumeSpecName: "kube-api-access-lv8qz") pod "87c361f2-a89b-475e-b485-55438adf0c71" (UID: "87c361f2-a89b-475e-b485-55438adf0c71"). InnerVolumeSpecName "kube-api-access-lv8qz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.029036 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "87c361f2-a89b-475e-b485-55438adf0c71" (UID: "87c361f2-a89b-475e-b485-55438adf0c71"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.030410 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-scripts" (OuterVolumeSpecName: "scripts") pod "87c361f2-a89b-475e-b485-55438adf0c71" (UID: "87c361f2-a89b-475e-b485-55438adf0c71"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.089886 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87c361f2-a89b-475e-b485-55438adf0c71" (UID: "87c361f2-a89b-475e-b485-55438adf0c71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.125089 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87c361f2-a89b-475e-b485-55438adf0c71-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.125120 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.125130 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.125139 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv8qz\" (UniqueName: \"kubernetes.io/projected/87c361f2-a89b-475e-b485-55438adf0c71-kube-api-access-lv8qz\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.125150 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.125912 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data" (OuterVolumeSpecName: "config-data") pod "87c361f2-a89b-475e-b485-55438adf0c71" (UID: "87c361f2-a89b-475e-b485-55438adf0c71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.226839 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87c361f2-a89b-475e-b485-55438adf0c71-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.548436 4846 generic.go:334] "Generic (PLEG): container finished" podID="87c361f2-a89b-475e-b485-55438adf0c71" containerID="c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e" exitCode=0 Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.548484 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87c361f2-a89b-475e-b485-55438adf0c71","Type":"ContainerDied","Data":"c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e"} Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.548541 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"87c361f2-a89b-475e-b485-55438adf0c71","Type":"ContainerDied","Data":"0ac8881f488d257a91d1cf9481cb05fc16d234cc06fe709015505bebaf77d57d"} Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.548566 4846 scope.go:117] "RemoveContainer" containerID="d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.548723 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.550038 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.582698 4846 scope.go:117] "RemoveContainer" containerID="c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.588657 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.603281 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.617736 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:30:09 crc kubenswrapper[4846]: E0202 12:30:09.618201 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c361f2-a89b-475e-b485-55438adf0c71" containerName="probe" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.618224 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c361f2-a89b-475e-b485-55438adf0c71" containerName="probe" Feb 02 12:30:09 crc kubenswrapper[4846]: E0202 12:30:09.618250 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87c361f2-a89b-475e-b485-55438adf0c71" containerName="cinder-scheduler" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.618259 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="87c361f2-a89b-475e-b485-55438adf0c71" containerName="cinder-scheduler" Feb 02 12:30:09 crc kubenswrapper[4846]: E0202 12:30:09.618277 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="832895eb-b40b-427c-9485-166c86ef9d83" containerName="dnsmasq-dns" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.618285 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="832895eb-b40b-427c-9485-166c86ef9d83" containerName="dnsmasq-dns" Feb 02 12:30:09 crc kubenswrapper[4846]: E0202 12:30:09.618297 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="832895eb-b40b-427c-9485-166c86ef9d83" containerName="init" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.618305 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="832895eb-b40b-427c-9485-166c86ef9d83" containerName="init" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.618504 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c361f2-a89b-475e-b485-55438adf0c71" containerName="cinder-scheduler" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.618525 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="87c361f2-a89b-475e-b485-55438adf0c71" containerName="probe" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.618537 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="832895eb-b40b-427c-9485-166c86ef9d83" containerName="dnsmasq-dns" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.620006 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.623609 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.630405 4846 scope.go:117] "RemoveContainer" containerID="d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.631447 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:30:09 crc kubenswrapper[4846]: E0202 12:30:09.634172 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e\": container with ID starting with d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e not found: ID does not exist" containerID="d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.634416 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e"} err="failed to get container status \"d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e\": rpc error: code = NotFound desc = could not find container \"d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e\": container with ID starting with d02fb019815765553d7f83a0f2827317396d0b91a5a89fa93befaf687e1ee95e not found: ID does not exist" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.634516 4846 scope.go:117] "RemoveContainer" containerID="c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e" Feb 02 12:30:09 crc kubenswrapper[4846]: E0202 12:30:09.638009 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e\": container with ID starting with c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e not found: ID does not exist" containerID="c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.638150 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e"} err="failed to get container status \"c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e\": rpc error: code = NotFound desc = could not find container \"c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e\": container with ID starting with c2f4e2ad7628fc0805947a06a76541ce28380a6bf119087ee0673d5eb3b6e34e not found: ID does not exist" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.733646 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.734141 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.734299 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9tf7\" (UniqueName: \"kubernetes.io/projected/444e13f0-bc18-4bc9-8d92-af6a379e22ab-kube-api-access-l9tf7\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.734556 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/444e13f0-bc18-4bc9-8d92-af6a379e22ab-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.734743 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.734879 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.836821 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.837236 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.837270 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9tf7\" (UniqueName: \"kubernetes.io/projected/444e13f0-bc18-4bc9-8d92-af6a379e22ab-kube-api-access-l9tf7\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.837339 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/444e13f0-bc18-4bc9-8d92-af6a379e22ab-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.837373 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.837423 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.839408 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/444e13f0-bc18-4bc9-8d92-af6a379e22ab-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.848325 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.849365 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.856090 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.856224 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.865957 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9tf7\" (UniqueName: \"kubernetes.io/projected/444e13f0-bc18-4bc9-8d92-af6a379e22ab-kube-api-access-l9tf7\") pod \"cinder-scheduler-0\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " pod="openstack/cinder-scheduler-0" Feb 02 12:30:09 crc kubenswrapper[4846]: I0202 12:30:09.935524 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.455219 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.563694 4846 generic.go:334] "Generic (PLEG): container finished" podID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerID="d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6" exitCode=0 Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.563798 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85c76b7df7-9b4xg" event={"ID":"0374610f-a1fd-4650-8ad8-cc4455c06611","Type":"ContainerDied","Data":"d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6"} Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.563853 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85c76b7df7-9b4xg" event={"ID":"0374610f-a1fd-4650-8ad8-cc4455c06611","Type":"ContainerDied","Data":"e8699d00e2a7950dfaf6b63f50aa81c359403d573b6117eacfab24ad2a6ff406"} Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.563850 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85c76b7df7-9b4xg" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.563896 4846 scope.go:117] "RemoveContainer" containerID="39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.613096 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.627535 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.636138 4846 scope.go:117] "RemoveContainer" containerID="d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6" Feb 02 12:30:10 crc kubenswrapper[4846]: W0202 12:30:10.639418 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod444e13f0_bc18_4bc9_8d92_af6a379e22ab.slice/crio-3b49c7fefa01d749cd32612414a9a283bf51dc9e40af50dd67a7ce30bc8446f4 WatchSource:0}: Error finding container 3b49c7fefa01d749cd32612414a9a283bf51dc9e40af50dd67a7ce30bc8446f4: Status 404 returned error can't find the container with id 3b49c7fefa01d749cd32612414a9a283bf51dc9e40af50dd67a7ce30bc8446f4 Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.655799 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-combined-ca-bundle\") pod \"0374610f-a1fd-4650-8ad8-cc4455c06611\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.656375 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-ovndb-tls-certs\") pod \"0374610f-a1fd-4650-8ad8-cc4455c06611\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.656539 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86rcl\" (UniqueName: \"kubernetes.io/projected/0374610f-a1fd-4650-8ad8-cc4455c06611-kube-api-access-86rcl\") pod \"0374610f-a1fd-4650-8ad8-cc4455c06611\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.656727 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-httpd-config\") pod \"0374610f-a1fd-4650-8ad8-cc4455c06611\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.656854 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-internal-tls-certs\") pod \"0374610f-a1fd-4650-8ad8-cc4455c06611\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.656982 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-config\") pod \"0374610f-a1fd-4650-8ad8-cc4455c06611\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.657074 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-public-tls-certs\") pod \"0374610f-a1fd-4650-8ad8-cc4455c06611\" (UID: \"0374610f-a1fd-4650-8ad8-cc4455c06611\") " Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.666208 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "0374610f-a1fd-4650-8ad8-cc4455c06611" (UID: "0374610f-a1fd-4650-8ad8-cc4455c06611"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.668042 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0374610f-a1fd-4650-8ad8-cc4455c06611-kube-api-access-86rcl" (OuterVolumeSpecName: "kube-api-access-86rcl") pod "0374610f-a1fd-4650-8ad8-cc4455c06611" (UID: "0374610f-a1fd-4650-8ad8-cc4455c06611"). InnerVolumeSpecName "kube-api-access-86rcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.733722 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-config" (OuterVolumeSpecName: "config") pod "0374610f-a1fd-4650-8ad8-cc4455c06611" (UID: "0374610f-a1fd-4650-8ad8-cc4455c06611"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.759778 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86rcl\" (UniqueName: \"kubernetes.io/projected/0374610f-a1fd-4650-8ad8-cc4455c06611-kube-api-access-86rcl\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.761219 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.761248 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.763781 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0374610f-a1fd-4650-8ad8-cc4455c06611" (UID: "0374610f-a1fd-4650-8ad8-cc4455c06611"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.809387 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0374610f-a1fd-4650-8ad8-cc4455c06611" (UID: "0374610f-a1fd-4650-8ad8-cc4455c06611"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.845417 4846 scope.go:117] "RemoveContainer" containerID="39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca" Feb 02 12:30:10 crc kubenswrapper[4846]: E0202 12:30:10.846898 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca\": container with ID starting with 39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca not found: ID does not exist" containerID="39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.846960 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca"} err="failed to get container status \"39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca\": rpc error: code = NotFound desc = could not find container \"39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca\": container with ID starting with 39bc490a4fcce77cadf858a3fe25ceb583d569d03d4154ae655c6bea873a13ca not found: ID does not exist" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.846991 4846 scope.go:117] "RemoveContainer" containerID="d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6" Feb 02 12:30:10 crc kubenswrapper[4846]: E0202 12:30:10.847660 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6\": container with ID starting with d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6 not found: ID does not exist" containerID="d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.847676 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6"} err="failed to get container status \"d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6\": rpc error: code = NotFound desc = could not find container \"d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6\": container with ID starting with d9b4fd0de7885fdc2a998d0029240677e4309217fb6f5ab6f97d7e5ce5f0dbc6 not found: ID does not exist" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.848184 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "0374610f-a1fd-4650-8ad8-cc4455c06611" (UID: "0374610f-a1fd-4650-8ad8-cc4455c06611"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.848738 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0374610f-a1fd-4650-8ad8-cc4455c06611" (UID: "0374610f-a1fd-4650-8ad8-cc4455c06611"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.862573 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.862613 4846 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.862640 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.862649 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0374610f-a1fd-4650-8ad8-cc4455c06611-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.896486 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.917270 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85c76b7df7-9b4xg"] Feb 02 12:30:10 crc kubenswrapper[4846]: I0202 12:30:10.934403 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-85c76b7df7-9b4xg"] Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.293826 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c75698f54-h7x5d"] Feb 02 12:30:11 crc kubenswrapper[4846]: E0202 12:30:11.295352 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerName="neutron-api" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.295801 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerName="neutron-api" Feb 02 12:30:11 crc kubenswrapper[4846]: E0202 12:30:11.295894 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerName="neutron-httpd" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.295954 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerName="neutron-httpd" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.296384 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerName="neutron-api" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.296465 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" containerName="neutron-httpd" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.298129 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.340458 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c75698f54-h7x5d"] Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.381817 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-public-tls-certs\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.381883 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-internal-tls-certs\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.381901 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-combined-ca-bundle\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.381923 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-scripts\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.381944 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps9lh\" (UniqueName: \"kubernetes.io/projected/525708d2-15d1-4bfb-8258-bf216ebc4bab-kube-api-access-ps9lh\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.381972 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/525708d2-15d1-4bfb-8258-bf216ebc4bab-logs\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.382029 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-config-data\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.443409 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0374610f-a1fd-4650-8ad8-cc4455c06611" path="/var/lib/kubelet/pods/0374610f-a1fd-4650-8ad8-cc4455c06611/volumes" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.444213 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87c361f2-a89b-475e-b485-55438adf0c71" path="/var/lib/kubelet/pods/87c361f2-a89b-475e-b485-55438adf0c71/volumes" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.485587 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-internal-tls-certs\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.485898 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-combined-ca-bundle\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.485982 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-scripts\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.486067 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps9lh\" (UniqueName: \"kubernetes.io/projected/525708d2-15d1-4bfb-8258-bf216ebc4bab-kube-api-access-ps9lh\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.486150 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/525708d2-15d1-4bfb-8258-bf216ebc4bab-logs\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.486713 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-config-data\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.486904 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-public-tls-certs\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.487445 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/525708d2-15d1-4bfb-8258-bf216ebc4bab-logs\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.497670 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-internal-tls-certs\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.497752 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-config-data\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.497989 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-scripts\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.498343 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-combined-ca-bundle\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.508141 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-public-tls-certs\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.515692 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps9lh\" (UniqueName: \"kubernetes.io/projected/525708d2-15d1-4bfb-8258-bf216ebc4bab-kube-api-access-ps9lh\") pod \"placement-c75698f54-h7x5d\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.579833 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"444e13f0-bc18-4bc9-8d92-af6a379e22ab","Type":"ContainerStarted","Data":"7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806"} Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.579887 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"444e13f0-bc18-4bc9-8d92-af6a379e22ab","Type":"ContainerStarted","Data":"3b49c7fefa01d749cd32612414a9a283bf51dc9e40af50dd67a7ce30bc8446f4"} Feb 02 12:30:11 crc kubenswrapper[4846]: I0202 12:30:11.651653 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.261433 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c75698f54-h7x5d"] Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.593558 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c75698f54-h7x5d" event={"ID":"525708d2-15d1-4bfb-8258-bf216ebc4bab","Type":"ContainerStarted","Data":"4f2c0bed69df185a3aca7bd0bcabf6f5afe33072cfbe8448804ba379ea4ff18e"} Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.593611 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c75698f54-h7x5d" event={"ID":"525708d2-15d1-4bfb-8258-bf216ebc4bab","Type":"ContainerStarted","Data":"359771b6d080e100102b52000fea7f8597862be6e2195eb43d0b8aab5071233c"} Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.602104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"444e13f0-bc18-4bc9-8d92-af6a379e22ab","Type":"ContainerStarted","Data":"642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210"} Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.626998 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.626973529 podStartE2EDuration="3.626973529s" podCreationTimestamp="2026-02-02 12:30:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:30:12.622602569 +0000 UTC m=+1243.851189442" watchObservedRunningTime="2026-02-02 12:30:12.626973529 +0000 UTC m=+1243.855560402" Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.632207 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.777279 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.826114 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.856751 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f59fdf6f6-rsrcc"] Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.857211 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerName="barbican-api-log" containerID="cri-o://616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5" gracePeriod=30 Feb 02 12:30:12 crc kubenswrapper[4846]: I0202 12:30:12.857956 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerName="barbican-api" containerID="cri-o://662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae" gracePeriod=30 Feb 02 12:30:13 crc kubenswrapper[4846]: I0202 12:30:13.618537 4846 generic.go:334] "Generic (PLEG): container finished" podID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerID="616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5" exitCode=143 Feb 02 12:30:13 crc kubenswrapper[4846]: I0202 12:30:13.618791 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" event={"ID":"a4500640-81f0-49dd-a0f4-e9b5a4e211b0","Type":"ContainerDied","Data":"616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5"} Feb 02 12:30:13 crc kubenswrapper[4846]: I0202 12:30:13.625368 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c75698f54-h7x5d" event={"ID":"525708d2-15d1-4bfb-8258-bf216ebc4bab","Type":"ContainerStarted","Data":"f7527fb5dfdb0a8afb41a261ae73e292db1a4a3cb90a3972583cde5f54856e3b"} Feb 02 12:30:13 crc kubenswrapper[4846]: I0202 12:30:13.625429 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:13 crc kubenswrapper[4846]: I0202 12:30:13.625442 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:13 crc kubenswrapper[4846]: I0202 12:30:13.652140 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-c75698f54-h7x5d" podStartSLOduration=2.652118635 podStartE2EDuration="2.652118635s" podCreationTimestamp="2026-02-02 12:30:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:30:13.643925021 +0000 UTC m=+1244.872511894" watchObservedRunningTime="2026-02-02 12:30:13.652118635 +0000 UTC m=+1244.880705498" Feb 02 12:30:14 crc kubenswrapper[4846]: I0202 12:30:14.936982 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 02 12:30:15 crc kubenswrapper[4846]: I0202 12:30:15.365216 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.162999 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:59080->10.217.0.159:9311: read: connection reset by peer" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.163237 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:59066->10.217.0.159:9311: read: connection reset by peer" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.580613 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.647081 4846 generic.go:334] "Generic (PLEG): container finished" podID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerID="662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae" exitCode=0 Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.647120 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" event={"ID":"a4500640-81f0-49dd-a0f4-e9b5a4e211b0","Type":"ContainerDied","Data":"662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae"} Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.647130 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.647148 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f59fdf6f6-rsrcc" event={"ID":"a4500640-81f0-49dd-a0f4-e9b5a4e211b0","Type":"ContainerDied","Data":"e1397205faac1ccd81752df80c4a7f4c972080add1cc8d8c3af334e9b4d74a0a"} Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.647165 4846 scope.go:117] "RemoveContainer" containerID="662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.668556 4846 scope.go:117] "RemoveContainer" containerID="616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.688027 4846 scope.go:117] "RemoveContainer" containerID="662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae" Feb 02 12:30:16 crc kubenswrapper[4846]: E0202 12:30:16.688534 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae\": container with ID starting with 662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae not found: ID does not exist" containerID="662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.688578 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae"} err="failed to get container status \"662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae\": rpc error: code = NotFound desc = could not find container \"662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae\": container with ID starting with 662ca3e7799664829f50933a3e7fa2b1be67050bcd806b56bd80489417caacae not found: ID does not exist" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.688605 4846 scope.go:117] "RemoveContainer" containerID="616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5" Feb 02 12:30:16 crc kubenswrapper[4846]: E0202 12:30:16.689157 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5\": container with ID starting with 616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5 not found: ID does not exist" containerID="616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.689207 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5"} err="failed to get container status \"616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5\": rpc error: code = NotFound desc = could not find container \"616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5\": container with ID starting with 616ccd95f616f7313fd5d3cd5b2f40c08bca89904664cc1e9390f06230a370e5 not found: ID does not exist" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.719792 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data\") pod \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.719937 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx95l\" (UniqueName: \"kubernetes.io/projected/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-kube-api-access-tx95l\") pod \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.719967 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data-custom\") pod \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.720017 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-logs\") pod \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.720081 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-combined-ca-bundle\") pod \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\" (UID: \"a4500640-81f0-49dd-a0f4-e9b5a4e211b0\") " Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.720453 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-logs" (OuterVolumeSpecName: "logs") pod "a4500640-81f0-49dd-a0f4-e9b5a4e211b0" (UID: "a4500640-81f0-49dd-a0f4-e9b5a4e211b0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.720550 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.725572 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a4500640-81f0-49dd-a0f4-e9b5a4e211b0" (UID: "a4500640-81f0-49dd-a0f4-e9b5a4e211b0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.741581 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-kube-api-access-tx95l" (OuterVolumeSpecName: "kube-api-access-tx95l") pod "a4500640-81f0-49dd-a0f4-e9b5a4e211b0" (UID: "a4500640-81f0-49dd-a0f4-e9b5a4e211b0"). InnerVolumeSpecName "kube-api-access-tx95l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.745257 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4500640-81f0-49dd-a0f4-e9b5a4e211b0" (UID: "a4500640-81f0-49dd-a0f4-e9b5a4e211b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.776314 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data" (OuterVolumeSpecName: "config-data") pod "a4500640-81f0-49dd-a0f4-e9b5a4e211b0" (UID: "a4500640-81f0-49dd-a0f4-e9b5a4e211b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.822692 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.822725 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx95l\" (UniqueName: \"kubernetes.io/projected/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-kube-api-access-tx95l\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.822737 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.822748 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4500640-81f0-49dd-a0f4-e9b5a4e211b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.979596 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f59fdf6f6-rsrcc"] Feb 02 12:30:16 crc kubenswrapper[4846]: I0202 12:30:16.989007 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6f59fdf6f6-rsrcc"] Feb 02 12:30:17 crc kubenswrapper[4846]: I0202 12:30:17.433746 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" path="/var/lib/kubelet/pods/a4500640-81f0-49dd-a0f4-e9b5a4e211b0/volumes" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.929406 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-68f9885b6f-c5hz4"] Feb 02 12:30:18 crc kubenswrapper[4846]: E0202 12:30:18.930016 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerName="barbican-api-log" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.930030 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerName="barbican-api-log" Feb 02 12:30:18 crc kubenswrapper[4846]: E0202 12:30:18.930057 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerName="barbican-api" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.930063 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerName="barbican-api" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.930205 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerName="barbican-api-log" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.930221 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4500640-81f0-49dd-a0f4-e9b5a4e211b0" containerName="barbican-api" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.931098 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.933048 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.933120 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.933150 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.949171 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-68f9885b6f-c5hz4"] Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.979387 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-run-httpd\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.979575 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-public-tls-certs\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.979709 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-combined-ca-bundle\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.979779 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-internal-tls-certs\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.979798 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-config-data\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.979942 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-log-httpd\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.979963 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvlnm\" (UniqueName: \"kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-kube-api-access-dvlnm\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:18 crc kubenswrapper[4846]: I0202 12:30:18.979994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-etc-swift\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.081331 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-run-httpd\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.081414 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-public-tls-certs\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.081479 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-combined-ca-bundle\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.081875 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-run-httpd\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.082850 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-internal-tls-certs\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.082889 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-config-data\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.083055 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-log-httpd\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.083078 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvlnm\" (UniqueName: \"kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-kube-api-access-dvlnm\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.083116 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-etc-swift\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.083576 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-log-httpd\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.086951 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-combined-ca-bundle\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.086957 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-public-tls-certs\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.087277 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-etc-swift\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.087293 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-internal-tls-certs\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.089202 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-config-data\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.100696 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvlnm\" (UniqueName: \"kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-kube-api-access-dvlnm\") pod \"swift-proxy-68f9885b6f-c5hz4\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.248506 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.639777 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.643179 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.644930 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.645223 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-4h4g7" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.646205 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.651744 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.707072 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.707160 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd6g4\" (UniqueName: \"kubernetes.io/projected/9fe862de-3e44-4a34-a6b6-4c6225a13839-kube-api-access-gd6g4\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.707195 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.707230 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config-secret\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.809060 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.809163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd6g4\" (UniqueName: \"kubernetes.io/projected/9fe862de-3e44-4a34-a6b6-4c6225a13839-kube-api-access-gd6g4\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.809199 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.809238 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config-secret\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.810437 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.814070 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config-secret\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.816172 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.833281 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-68f9885b6f-c5hz4"] Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.834420 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd6g4\" (UniqueName: \"kubernetes.io/projected/9fe862de-3e44-4a34-a6b6-4c6225a13839-kube-api-access-gd6g4\") pod \"openstackclient\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.863128 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.864442 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.876106 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.906678 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.908180 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 12:30:19 crc kubenswrapper[4846]: I0202 12:30:19.920788 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.012476 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt6kt\" (UniqueName: \"kubernetes.io/projected/ba822b20-88d5-404b-850c-7f9441933baf-kube-api-access-vt6kt\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.012665 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config-secret\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.012804 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.012942 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: E0202 12:30:20.050130 4846 log.go:32] "RunPodSandbox from runtime service failed" err=< Feb 02 12:30:20 crc kubenswrapper[4846]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_9fe862de-3e44-4a34-a6b6-4c6225a13839_0(abd46d042c26627872e5ec023d4562fdc29dd93b003fe3ab8cff7c02ce75b09a): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"abd46d042c26627872e5ec023d4562fdc29dd93b003fe3ab8cff7c02ce75b09a" Netns:"/var/run/netns/037d9be8-37c4-4ad2-b082-79053350dabc" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=abd46d042c26627872e5ec023d4562fdc29dd93b003fe3ab8cff7c02ce75b09a;K8S_POD_UID=9fe862de-3e44-4a34-a6b6-4c6225a13839" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/9fe862de-3e44-4a34-a6b6-4c6225a13839]: expected pod UID "9fe862de-3e44-4a34-a6b6-4c6225a13839" but got "ba822b20-88d5-404b-850c-7f9441933baf" from Kube API Feb 02 12:30:20 crc kubenswrapper[4846]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 02 12:30:20 crc kubenswrapper[4846]: > Feb 02 12:30:20 crc kubenswrapper[4846]: E0202 12:30:20.050224 4846 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Feb 02 12:30:20 crc kubenswrapper[4846]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_9fe862de-3e44-4a34-a6b6-4c6225a13839_0(abd46d042c26627872e5ec023d4562fdc29dd93b003fe3ab8cff7c02ce75b09a): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"abd46d042c26627872e5ec023d4562fdc29dd93b003fe3ab8cff7c02ce75b09a" Netns:"/var/run/netns/037d9be8-37c4-4ad2-b082-79053350dabc" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=abd46d042c26627872e5ec023d4562fdc29dd93b003fe3ab8cff7c02ce75b09a;K8S_POD_UID=9fe862de-3e44-4a34-a6b6-4c6225a13839" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/9fe862de-3e44-4a34-a6b6-4c6225a13839]: expected pod UID "9fe862de-3e44-4a34-a6b6-4c6225a13839" but got "ba822b20-88d5-404b-850c-7f9441933baf" from Kube API Feb 02 12:30:20 crc kubenswrapper[4846]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 02 12:30:20 crc kubenswrapper[4846]: > pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.115373 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.115468 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.115504 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt6kt\" (UniqueName: \"kubernetes.io/projected/ba822b20-88d5-404b-850c-7f9441933baf-kube-api-access-vt6kt\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.115612 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config-secret\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.120012 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.121885 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config-secret\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.124498 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.140781 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt6kt\" (UniqueName: \"kubernetes.io/projected/ba822b20-88d5-404b-850c-7f9441933baf-kube-api-access-vt6kt\") pod \"openstackclient\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.165063 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.234048 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: W0202 12:30:20.690176 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba822b20_88d5_404b_850c_7f9441933baf.slice/crio-65df77da72dd654f39cea38818549aa37d5e678426e95bccd75aacdd891f9f4e WatchSource:0}: Error finding container 65df77da72dd654f39cea38818549aa37d5e678426e95bccd75aacdd891f9f4e: Status 404 returned error can't find the container with id 65df77da72dd654f39cea38818549aa37d5e678426e95bccd75aacdd891f9f4e Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.694052 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.726388 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68f9885b6f-c5hz4" event={"ID":"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b","Type":"ContainerStarted","Data":"fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d"} Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.726441 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68f9885b6f-c5hz4" event={"ID":"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b","Type":"ContainerStarted","Data":"a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce"} Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.726461 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.726475 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.726485 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68f9885b6f-c5hz4" event={"ID":"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b","Type":"ContainerStarted","Data":"efff0ccd3dbde79a2cc68d31b0063ea27185ac79942fa5cddcebdbd453a3635e"} Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.728661 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.729398 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ba822b20-88d5-404b-850c-7f9441933baf","Type":"ContainerStarted","Data":"65df77da72dd654f39cea38818549aa37d5e678426e95bccd75aacdd891f9f4e"} Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.739237 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.745570 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9fe862de-3e44-4a34-a6b6-4c6225a13839" podUID="ba822b20-88d5-404b-850c-7f9441933baf" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.747077 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-68f9885b6f-c5hz4" podStartSLOduration=2.747066078 podStartE2EDuration="2.747066078s" podCreationTimestamp="2026-02-02 12:30:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:30:20.742710199 +0000 UTC m=+1251.971297062" watchObservedRunningTime="2026-02-02 12:30:20.747066078 +0000 UTC m=+1251.975652941" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.829327 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config-secret\") pod \"9fe862de-3e44-4a34-a6b6-4c6225a13839\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.829766 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-combined-ca-bundle\") pod \"9fe862de-3e44-4a34-a6b6-4c6225a13839\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.829833 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd6g4\" (UniqueName: \"kubernetes.io/projected/9fe862de-3e44-4a34-a6b6-4c6225a13839-kube-api-access-gd6g4\") pod \"9fe862de-3e44-4a34-a6b6-4c6225a13839\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.829888 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config\") pod \"9fe862de-3e44-4a34-a6b6-4c6225a13839\" (UID: \"9fe862de-3e44-4a34-a6b6-4c6225a13839\") " Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.831420 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9fe862de-3e44-4a34-a6b6-4c6225a13839" (UID: "9fe862de-3e44-4a34-a6b6-4c6225a13839"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.836120 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fe862de-3e44-4a34-a6b6-4c6225a13839-kube-api-access-gd6g4" (OuterVolumeSpecName: "kube-api-access-gd6g4") pod "9fe862de-3e44-4a34-a6b6-4c6225a13839" (UID: "9fe862de-3e44-4a34-a6b6-4c6225a13839"). InnerVolumeSpecName "kube-api-access-gd6g4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.836747 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9fe862de-3e44-4a34-a6b6-4c6225a13839" (UID: "9fe862de-3e44-4a34-a6b6-4c6225a13839"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.838090 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fe862de-3e44-4a34-a6b6-4c6225a13839" (UID: "9fe862de-3e44-4a34-a6b6-4c6225a13839"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.931766 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.931803 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fe862de-3e44-4a34-a6b6-4c6225a13839-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.931812 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd6g4\" (UniqueName: \"kubernetes.io/projected/9fe862de-3e44-4a34-a6b6-4c6225a13839-kube-api-access-gd6g4\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:20 crc kubenswrapper[4846]: I0202 12:30:20.931820 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fe862de-3e44-4a34-a6b6-4c6225a13839-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.448208 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fe862de-3e44-4a34-a6b6-4c6225a13839" path="/var/lib/kubelet/pods/9fe862de-3e44-4a34-a6b6-4c6225a13839/volumes" Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.448786 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.449066 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="ceilometer-central-agent" containerID="cri-o://951f36c17c01a2fa6e3011d452524eeca1d82b0bd91bd8c6634e1795396413b0" gracePeriod=30 Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.450035 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="sg-core" containerID="cri-o://4420fc9c386fffb528982e84b27202fab05e19c602905ca638a1c18fb64455d7" gracePeriod=30 Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.450375 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="proxy-httpd" containerID="cri-o://7dc372b2f0c4719343735ced9862d29617e80edaf89ef2898292799a898c3e83" gracePeriod=30 Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.450426 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="ceilometer-notification-agent" containerID="cri-o://21ccdd76b8af8a5c289f14aa6a68389288abd0458ad44f6858c41a788230a56b" gracePeriod=30 Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.463091 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.163:3000/\": EOF" Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.756518 4846 generic.go:334] "Generic (PLEG): container finished" podID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerID="7dc372b2f0c4719343735ced9862d29617e80edaf89ef2898292799a898c3e83" exitCode=0 Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.756550 4846 generic.go:334] "Generic (PLEG): container finished" podID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerID="4420fc9c386fffb528982e84b27202fab05e19c602905ca638a1c18fb64455d7" exitCode=2 Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.757454 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f74434a-8ddb-40fd-8ecb-637c3a63c919","Type":"ContainerDied","Data":"7dc372b2f0c4719343735ced9862d29617e80edaf89ef2898292799a898c3e83"} Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.757488 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f74434a-8ddb-40fd-8ecb-637c3a63c919","Type":"ContainerDied","Data":"4420fc9c386fffb528982e84b27202fab05e19c602905ca638a1c18fb64455d7"} Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.757515 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 12:30:21 crc kubenswrapper[4846]: I0202 12:30:21.763258 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9fe862de-3e44-4a34-a6b6-4c6225a13839" podUID="ba822b20-88d5-404b-850c-7f9441933baf" Feb 02 12:30:22 crc kubenswrapper[4846]: I0202 12:30:22.770792 4846 generic.go:334] "Generic (PLEG): container finished" podID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerID="951f36c17c01a2fa6e3011d452524eeca1d82b0bd91bd8c6634e1795396413b0" exitCode=0 Feb 02 12:30:22 crc kubenswrapper[4846]: I0202 12:30:22.770830 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f74434a-8ddb-40fd-8ecb-637c3a63c919","Type":"ContainerDied","Data":"951f36c17c01a2fa6e3011d452524eeca1d82b0bd91bd8c6634e1795396413b0"} Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.783668 4846 generic.go:334] "Generic (PLEG): container finished" podID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerID="21ccdd76b8af8a5c289f14aa6a68389288abd0458ad44f6858c41a788230a56b" exitCode=0 Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.783754 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f74434a-8ddb-40fd-8ecb-637c3a63c919","Type":"ContainerDied","Data":"21ccdd76b8af8a5c289f14aa6a68389288abd0458ad44f6858c41a788230a56b"} Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.938398 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.987405 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-combined-ca-bundle\") pod \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.987487 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-run-httpd\") pod \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.987541 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-config-data\") pod \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.987593 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-sg-core-conf-yaml\") pod \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.987691 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-log-httpd\") pod \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.987769 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb6ts\" (UniqueName: \"kubernetes.io/projected/2f74434a-8ddb-40fd-8ecb-637c3a63c919-kube-api-access-gb6ts\") pod \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.987820 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-scripts\") pod \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.988112 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2f74434a-8ddb-40fd-8ecb-637c3a63c919" (UID: "2f74434a-8ddb-40fd-8ecb-637c3a63c919"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.988294 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2f74434a-8ddb-40fd-8ecb-637c3a63c919" (UID: "2f74434a-8ddb-40fd-8ecb-637c3a63c919"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:23 crc kubenswrapper[4846]: I0202 12:30:23.992757 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f74434a-8ddb-40fd-8ecb-637c3a63c919-kube-api-access-gb6ts" (OuterVolumeSpecName: "kube-api-access-gb6ts") pod "2f74434a-8ddb-40fd-8ecb-637c3a63c919" (UID: "2f74434a-8ddb-40fd-8ecb-637c3a63c919"). InnerVolumeSpecName "kube-api-access-gb6ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.009439 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-scripts" (OuterVolumeSpecName: "scripts") pod "2f74434a-8ddb-40fd-8ecb-637c3a63c919" (UID: "2f74434a-8ddb-40fd-8ecb-637c3a63c919"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.019418 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2f74434a-8ddb-40fd-8ecb-637c3a63c919" (UID: "2f74434a-8ddb-40fd-8ecb-637c3a63c919"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.090573 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f74434a-8ddb-40fd-8ecb-637c3a63c919" (UID: "2f74434a-8ddb-40fd-8ecb-637c3a63c919"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.091162 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-combined-ca-bundle\") pod \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\" (UID: \"2f74434a-8ddb-40fd-8ecb-637c3a63c919\") " Feb 02 12:30:24 crc kubenswrapper[4846]: W0202 12:30:24.091314 4846 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/2f74434a-8ddb-40fd-8ecb-637c3a63c919/volumes/kubernetes.io~secret/combined-ca-bundle Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.091333 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f74434a-8ddb-40fd-8ecb-637c3a63c919" (UID: "2f74434a-8ddb-40fd-8ecb-637c3a63c919"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.092298 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.092369 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb6ts\" (UniqueName: \"kubernetes.io/projected/2f74434a-8ddb-40fd-8ecb-637c3a63c919-kube-api-access-gb6ts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.092385 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.092397 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.092412 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f74434a-8ddb-40fd-8ecb-637c3a63c919-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.092424 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.118666 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-config-data" (OuterVolumeSpecName: "config-data") pod "2f74434a-8ddb-40fd-8ecb-637c3a63c919" (UID: "2f74434a-8ddb-40fd-8ecb-637c3a63c919"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.194005 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f74434a-8ddb-40fd-8ecb-637c3a63c919-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.797442 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f74434a-8ddb-40fd-8ecb-637c3a63c919","Type":"ContainerDied","Data":"531541ff093aac860ab0078d5ef28a6104f5f755676ac72d600ac38cd351dea1"} Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.797504 4846 scope.go:117] "RemoveContainer" containerID="7dc372b2f0c4719343735ced9862d29617e80edaf89ef2898292799a898c3e83" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.797514 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.832337 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.849977 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.860557 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:24 crc kubenswrapper[4846]: E0202 12:30:24.861116 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="proxy-httpd" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.861140 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="proxy-httpd" Feb 02 12:30:24 crc kubenswrapper[4846]: E0202 12:30:24.861165 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="ceilometer-central-agent" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.861173 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="ceilometer-central-agent" Feb 02 12:30:24 crc kubenswrapper[4846]: E0202 12:30:24.861188 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="sg-core" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.861196 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="sg-core" Feb 02 12:30:24 crc kubenswrapper[4846]: E0202 12:30:24.861236 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="ceilometer-notification-agent" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.861246 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="ceilometer-notification-agent" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.861449 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="proxy-httpd" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.861470 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="ceilometer-central-agent" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.861487 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="sg-core" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.861508 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" containerName="ceilometer-notification-agent" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.865286 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.870118 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.872405 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.896684 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.905360 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.905419 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjmwr\" (UniqueName: \"kubernetes.io/projected/4264e2ff-9f25-41df-beeb-487e9378f228-kube-api-access-hjmwr\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.905542 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.905716 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-scripts\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.905834 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-config-data\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.905905 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-run-httpd\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:24 crc kubenswrapper[4846]: I0202 12:30:24.905945 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-log-httpd\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.007916 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-scripts\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.008345 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-config-data\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.008866 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-run-httpd\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.008914 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-log-httpd\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.008959 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.008982 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjmwr\" (UniqueName: \"kubernetes.io/projected/4264e2ff-9f25-41df-beeb-487e9378f228-kube-api-access-hjmwr\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.009011 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.009386 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-run-httpd\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.009681 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-log-httpd\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.013281 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.013734 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-config-data\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.014806 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.022512 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-scripts\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.024686 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjmwr\" (UniqueName: \"kubernetes.io/projected/4264e2ff-9f25-41df-beeb-487e9378f228-kube-api-access-hjmwr\") pod \"ceilometer-0\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.190019 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:25 crc kubenswrapper[4846]: I0202 12:30:25.435852 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f74434a-8ddb-40fd-8ecb-637c3a63c919" path="/var/lib/kubelet/pods/2f74434a-8ddb-40fd-8ecb-637c3a63c919/volumes" Feb 02 12:30:29 crc kubenswrapper[4846]: I0202 12:30:29.115581 4846 scope.go:117] "RemoveContainer" containerID="4420fc9c386fffb528982e84b27202fab05e19c602905ca638a1c18fb64455d7" Feb 02 12:30:29 crc kubenswrapper[4846]: I0202 12:30:29.189449 4846 scope.go:117] "RemoveContainer" containerID="21ccdd76b8af8a5c289f14aa6a68389288abd0458ad44f6858c41a788230a56b" Feb 02 12:30:29 crc kubenswrapper[4846]: I0202 12:30:29.258156 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:29 crc kubenswrapper[4846]: I0202 12:30:29.261828 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:30:29 crc kubenswrapper[4846]: I0202 12:30:29.338757 4846 scope.go:117] "RemoveContainer" containerID="951f36c17c01a2fa6e3011d452524eeca1d82b0bd91bd8c6634e1795396413b0" Feb 02 12:30:29 crc kubenswrapper[4846]: I0202 12:30:29.574841 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:29 crc kubenswrapper[4846]: W0202 12:30:29.576554 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4264e2ff_9f25_41df_beeb_487e9378f228.slice/crio-41479d2195fcd320d23a926af739c2b7639b72ff3ce57536542abb2e06c8e600 WatchSource:0}: Error finding container 41479d2195fcd320d23a926af739c2b7639b72ff3ce57536542abb2e06c8e600: Status 404 returned error can't find the container with id 41479d2195fcd320d23a926af739c2b7639b72ff3ce57536542abb2e06c8e600 Feb 02 12:30:29 crc kubenswrapper[4846]: I0202 12:30:29.917328 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4264e2ff-9f25-41df-beeb-487e9378f228","Type":"ContainerStarted","Data":"41479d2195fcd320d23a926af739c2b7639b72ff3ce57536542abb2e06c8e600"} Feb 02 12:30:29 crc kubenswrapper[4846]: I0202 12:30:29.920790 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ba822b20-88d5-404b-850c-7f9441933baf","Type":"ContainerStarted","Data":"b4630774f6b8174e6eaa2407dc38ee14e1e53c6776c23ce799141cd61d5f6c49"} Feb 02 12:30:30 crc kubenswrapper[4846]: I0202 12:30:30.479546 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:30:30 crc kubenswrapper[4846]: I0202 12:30:30.479869 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:30:30 crc kubenswrapper[4846]: I0202 12:30:30.933232 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4264e2ff-9f25-41df-beeb-487e9378f228","Type":"ContainerStarted","Data":"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042"} Feb 02 12:30:31 crc kubenswrapper[4846]: I0202 12:30:31.952888 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4264e2ff-9f25-41df-beeb-487e9378f228","Type":"ContainerStarted","Data":"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb"} Feb 02 12:30:31 crc kubenswrapper[4846]: I0202 12:30:31.953429 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4264e2ff-9f25-41df-beeb-487e9378f228","Type":"ContainerStarted","Data":"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18"} Feb 02 12:30:32 crc kubenswrapper[4846]: I0202 12:30:32.250254 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=4.781662263 podStartE2EDuration="13.250218722s" podCreationTimestamp="2026-02-02 12:30:19 +0000 UTC" firstStartedPulling="2026-02-02 12:30:20.69283692 +0000 UTC m=+1251.921423783" lastFinishedPulling="2026-02-02 12:30:29.161393369 +0000 UTC m=+1260.389980242" observedRunningTime="2026-02-02 12:30:29.95900582 +0000 UTC m=+1261.187592683" watchObservedRunningTime="2026-02-02 12:30:32.250218722 +0000 UTC m=+1263.478805585" Feb 02 12:30:32 crc kubenswrapper[4846]: I0202 12:30:32.258488 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:32 crc kubenswrapper[4846]: I0202 12:30:32.973987 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:30:32 crc kubenswrapper[4846]: I0202 12:30:32.974271 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" containerName="glance-log" containerID="cri-o://0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952" gracePeriod=30 Feb 02 12:30:32 crc kubenswrapper[4846]: I0202 12:30:32.974306 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" containerName="glance-httpd" containerID="cri-o://a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90" gracePeriod=30 Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.727133 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-98cnq"] Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.728492 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-98cnq" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.742585 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-98cnq"] Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.784453 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zl6d\" (UniqueName: \"kubernetes.io/projected/203788b8-50cf-402c-afb3-a546b4753707-kube-api-access-7zl6d\") pod \"nova-api-db-create-98cnq\" (UID: \"203788b8-50cf-402c-afb3-a546b4753707\") " pod="openstack/nova-api-db-create-98cnq" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.784510 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/203788b8-50cf-402c-afb3-a546b4753707-operator-scripts\") pod \"nova-api-db-create-98cnq\" (UID: \"203788b8-50cf-402c-afb3-a546b4753707\") " pod="openstack/nova-api-db-create-98cnq" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.804233 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-4xltk"] Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.805318 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4xltk" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.813304 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4xltk"] Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.885882 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c146f388-9b5f-4340-a68d-4a6f32d634cc-operator-scripts\") pod \"nova-cell0-db-create-4xltk\" (UID: \"c146f388-9b5f-4340-a68d-4a6f32d634cc\") " pod="openstack/nova-cell0-db-create-4xltk" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.885955 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zl6d\" (UniqueName: \"kubernetes.io/projected/203788b8-50cf-402c-afb3-a546b4753707-kube-api-access-7zl6d\") pod \"nova-api-db-create-98cnq\" (UID: \"203788b8-50cf-402c-afb3-a546b4753707\") " pod="openstack/nova-api-db-create-98cnq" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.886000 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/203788b8-50cf-402c-afb3-a546b4753707-operator-scripts\") pod \"nova-api-db-create-98cnq\" (UID: \"203788b8-50cf-402c-afb3-a546b4753707\") " pod="openstack/nova-api-db-create-98cnq" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.886052 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w59tn\" (UniqueName: \"kubernetes.io/projected/c146f388-9b5f-4340-a68d-4a6f32d634cc-kube-api-access-w59tn\") pod \"nova-cell0-db-create-4xltk\" (UID: \"c146f388-9b5f-4340-a68d-4a6f32d634cc\") " pod="openstack/nova-cell0-db-create-4xltk" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.889482 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/203788b8-50cf-402c-afb3-a546b4753707-operator-scripts\") pod \"nova-api-db-create-98cnq\" (UID: \"203788b8-50cf-402c-afb3-a546b4753707\") " pod="openstack/nova-api-db-create-98cnq" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.929169 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-798b-account-create-update-gjnrv"] Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.947513 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zl6d\" (UniqueName: \"kubernetes.io/projected/203788b8-50cf-402c-afb3-a546b4753707-kube-api-access-7zl6d\") pod \"nova-api-db-create-98cnq\" (UID: \"203788b8-50cf-402c-afb3-a546b4753707\") " pod="openstack/nova-api-db-create-98cnq" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.951611 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-798b-account-create-update-gjnrv" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.953640 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.990465 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w59tn\" (UniqueName: \"kubernetes.io/projected/c146f388-9b5f-4340-a68d-4a6f32d634cc-kube-api-access-w59tn\") pod \"nova-cell0-db-create-4xltk\" (UID: \"c146f388-9b5f-4340-a68d-4a6f32d634cc\") " pod="openstack/nova-cell0-db-create-4xltk" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.990558 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/037eb6da-9f07-4b72-9efe-5000aaa83b94-operator-scripts\") pod \"nova-api-798b-account-create-update-gjnrv\" (UID: \"037eb6da-9f07-4b72-9efe-5000aaa83b94\") " pod="openstack/nova-api-798b-account-create-update-gjnrv" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.990588 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6gvp\" (UniqueName: \"kubernetes.io/projected/037eb6da-9f07-4b72-9efe-5000aaa83b94-kube-api-access-n6gvp\") pod \"nova-api-798b-account-create-update-gjnrv\" (UID: \"037eb6da-9f07-4b72-9efe-5000aaa83b94\") " pod="openstack/nova-api-798b-account-create-update-gjnrv" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.991854 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c146f388-9b5f-4340-a68d-4a6f32d634cc-operator-scripts\") pod \"nova-cell0-db-create-4xltk\" (UID: \"c146f388-9b5f-4340-a68d-4a6f32d634cc\") " pod="openstack/nova-cell0-db-create-4xltk" Feb 02 12:30:33 crc kubenswrapper[4846]: I0202 12:30:33.992826 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c146f388-9b5f-4340-a68d-4a6f32d634cc-operator-scripts\") pod \"nova-cell0-db-create-4xltk\" (UID: \"c146f388-9b5f-4340-a68d-4a6f32d634cc\") " pod="openstack/nova-cell0-db-create-4xltk" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.006194 4846 generic.go:334] "Generic (PLEG): container finished" podID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" containerID="0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952" exitCode=143 Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.006386 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3be8e6d-863b-49ba-b177-5ee2bb76ba51","Type":"ContainerDied","Data":"0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952"} Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.023657 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-gkxqm"] Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.025059 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gkxqm" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.025261 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w59tn\" (UniqueName: \"kubernetes.io/projected/c146f388-9b5f-4340-a68d-4a6f32d634cc-kube-api-access-w59tn\") pod \"nova-cell0-db-create-4xltk\" (UID: \"c146f388-9b5f-4340-a68d-4a6f32d634cc\") " pod="openstack/nova-cell0-db-create-4xltk" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.042607 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-gkxqm"] Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.049632 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-98cnq" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.056513 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-798b-account-create-update-gjnrv"] Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.094111 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acd7f59a-b591-4a5c-b183-7006ae325ed8-operator-scripts\") pod \"nova-cell1-db-create-gkxqm\" (UID: \"acd7f59a-b591-4a5c-b183-7006ae325ed8\") " pod="openstack/nova-cell1-db-create-gkxqm" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.094192 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/037eb6da-9f07-4b72-9efe-5000aaa83b94-operator-scripts\") pod \"nova-api-798b-account-create-update-gjnrv\" (UID: \"037eb6da-9f07-4b72-9efe-5000aaa83b94\") " pod="openstack/nova-api-798b-account-create-update-gjnrv" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.094228 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6gvp\" (UniqueName: \"kubernetes.io/projected/037eb6da-9f07-4b72-9efe-5000aaa83b94-kube-api-access-n6gvp\") pod \"nova-api-798b-account-create-update-gjnrv\" (UID: \"037eb6da-9f07-4b72-9efe-5000aaa83b94\") " pod="openstack/nova-api-798b-account-create-update-gjnrv" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.094293 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtdcv\" (UniqueName: \"kubernetes.io/projected/acd7f59a-b591-4a5c-b183-7006ae325ed8-kube-api-access-rtdcv\") pod \"nova-cell1-db-create-gkxqm\" (UID: \"acd7f59a-b591-4a5c-b183-7006ae325ed8\") " pod="openstack/nova-cell1-db-create-gkxqm" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.095314 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/037eb6da-9f07-4b72-9efe-5000aaa83b94-operator-scripts\") pod \"nova-api-798b-account-create-update-gjnrv\" (UID: \"037eb6da-9f07-4b72-9efe-5000aaa83b94\") " pod="openstack/nova-api-798b-account-create-update-gjnrv" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.118649 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-s2hlh"] Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.120033 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.122469 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.125179 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6gvp\" (UniqueName: \"kubernetes.io/projected/037eb6da-9f07-4b72-9efe-5000aaa83b94-kube-api-access-n6gvp\") pod \"nova-api-798b-account-create-update-gjnrv\" (UID: \"037eb6da-9f07-4b72-9efe-5000aaa83b94\") " pod="openstack/nova-api-798b-account-create-update-gjnrv" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.125741 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4xltk" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.148148 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-s2hlh"] Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.202870 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acd7f59a-b591-4a5c-b183-7006ae325ed8-operator-scripts\") pod \"nova-cell1-db-create-gkxqm\" (UID: \"acd7f59a-b591-4a5c-b183-7006ae325ed8\") " pod="openstack/nova-cell1-db-create-gkxqm" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.203172 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-858ln\" (UniqueName: \"kubernetes.io/projected/77c4e938-e657-40a2-9d0d-7261afc00a38-kube-api-access-858ln\") pod \"nova-cell0-ad5a-account-create-update-s2hlh\" (UID: \"77c4e938-e657-40a2-9d0d-7261afc00a38\") " pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.203217 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtdcv\" (UniqueName: \"kubernetes.io/projected/acd7f59a-b591-4a5c-b183-7006ae325ed8-kube-api-access-rtdcv\") pod \"nova-cell1-db-create-gkxqm\" (UID: \"acd7f59a-b591-4a5c-b183-7006ae325ed8\") " pod="openstack/nova-cell1-db-create-gkxqm" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.203283 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77c4e938-e657-40a2-9d0d-7261afc00a38-operator-scripts\") pod \"nova-cell0-ad5a-account-create-update-s2hlh\" (UID: \"77c4e938-e657-40a2-9d0d-7261afc00a38\") " pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.203858 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acd7f59a-b591-4a5c-b183-7006ae325ed8-operator-scripts\") pod \"nova-cell1-db-create-gkxqm\" (UID: \"acd7f59a-b591-4a5c-b183-7006ae325ed8\") " pod="openstack/nova-cell1-db-create-gkxqm" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.233123 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtdcv\" (UniqueName: \"kubernetes.io/projected/acd7f59a-b591-4a5c-b183-7006ae325ed8-kube-api-access-rtdcv\") pod \"nova-cell1-db-create-gkxqm\" (UID: \"acd7f59a-b591-4a5c-b183-7006ae325ed8\") " pod="openstack/nova-cell1-db-create-gkxqm" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.308133 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-858ln\" (UniqueName: \"kubernetes.io/projected/77c4e938-e657-40a2-9d0d-7261afc00a38-kube-api-access-858ln\") pod \"nova-cell0-ad5a-account-create-update-s2hlh\" (UID: \"77c4e938-e657-40a2-9d0d-7261afc00a38\") " pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.308185 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77c4e938-e657-40a2-9d0d-7261afc00a38-operator-scripts\") pod \"nova-cell0-ad5a-account-create-update-s2hlh\" (UID: \"77c4e938-e657-40a2-9d0d-7261afc00a38\") " pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.309116 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77c4e938-e657-40a2-9d0d-7261afc00a38-operator-scripts\") pod \"nova-cell0-ad5a-account-create-update-s2hlh\" (UID: \"77c4e938-e657-40a2-9d0d-7261afc00a38\") " pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.325537 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f3b0-account-create-update-5xhdn"] Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.335327 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.345413 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.353427 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f3b0-account-create-update-5xhdn"] Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.354539 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-858ln\" (UniqueName: \"kubernetes.io/projected/77c4e938-e657-40a2-9d0d-7261afc00a38-kube-api-access-858ln\") pod \"nova-cell0-ad5a-account-create-update-s2hlh\" (UID: \"77c4e938-e657-40a2-9d0d-7261afc00a38\") " pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.366264 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-798b-account-create-update-gjnrv" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.410282 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcfh7\" (UniqueName: \"kubernetes.io/projected/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-kube-api-access-zcfh7\") pod \"nova-cell1-f3b0-account-create-update-5xhdn\" (UID: \"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531\") " pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.410395 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-operator-scripts\") pod \"nova-cell1-f3b0-account-create-update-5xhdn\" (UID: \"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531\") " pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.471675 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gkxqm" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.481764 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.518525 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-operator-scripts\") pod \"nova-cell1-f3b0-account-create-update-5xhdn\" (UID: \"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531\") " pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.518755 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcfh7\" (UniqueName: \"kubernetes.io/projected/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-kube-api-access-zcfh7\") pod \"nova-cell1-f3b0-account-create-update-5xhdn\" (UID: \"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531\") " pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.528190 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-operator-scripts\") pod \"nova-cell1-f3b0-account-create-update-5xhdn\" (UID: \"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531\") " pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.529183 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.529525 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" containerName="glance-log" containerID="cri-o://daf627fdb7650538b5227b20b9c3661cc4f32bff6d3c41d2a44219cc248b6742" gracePeriod=30 Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.530125 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" containerName="glance-httpd" containerID="cri-o://78297be1a506fe0c9e7ba8b6348cafd64ffab90d1e75cb6ad2177171c185713b" gracePeriod=30 Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.560889 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcfh7\" (UniqueName: \"kubernetes.io/projected/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-kube-api-access-zcfh7\") pod \"nova-cell1-f3b0-account-create-update-5xhdn\" (UID: \"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531\") " pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.620340 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-98cnq"] Feb 02 12:30:34 crc kubenswrapper[4846]: W0202 12:30:34.623391 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod203788b8_50cf_402c_afb3_a546b4753707.slice/crio-e3a7fd061f56a5c7e04a9026ad1ebffeb2c0b17cd045fa8cd7b99861c6a1248e WatchSource:0}: Error finding container e3a7fd061f56a5c7e04a9026ad1ebffeb2c0b17cd045fa8cd7b99861c6a1248e: Status 404 returned error can't find the container with id e3a7fd061f56a5c7e04a9026ad1ebffeb2c0b17cd045fa8cd7b99861c6a1248e Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.794277 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.897591 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4xltk"] Feb 02 12:30:34 crc kubenswrapper[4846]: W0202 12:30:34.907919 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc146f388_9b5f_4340_a68d_4a6f32d634cc.slice/crio-71c293a73d46a351a5de8b04f300b91e7f1c554bbadd34cd8c68eae9ef1e1cd4 WatchSource:0}: Error finding container 71c293a73d46a351a5de8b04f300b91e7f1c554bbadd34cd8c68eae9ef1e1cd4: Status 404 returned error can't find the container with id 71c293a73d46a351a5de8b04f300b91e7f1c554bbadd34cd8c68eae9ef1e1cd4 Feb 02 12:30:34 crc kubenswrapper[4846]: I0202 12:30:34.976927 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-798b-account-create-update-gjnrv"] Feb 02 12:30:34 crc kubenswrapper[4846]: W0202 12:30:34.985740 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod037eb6da_9f07_4b72_9efe_5000aaa83b94.slice/crio-5d3706561607e7f085f8c50ddef82ae888b6bb83172d05a0e281b3c4c121902f WatchSource:0}: Error finding container 5d3706561607e7f085f8c50ddef82ae888b6bb83172d05a0e281b3c4c121902f: Status 404 returned error can't find the container with id 5d3706561607e7f085f8c50ddef82ae888b6bb83172d05a0e281b3c4c121902f Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.033120 4846 generic.go:334] "Generic (PLEG): container finished" podID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" containerID="daf627fdb7650538b5227b20b9c3661cc4f32bff6d3c41d2a44219cc248b6742" exitCode=143 Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.033671 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d51c1b58-d44f-44cd-a513-af4fb272c0a7","Type":"ContainerDied","Data":"daf627fdb7650538b5227b20b9c3661cc4f32bff6d3c41d2a44219cc248b6742"} Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.040994 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4264e2ff-9f25-41df-beeb-487e9378f228","Type":"ContainerStarted","Data":"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45"} Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.041045 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="ceilometer-central-agent" containerID="cri-o://3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042" gracePeriod=30 Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.041110 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.041159 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="ceilometer-notification-agent" containerID="cri-o://1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18" gracePeriod=30 Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.041159 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="sg-core" containerID="cri-o://cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb" gracePeriod=30 Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.041189 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="proxy-httpd" containerID="cri-o://f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45" gracePeriod=30 Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.046973 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-798b-account-create-update-gjnrv" event={"ID":"037eb6da-9f07-4b72-9efe-5000aaa83b94","Type":"ContainerStarted","Data":"5d3706561607e7f085f8c50ddef82ae888b6bb83172d05a0e281b3c4c121902f"} Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.049513 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4xltk" event={"ID":"c146f388-9b5f-4340-a68d-4a6f32d634cc","Type":"ContainerStarted","Data":"71c293a73d46a351a5de8b04f300b91e7f1c554bbadd34cd8c68eae9ef1e1cd4"} Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.067309 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.708551179 podStartE2EDuration="11.067285721s" podCreationTimestamp="2026-02-02 12:30:24 +0000 UTC" firstStartedPulling="2026-02-02 12:30:29.579198982 +0000 UTC m=+1260.807785845" lastFinishedPulling="2026-02-02 12:30:33.937933524 +0000 UTC m=+1265.166520387" observedRunningTime="2026-02-02 12:30:35.060759479 +0000 UTC m=+1266.289346342" watchObservedRunningTime="2026-02-02 12:30:35.067285721 +0000 UTC m=+1266.295872584" Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.078702 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-98cnq" event={"ID":"203788b8-50cf-402c-afb3-a546b4753707","Type":"ContainerStarted","Data":"ca013b760fcb0dbd5895aba5a0850c6f7f8aa9da7493a41b415f4331df665b88"} Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.078753 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-98cnq" event={"ID":"203788b8-50cf-402c-afb3-a546b4753707","Type":"ContainerStarted","Data":"e3a7fd061f56a5c7e04a9026ad1ebffeb2c0b17cd045fa8cd7b99861c6a1248e"} Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.108477 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-98cnq" podStartSLOduration=2.108455644 podStartE2EDuration="2.108455644s" podCreationTimestamp="2026-02-02 12:30:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:30:35.099984933 +0000 UTC m=+1266.328571796" watchObservedRunningTime="2026-02-02 12:30:35.108455644 +0000 UTC m=+1266.337042507" Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.182840 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-s2hlh"] Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.209166 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.286373 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5564fdb9d4-xt6f7"] Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.286882 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5564fdb9d4-xt6f7" podUID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" containerName="neutron-api" containerID="cri-o://a7c5dc480cc9c268bfef70458e359db736dad2aa52a7110af88fdcd9bdce0c4a" gracePeriod=30 Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.287461 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5564fdb9d4-xt6f7" podUID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" containerName="neutron-httpd" containerID="cri-o://6af2e184bc53c534dc08e92be427ce1c5bdc36efdacf88e01825552ab6c352fb" gracePeriod=30 Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.334318 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-gkxqm"] Feb 02 12:30:35 crc kubenswrapper[4846]: W0202 12:30:35.357448 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacd7f59a_b591_4a5c_b183_7006ae325ed8.slice/crio-66ec2029348e6d5b91b33d411b678eaf86c690427ce253fd15fb47b9d03ca33b WatchSource:0}: Error finding container 66ec2029348e6d5b91b33d411b678eaf86c690427ce253fd15fb47b9d03ca33b: Status 404 returned error can't find the container with id 66ec2029348e6d5b91b33d411b678eaf86c690427ce253fd15fb47b9d03ca33b Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.383154 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f3b0-account-create-update-5xhdn"] Feb 02 12:30:35 crc kubenswrapper[4846]: I0202 12:30:35.894231 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.057522 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-log-httpd\") pod \"4264e2ff-9f25-41df-beeb-487e9378f228\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.057588 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-combined-ca-bundle\") pod \"4264e2ff-9f25-41df-beeb-487e9378f228\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.057639 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-config-data\") pod \"4264e2ff-9f25-41df-beeb-487e9378f228\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.057682 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-scripts\") pod \"4264e2ff-9f25-41df-beeb-487e9378f228\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.057732 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjmwr\" (UniqueName: \"kubernetes.io/projected/4264e2ff-9f25-41df-beeb-487e9378f228-kube-api-access-hjmwr\") pod \"4264e2ff-9f25-41df-beeb-487e9378f228\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.057752 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-run-httpd\") pod \"4264e2ff-9f25-41df-beeb-487e9378f228\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.057794 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-sg-core-conf-yaml\") pod \"4264e2ff-9f25-41df-beeb-487e9378f228\" (UID: \"4264e2ff-9f25-41df-beeb-487e9378f228\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.060515 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4264e2ff-9f25-41df-beeb-487e9378f228" (UID: "4264e2ff-9f25-41df-beeb-487e9378f228"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.063166 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4264e2ff-9f25-41df-beeb-487e9378f228" (UID: "4264e2ff-9f25-41df-beeb-487e9378f228"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.064869 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-scripts" (OuterVolumeSpecName: "scripts") pod "4264e2ff-9f25-41df-beeb-487e9378f228" (UID: "4264e2ff-9f25-41df-beeb-487e9378f228"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.079849 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4264e2ff-9f25-41df-beeb-487e9378f228-kube-api-access-hjmwr" (OuterVolumeSpecName: "kube-api-access-hjmwr") pod "4264e2ff-9f25-41df-beeb-487e9378f228" (UID: "4264e2ff-9f25-41df-beeb-487e9378f228"). InnerVolumeSpecName "kube-api-access-hjmwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113104 4846 generic.go:334] "Generic (PLEG): container finished" podID="4264e2ff-9f25-41df-beeb-487e9378f228" containerID="f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45" exitCode=0 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113354 4846 generic.go:334] "Generic (PLEG): container finished" podID="4264e2ff-9f25-41df-beeb-487e9378f228" containerID="cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb" exitCode=2 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113463 4846 generic.go:334] "Generic (PLEG): container finished" podID="4264e2ff-9f25-41df-beeb-487e9378f228" containerID="1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18" exitCode=0 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113531 4846 generic.go:334] "Generic (PLEG): container finished" podID="4264e2ff-9f25-41df-beeb-487e9378f228" containerID="3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042" exitCode=0 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113169 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113184 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4264e2ff-9f25-41df-beeb-487e9378f228","Type":"ContainerDied","Data":"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113669 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4264e2ff-9f25-41df-beeb-487e9378f228","Type":"ContainerDied","Data":"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113685 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4264e2ff-9f25-41df-beeb-487e9378f228","Type":"ContainerDied","Data":"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113754 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4264e2ff-9f25-41df-beeb-487e9378f228","Type":"ContainerDied","Data":"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113765 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4264e2ff-9f25-41df-beeb-487e9378f228","Type":"ContainerDied","Data":"41479d2195fcd320d23a926af739c2b7639b72ff3ce57536542abb2e06c8e600"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.113804 4846 scope.go:117] "RemoveContainer" containerID="f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.116946 4846 generic.go:334] "Generic (PLEG): container finished" podID="037eb6da-9f07-4b72-9efe-5000aaa83b94" containerID="cd9c89a15c32fa01cbe6922310a47bcd10ec6d5810bd28c0c659192a699b7d01" exitCode=0 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.116991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-798b-account-create-update-gjnrv" event={"ID":"037eb6da-9f07-4b72-9efe-5000aaa83b94","Type":"ContainerDied","Data":"cd9c89a15c32fa01cbe6922310a47bcd10ec6d5810bd28c0c659192a699b7d01"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.119581 4846 generic.go:334] "Generic (PLEG): container finished" podID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" containerID="6af2e184bc53c534dc08e92be427ce1c5bdc36efdacf88e01825552ab6c352fb" exitCode=0 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.119658 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5564fdb9d4-xt6f7" event={"ID":"447fc9d5-7789-43d6-8d11-d5ef5b00c371","Type":"ContainerDied","Data":"6af2e184bc53c534dc08e92be427ce1c5bdc36efdacf88e01825552ab6c352fb"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.121182 4846 generic.go:334] "Generic (PLEG): container finished" podID="c146f388-9b5f-4340-a68d-4a6f32d634cc" containerID="4030d8eaddee56f1991428cfeacaa9dffb7cdec1411b2eaff3e458cb869e836c" exitCode=0 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.121219 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4xltk" event={"ID":"c146f388-9b5f-4340-a68d-4a6f32d634cc","Type":"ContainerDied","Data":"4030d8eaddee56f1991428cfeacaa9dffb7cdec1411b2eaff3e458cb869e836c"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.123739 4846 generic.go:334] "Generic (PLEG): container finished" podID="acd7f59a-b591-4a5c-b183-7006ae325ed8" containerID="2e665932988445cbfa0eb4e53f1f8f931bda8deb94a40abbec8844a52aed5529" exitCode=0 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.123811 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gkxqm" event={"ID":"acd7f59a-b591-4a5c-b183-7006ae325ed8","Type":"ContainerDied","Data":"2e665932988445cbfa0eb4e53f1f8f931bda8deb94a40abbec8844a52aed5529"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.123836 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gkxqm" event={"ID":"acd7f59a-b591-4a5c-b183-7006ae325ed8","Type":"ContainerStarted","Data":"66ec2029348e6d5b91b33d411b678eaf86c690427ce253fd15fb47b9d03ca33b"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.126385 4846 generic.go:334] "Generic (PLEG): container finished" podID="bdb0f9a1-5ddc-45ba-af84-87dc5e2af531" containerID="fe143b8c77eda416e18ab2d5f5660e969ca46e9b40cf1c43b4cfcf689c72cfa1" exitCode=0 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.126427 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" event={"ID":"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531","Type":"ContainerDied","Data":"fe143b8c77eda416e18ab2d5f5660e969ca46e9b40cf1c43b4cfcf689c72cfa1"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.126444 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" event={"ID":"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531","Type":"ContainerStarted","Data":"069cfa7aa15110f31674089d99f7e9f885fe54971e2a724133363c1f46e3f7a5"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.136744 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4264e2ff-9f25-41df-beeb-487e9378f228" (UID: "4264e2ff-9f25-41df-beeb-487e9378f228"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.136986 4846 generic.go:334] "Generic (PLEG): container finished" podID="203788b8-50cf-402c-afb3-a546b4753707" containerID="ca013b760fcb0dbd5895aba5a0850c6f7f8aa9da7493a41b415f4331df665b88" exitCode=0 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.137065 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-98cnq" event={"ID":"203788b8-50cf-402c-afb3-a546b4753707","Type":"ContainerDied","Data":"ca013b760fcb0dbd5895aba5a0850c6f7f8aa9da7493a41b415f4331df665b88"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.146282 4846 generic.go:334] "Generic (PLEG): container finished" podID="77c4e938-e657-40a2-9d0d-7261afc00a38" containerID="5e87284be07cce9a9217396923aec40f910b6e6f2a348d1a35df8d98c6249d80" exitCode=0 Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.146329 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" event={"ID":"77c4e938-e657-40a2-9d0d-7261afc00a38","Type":"ContainerDied","Data":"5e87284be07cce9a9217396923aec40f910b6e6f2a348d1a35df8d98c6249d80"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.146353 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" event={"ID":"77c4e938-e657-40a2-9d0d-7261afc00a38","Type":"ContainerStarted","Data":"5893dee30c4ad745bda15c435ca8e7c0feec0b271ed5d2cdf0235a7a537078b3"} Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.157790 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4264e2ff-9f25-41df-beeb-487e9378f228" (UID: "4264e2ff-9f25-41df-beeb-487e9378f228"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.160683 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.160713 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.160723 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.160732 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjmwr\" (UniqueName: \"kubernetes.io/projected/4264e2ff-9f25-41df-beeb-487e9378f228-kube-api-access-hjmwr\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.160747 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4264e2ff-9f25-41df-beeb-487e9378f228-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.160755 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.226521 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-config-data" (OuterVolumeSpecName: "config-data") pod "4264e2ff-9f25-41df-beeb-487e9378f228" (UID: "4264e2ff-9f25-41df-beeb-487e9378f228"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.262064 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4264e2ff-9f25-41df-beeb-487e9378f228-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.377245 4846 scope.go:117] "RemoveContainer" containerID="cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.408100 4846 scope.go:117] "RemoveContainer" containerID="1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.433506 4846 scope.go:117] "RemoveContainer" containerID="3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.452842 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.470590 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.487009 4846 scope.go:117] "RemoveContainer" containerID="f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45" Feb 02 12:30:36 crc kubenswrapper[4846]: E0202 12:30:36.487554 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45\": container with ID starting with f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45 not found: ID does not exist" containerID="f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.487594 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45"} err="failed to get container status \"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45\": rpc error: code = NotFound desc = could not find container \"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45\": container with ID starting with f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.487693 4846 scope.go:117] "RemoveContainer" containerID="cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb" Feb 02 12:30:36 crc kubenswrapper[4846]: E0202 12:30:36.488016 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb\": container with ID starting with cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb not found: ID does not exist" containerID="cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.488037 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb"} err="failed to get container status \"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb\": rpc error: code = NotFound desc = could not find container \"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb\": container with ID starting with cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.488053 4846 scope.go:117] "RemoveContainer" containerID="1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18" Feb 02 12:30:36 crc kubenswrapper[4846]: E0202 12:30:36.488285 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18\": container with ID starting with 1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18 not found: ID does not exist" containerID="1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.488308 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18"} err="failed to get container status \"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18\": rpc error: code = NotFound desc = could not find container \"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18\": container with ID starting with 1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.488327 4846 scope.go:117] "RemoveContainer" containerID="3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042" Feb 02 12:30:36 crc kubenswrapper[4846]: E0202 12:30:36.489309 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042\": container with ID starting with 3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042 not found: ID does not exist" containerID="3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.489373 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042"} err="failed to get container status \"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042\": rpc error: code = NotFound desc = could not find container \"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042\": container with ID starting with 3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.489396 4846 scope.go:117] "RemoveContainer" containerID="f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.489677 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45"} err="failed to get container status \"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45\": rpc error: code = NotFound desc = could not find container \"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45\": container with ID starting with f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.489699 4846 scope.go:117] "RemoveContainer" containerID="cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.490040 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb"} err="failed to get container status \"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb\": rpc error: code = NotFound desc = could not find container \"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb\": container with ID starting with cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.490089 4846 scope.go:117] "RemoveContainer" containerID="1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.492454 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18"} err="failed to get container status \"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18\": rpc error: code = NotFound desc = could not find container \"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18\": container with ID starting with 1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.492485 4846 scope.go:117] "RemoveContainer" containerID="3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.493795 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042"} err="failed to get container status \"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042\": rpc error: code = NotFound desc = could not find container \"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042\": container with ID starting with 3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.493825 4846 scope.go:117] "RemoveContainer" containerID="f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494067 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45"} err="failed to get container status \"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45\": rpc error: code = NotFound desc = could not find container \"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45\": container with ID starting with f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494085 4846 scope.go:117] "RemoveContainer" containerID="cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494364 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494354 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb"} err="failed to get container status \"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb\": rpc error: code = NotFound desc = could not find container \"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb\": container with ID starting with cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494404 4846 scope.go:117] "RemoveContainer" containerID="1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494633 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18"} err="failed to get container status \"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18\": rpc error: code = NotFound desc = could not find container \"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18\": container with ID starting with 1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494648 4846 scope.go:117] "RemoveContainer" containerID="3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494870 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042"} err="failed to get container status \"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042\": rpc error: code = NotFound desc = could not find container \"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042\": container with ID starting with 3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494886 4846 scope.go:117] "RemoveContainer" containerID="f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45" Feb 02 12:30:36 crc kubenswrapper[4846]: E0202 12:30:36.494893 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="ceilometer-notification-agent" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494911 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="ceilometer-notification-agent" Feb 02 12:30:36 crc kubenswrapper[4846]: E0202 12:30:36.494934 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="proxy-httpd" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.494999 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="proxy-httpd" Feb 02 12:30:36 crc kubenswrapper[4846]: E0202 12:30:36.495026 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="ceilometer-central-agent" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.495034 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="ceilometer-central-agent" Feb 02 12:30:36 crc kubenswrapper[4846]: E0202 12:30:36.495063 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="sg-core" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.495070 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="sg-core" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.495066 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45"} err="failed to get container status \"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45\": rpc error: code = NotFound desc = could not find container \"f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45\": container with ID starting with f7ebc2545abb06ece657fed4068f068f5007ba78e4d73da97b83759e613ccc45 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.495085 4846 scope.go:117] "RemoveContainer" containerID="cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.495293 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="ceilometer-notification-agent" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.495311 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="sg-core" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.495327 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="proxy-httpd" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.495343 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" containerName="ceilometer-central-agent" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.496667 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb"} err="failed to get container status \"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb\": rpc error: code = NotFound desc = could not find container \"cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb\": container with ID starting with cfb6c20435be86e769de63736422713c1a632985ac3add514aa5486b4ce9bfcb not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.496689 4846 scope.go:117] "RemoveContainer" containerID="1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.498428 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.498657 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18"} err="failed to get container status \"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18\": rpc error: code = NotFound desc = could not find container \"1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18\": container with ID starting with 1a016ff9dcbf6a04f79c21d1e5e8d00eb745b529157b7d8419e097952d23ab18 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.498696 4846 scope.go:117] "RemoveContainer" containerID="3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.499384 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042"} err="failed to get container status \"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042\": rpc error: code = NotFound desc = could not find container \"3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042\": container with ID starting with 3add1a48405c62e67de7234deb9098a4e10a28c195692dc6e5d75d7299664042 not found: ID does not exist" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.501947 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.502167 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.524002 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.572953 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-scripts\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.573009 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-log-httpd\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.573057 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-config-data\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.573155 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-run-httpd\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.573179 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75xfn\" (UniqueName: \"kubernetes.io/projected/cd738edb-655a-40fc-937f-cdfbbd5d18db-kube-api-access-75xfn\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.573205 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.573233 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.668841 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.674484 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqjv9\" (UniqueName: \"kubernetes.io/projected/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-kube-api-access-pqjv9\") pod \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.674571 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-httpd-run\") pod \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.674600 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-config-data\") pod \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.674812 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-public-tls-certs\") pod \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.674847 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-scripts\") pod \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.674872 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-combined-ca-bundle\") pod \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.674901 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-logs\") pod \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.675014 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\" (UID: \"a3be8e6d-863b-49ba-b177-5ee2bb76ba51\") " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.675214 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-run-httpd\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.675252 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75xfn\" (UniqueName: \"kubernetes.io/projected/cd738edb-655a-40fc-937f-cdfbbd5d18db-kube-api-access-75xfn\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.675261 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a3be8e6d-863b-49ba-b177-5ee2bb76ba51" (UID: "a3be8e6d-863b-49ba-b177-5ee2bb76ba51"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.675288 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.675393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.675948 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-logs" (OuterVolumeSpecName: "logs") pod "a3be8e6d-863b-49ba-b177-5ee2bb76ba51" (UID: "a3be8e6d-863b-49ba-b177-5ee2bb76ba51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.676053 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-run-httpd\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.677481 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-scripts\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.677566 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-log-httpd\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.677632 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-config-data\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.677718 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.677728 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.678488 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-log-httpd\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.680016 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-kube-api-access-pqjv9" (OuterVolumeSpecName: "kube-api-access-pqjv9") pod "a3be8e6d-863b-49ba-b177-5ee2bb76ba51" (UID: "a3be8e6d-863b-49ba-b177-5ee2bb76ba51"). InnerVolumeSpecName "kube-api-access-pqjv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.682977 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "a3be8e6d-863b-49ba-b177-5ee2bb76ba51" (UID: "a3be8e6d-863b-49ba-b177-5ee2bb76ba51"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.691525 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-scripts" (OuterVolumeSpecName: "scripts") pod "a3be8e6d-863b-49ba-b177-5ee2bb76ba51" (UID: "a3be8e6d-863b-49ba-b177-5ee2bb76ba51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.693401 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-config-data\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.700086 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75xfn\" (UniqueName: \"kubernetes.io/projected/cd738edb-655a-40fc-937f-cdfbbd5d18db-kube-api-access-75xfn\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.701092 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-scripts\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.707697 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.714194 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.732877 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3be8e6d-863b-49ba-b177-5ee2bb76ba51" (UID: "a3be8e6d-863b-49ba-b177-5ee2bb76ba51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.781730 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.781879 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.781959 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.782019 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqjv9\" (UniqueName: \"kubernetes.io/projected/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-kube-api-access-pqjv9\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.786676 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-config-data" (OuterVolumeSpecName: "config-data") pod "a3be8e6d-863b-49ba-b177-5ee2bb76ba51" (UID: "a3be8e6d-863b-49ba-b177-5ee2bb76ba51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.794766 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a3be8e6d-863b-49ba-b177-5ee2bb76ba51" (UID: "a3be8e6d-863b-49ba-b177-5ee2bb76ba51"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.806517 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.823405 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.883371 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.883404 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:36 crc kubenswrapper[4846]: I0202 12:30:36.883416 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3be8e6d-863b-49ba-b177-5ee2bb76ba51-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.156964 4846 generic.go:334] "Generic (PLEG): container finished" podID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" containerID="a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90" exitCode=0 Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.157013 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3be8e6d-863b-49ba-b177-5ee2bb76ba51","Type":"ContainerDied","Data":"a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90"} Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.157040 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.157422 4846 scope.go:117] "RemoveContainer" containerID="a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.157407 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a3be8e6d-863b-49ba-b177-5ee2bb76ba51","Type":"ContainerDied","Data":"7a0837b7c2c6d34b2e66db2be02cd58c4b57ec9bf10d0675de223e0bbf85adc7"} Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.229349 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.250677 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.283007 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:30:37 crc kubenswrapper[4846]: E0202 12:30:37.284780 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" containerName="glance-log" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.284802 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" containerName="glance-log" Feb 02 12:30:37 crc kubenswrapper[4846]: E0202 12:30:37.284839 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" containerName="glance-httpd" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.284847 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" containerName="glance-httpd" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.285266 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" containerName="glance-log" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.285292 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" containerName="glance-httpd" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.286763 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.291157 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.294587 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.300214 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.405806 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-scripts\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.405901 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-config-data\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.405925 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jrtf\" (UniqueName: \"kubernetes.io/projected/271eeb3d-6d83-49c6-9d80-75b2276a304a-kube-api-access-7jrtf\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.405946 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.405976 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-logs\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.405991 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.406013 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.406041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.438742 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4264e2ff-9f25-41df-beeb-487e9378f228" path="/var/lib/kubelet/pods/4264e2ff-9f25-41df-beeb-487e9378f228/volumes" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.439589 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3be8e6d-863b-49ba-b177-5ee2bb76ba51" path="/var/lib/kubelet/pods/a3be8e6d-863b-49ba-b177-5ee2bb76ba51/volumes" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.507718 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.507877 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-scripts\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.508063 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-config-data\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.508131 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jrtf\" (UniqueName: \"kubernetes.io/projected/271eeb3d-6d83-49c6-9d80-75b2276a304a-kube-api-access-7jrtf\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.508194 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.508285 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-logs\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.508328 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.508397 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.509365 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.509867 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.513535 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.515719 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-scripts\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.517428 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-config-data\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.522143 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-logs\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.524057 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.525720 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jrtf\" (UniqueName: \"kubernetes.io/projected/271eeb3d-6d83-49c6-9d80-75b2276a304a-kube-api-access-7jrtf\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.537825 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.650082 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.892838 4846 scope.go:117] "RemoveContainer" containerID="0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.896006 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.935613 4846 scope.go:117] "RemoveContainer" containerID="a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90" Feb 02 12:30:37 crc kubenswrapper[4846]: E0202 12:30:37.939992 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90\": container with ID starting with a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90 not found: ID does not exist" containerID="a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.940064 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90"} err="failed to get container status \"a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90\": rpc error: code = NotFound desc = could not find container \"a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90\": container with ID starting with a09e6efeae85c9abc9793144c4d69bf97a7011c55f51a3d4fef67833be6feb90 not found: ID does not exist" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.940102 4846 scope.go:117] "RemoveContainer" containerID="0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952" Feb 02 12:30:37 crc kubenswrapper[4846]: E0202 12:30:37.940530 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952\": container with ID starting with 0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952 not found: ID does not exist" containerID="0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952" Feb 02 12:30:37 crc kubenswrapper[4846]: I0202 12:30:37.940573 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952"} err="failed to get container status \"0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952\": rpc error: code = NotFound desc = could not find container \"0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952\": container with ID starting with 0f1719d595a168f47759737ac43962e06103d9432329bedd5536d0e4c7f44952 not found: ID does not exist" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.091319 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4xltk" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.171993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4xltk" event={"ID":"c146f388-9b5f-4340-a68d-4a6f32d634cc","Type":"ContainerDied","Data":"71c293a73d46a351a5de8b04f300b91e7f1c554bbadd34cd8c68eae9ef1e1cd4"} Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.172029 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71c293a73d46a351a5de8b04f300b91e7f1c554bbadd34cd8c68eae9ef1e1cd4" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.172080 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4xltk" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.175514 4846 generic.go:334] "Generic (PLEG): container finished" podID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" containerID="78297be1a506fe0c9e7ba8b6348cafd64ffab90d1e75cb6ad2177171c185713b" exitCode=0 Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.175576 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d51c1b58-d44f-44cd-a513-af4fb272c0a7","Type":"ContainerDied","Data":"78297be1a506fe0c9e7ba8b6348cafd64ffab90d1e75cb6ad2177171c185713b"} Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.176763 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd738edb-655a-40fc-937f-cdfbbd5d18db","Type":"ContainerStarted","Data":"a7e61f6bc2901eace0444585c513b94fb5b384f895be35d589561d145e317a41"} Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.223485 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c146f388-9b5f-4340-a68d-4a6f32d634cc-operator-scripts\") pod \"c146f388-9b5f-4340-a68d-4a6f32d634cc\" (UID: \"c146f388-9b5f-4340-a68d-4a6f32d634cc\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.223641 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w59tn\" (UniqueName: \"kubernetes.io/projected/c146f388-9b5f-4340-a68d-4a6f32d634cc-kube-api-access-w59tn\") pod \"c146f388-9b5f-4340-a68d-4a6f32d634cc\" (UID: \"c146f388-9b5f-4340-a68d-4a6f32d634cc\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.224843 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c146f388-9b5f-4340-a68d-4a6f32d634cc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c146f388-9b5f-4340-a68d-4a6f32d634cc" (UID: "c146f388-9b5f-4340-a68d-4a6f32d634cc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.229947 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c146f388-9b5f-4340-a68d-4a6f32d634cc-kube-api-access-w59tn" (OuterVolumeSpecName: "kube-api-access-w59tn") pod "c146f388-9b5f-4340-a68d-4a6f32d634cc" (UID: "c146f388-9b5f-4340-a68d-4a6f32d634cc"). InnerVolumeSpecName "kube-api-access-w59tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.325343 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c146f388-9b5f-4340-a68d-4a6f32d634cc-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.325642 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w59tn\" (UniqueName: \"kubernetes.io/projected/c146f388-9b5f-4340-a68d-4a6f32d634cc-kube-api-access-w59tn\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.423746 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.436335 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-798b-account-create-update-gjnrv" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.447017 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.454422 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.461967 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gkxqm" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.484481 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-98cnq" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.528581 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-858ln\" (UniqueName: \"kubernetes.io/projected/77c4e938-e657-40a2-9d0d-7261afc00a38-kube-api-access-858ln\") pod \"77c4e938-e657-40a2-9d0d-7261afc00a38\" (UID: \"77c4e938-e657-40a2-9d0d-7261afc00a38\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.528682 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77c4e938-e657-40a2-9d0d-7261afc00a38-operator-scripts\") pod \"77c4e938-e657-40a2-9d0d-7261afc00a38\" (UID: \"77c4e938-e657-40a2-9d0d-7261afc00a38\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.529455 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77c4e938-e657-40a2-9d0d-7261afc00a38-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "77c4e938-e657-40a2-9d0d-7261afc00a38" (UID: "77c4e938-e657-40a2-9d0d-7261afc00a38"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.537797 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77c4e938-e657-40a2-9d0d-7261afc00a38-kube-api-access-858ln" (OuterVolumeSpecName: "kube-api-access-858ln") pod "77c4e938-e657-40a2-9d0d-7261afc00a38" (UID: "77c4e938-e657-40a2-9d0d-7261afc00a38"). InnerVolumeSpecName "kube-api-access-858ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.630187 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/203788b8-50cf-402c-afb3-a546b4753707-operator-scripts\") pod \"203788b8-50cf-402c-afb3-a546b4753707\" (UID: \"203788b8-50cf-402c-afb3-a546b4753707\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.630247 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acd7f59a-b591-4a5c-b183-7006ae325ed8-operator-scripts\") pod \"acd7f59a-b591-4a5c-b183-7006ae325ed8\" (UID: \"acd7f59a-b591-4a5c-b183-7006ae325ed8\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.630275 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-operator-scripts\") pod \"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531\" (UID: \"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.630369 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6gvp\" (UniqueName: \"kubernetes.io/projected/037eb6da-9f07-4b72-9efe-5000aaa83b94-kube-api-access-n6gvp\") pod \"037eb6da-9f07-4b72-9efe-5000aaa83b94\" (UID: \"037eb6da-9f07-4b72-9efe-5000aaa83b94\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.630397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zl6d\" (UniqueName: \"kubernetes.io/projected/203788b8-50cf-402c-afb3-a546b4753707-kube-api-access-7zl6d\") pod \"203788b8-50cf-402c-afb3-a546b4753707\" (UID: \"203788b8-50cf-402c-afb3-a546b4753707\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.630470 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtdcv\" (UniqueName: \"kubernetes.io/projected/acd7f59a-b591-4a5c-b183-7006ae325ed8-kube-api-access-rtdcv\") pod \"acd7f59a-b591-4a5c-b183-7006ae325ed8\" (UID: \"acd7f59a-b591-4a5c-b183-7006ae325ed8\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.630643 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcfh7\" (UniqueName: \"kubernetes.io/projected/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-kube-api-access-zcfh7\") pod \"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531\" (UID: \"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.630683 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/037eb6da-9f07-4b72-9efe-5000aaa83b94-operator-scripts\") pod \"037eb6da-9f07-4b72-9efe-5000aaa83b94\" (UID: \"037eb6da-9f07-4b72-9efe-5000aaa83b94\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.631186 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-858ln\" (UniqueName: \"kubernetes.io/projected/77c4e938-e657-40a2-9d0d-7261afc00a38-kube-api-access-858ln\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.631211 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77c4e938-e657-40a2-9d0d-7261afc00a38-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.631615 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/037eb6da-9f07-4b72-9efe-5000aaa83b94-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "037eb6da-9f07-4b72-9efe-5000aaa83b94" (UID: "037eb6da-9f07-4b72-9efe-5000aaa83b94"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.632039 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/203788b8-50cf-402c-afb3-a546b4753707-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "203788b8-50cf-402c-afb3-a546b4753707" (UID: "203788b8-50cf-402c-afb3-a546b4753707"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.632416 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd7f59a-b591-4a5c-b183-7006ae325ed8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "acd7f59a-b591-4a5c-b183-7006ae325ed8" (UID: "acd7f59a-b591-4a5c-b183-7006ae325ed8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.632810 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bdb0f9a1-5ddc-45ba-af84-87dc5e2af531" (UID: "bdb0f9a1-5ddc-45ba-af84-87dc5e2af531"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.641808 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/037eb6da-9f07-4b72-9efe-5000aaa83b94-kube-api-access-n6gvp" (OuterVolumeSpecName: "kube-api-access-n6gvp") pod "037eb6da-9f07-4b72-9efe-5000aaa83b94" (UID: "037eb6da-9f07-4b72-9efe-5000aaa83b94"). InnerVolumeSpecName "kube-api-access-n6gvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.648871 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-kube-api-access-zcfh7" (OuterVolumeSpecName: "kube-api-access-zcfh7") pod "bdb0f9a1-5ddc-45ba-af84-87dc5e2af531" (UID: "bdb0f9a1-5ddc-45ba-af84-87dc5e2af531"). InnerVolumeSpecName "kube-api-access-zcfh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.653785 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd7f59a-b591-4a5c-b183-7006ae325ed8-kube-api-access-rtdcv" (OuterVolumeSpecName: "kube-api-access-rtdcv") pod "acd7f59a-b591-4a5c-b183-7006ae325ed8" (UID: "acd7f59a-b591-4a5c-b183-7006ae325ed8"). InnerVolumeSpecName "kube-api-access-rtdcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.663873 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/203788b8-50cf-402c-afb3-a546b4753707-kube-api-access-7zl6d" (OuterVolumeSpecName: "kube-api-access-7zl6d") pod "203788b8-50cf-402c-afb3-a546b4753707" (UID: "203788b8-50cf-402c-afb3-a546b4753707"). InnerVolumeSpecName "kube-api-access-7zl6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.732777 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcfh7\" (UniqueName: \"kubernetes.io/projected/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-kube-api-access-zcfh7\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.732805 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/037eb6da-9f07-4b72-9efe-5000aaa83b94-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.732814 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/203788b8-50cf-402c-afb3-a546b4753707-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.732821 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acd7f59a-b591-4a5c-b183-7006ae325ed8-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.732830 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.732839 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6gvp\" (UniqueName: \"kubernetes.io/projected/037eb6da-9f07-4b72-9efe-5000aaa83b94-kube-api-access-n6gvp\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.732847 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zl6d\" (UniqueName: \"kubernetes.io/projected/203788b8-50cf-402c-afb3-a546b4753707-kube-api-access-7zl6d\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.732855 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtdcv\" (UniqueName: \"kubernetes.io/projected/acd7f59a-b591-4a5c-b183-7006ae325ed8-kube-api-access-rtdcv\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.801031 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.937174 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.937569 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-httpd-run\") pod \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.937594 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4zxn\" (UniqueName: \"kubernetes.io/projected/d51c1b58-d44f-44cd-a513-af4fb272c0a7-kube-api-access-n4zxn\") pod \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.937664 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-scripts\") pod \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.937691 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-combined-ca-bundle\") pod \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.937771 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-config-data\") pod \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.937808 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-internal-tls-certs\") pod \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.937864 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-logs\") pod \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\" (UID: \"d51c1b58-d44f-44cd-a513-af4fb272c0a7\") " Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.938682 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-logs" (OuterVolumeSpecName: "logs") pod "d51c1b58-d44f-44cd-a513-af4fb272c0a7" (UID: "d51c1b58-d44f-44cd-a513-af4fb272c0a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.939291 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d51c1b58-d44f-44cd-a513-af4fb272c0a7" (UID: "d51c1b58-d44f-44cd-a513-af4fb272c0a7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.944088 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d51c1b58-d44f-44cd-a513-af4fb272c0a7-kube-api-access-n4zxn" (OuterVolumeSpecName: "kube-api-access-n4zxn") pod "d51c1b58-d44f-44cd-a513-af4fb272c0a7" (UID: "d51c1b58-d44f-44cd-a513-af4fb272c0a7"). InnerVolumeSpecName "kube-api-access-n4zxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.944164 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-scripts" (OuterVolumeSpecName: "scripts") pod "d51c1b58-d44f-44cd-a513-af4fb272c0a7" (UID: "d51c1b58-d44f-44cd-a513-af4fb272c0a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.946781 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "d51c1b58-d44f-44cd-a513-af4fb272c0a7" (UID: "d51c1b58-d44f-44cd-a513-af4fb272c0a7"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:30:38 crc kubenswrapper[4846]: I0202 12:30:38.975108 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d51c1b58-d44f-44cd-a513-af4fb272c0a7" (UID: "d51c1b58-d44f-44cd-a513-af4fb272c0a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.009702 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-config-data" (OuterVolumeSpecName: "config-data") pod "d51c1b58-d44f-44cd-a513-af4fb272c0a7" (UID: "d51c1b58-d44f-44cd-a513-af4fb272c0a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.011760 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.013979 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d51c1b58-d44f-44cd-a513-af4fb272c0a7" (UID: "d51c1b58-d44f-44cd-a513-af4fb272c0a7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.039744 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.039775 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.039787 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4zxn\" (UniqueName: \"kubernetes.io/projected/d51c1b58-d44f-44cd-a513-af4fb272c0a7-kube-api-access-n4zxn\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.039798 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.039808 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.039817 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.039824 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d51c1b58-d44f-44cd-a513-af4fb272c0a7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.039832 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d51c1b58-d44f-44cd-a513-af4fb272c0a7-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.060945 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.142098 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.186219 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" event={"ID":"77c4e938-e657-40a2-9d0d-7261afc00a38","Type":"ContainerDied","Data":"5893dee30c4ad745bda15c435ca8e7c0feec0b271ed5d2cdf0235a7a537078b3"} Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.186258 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5893dee30c4ad745bda15c435ca8e7c0feec0b271ed5d2cdf0235a7a537078b3" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.186314 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad5a-account-create-update-s2hlh" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.200782 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d51c1b58-d44f-44cd-a513-af4fb272c0a7","Type":"ContainerDied","Data":"d087ebbc17840f60e2cfa79c52c93f604e603a562b25bcde4928405b0199c93a"} Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.200821 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.200859 4846 scope.go:117] "RemoveContainer" containerID="78297be1a506fe0c9e7ba8b6348cafd64ffab90d1e75cb6ad2177171c185713b" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.203892 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd738edb-655a-40fc-937f-cdfbbd5d18db","Type":"ContainerStarted","Data":"bfda7d36305988bf593cee08d1f38c56c7cdb6fb74b7377ce314cd04037dadec"} Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.211415 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-798b-account-create-update-gjnrv" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.211566 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-798b-account-create-update-gjnrv" event={"ID":"037eb6da-9f07-4b72-9efe-5000aaa83b94","Type":"ContainerDied","Data":"5d3706561607e7f085f8c50ddef82ae888b6bb83172d05a0e281b3c4c121902f"} Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.211656 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d3706561607e7f085f8c50ddef82ae888b6bb83172d05a0e281b3c4c121902f" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.214492 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"271eeb3d-6d83-49c6-9d80-75b2276a304a","Type":"ContainerStarted","Data":"d8cfb3823d5db38975c107d55a8870660a3a63bc0167de20ac845189945afadc"} Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.217772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gkxqm" event={"ID":"acd7f59a-b591-4a5c-b183-7006ae325ed8","Type":"ContainerDied","Data":"66ec2029348e6d5b91b33d411b678eaf86c690427ce253fd15fb47b9d03ca33b"} Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.217799 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66ec2029348e6d5b91b33d411b678eaf86c690427ce253fd15fb47b9d03ca33b" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.217808 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gkxqm" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.219652 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-98cnq" event={"ID":"203788b8-50cf-402c-afb3-a546b4753707","Type":"ContainerDied","Data":"e3a7fd061f56a5c7e04a9026ad1ebffeb2c0b17cd045fa8cd7b99861c6a1248e"} Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.219677 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3a7fd061f56a5c7e04a9026ad1ebffeb2c0b17cd045fa8cd7b99861c6a1248e" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.219716 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-98cnq" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.230942 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" event={"ID":"bdb0f9a1-5ddc-45ba-af84-87dc5e2af531","Type":"ContainerDied","Data":"069cfa7aa15110f31674089d99f7e9f885fe54971e2a724133363c1f46e3f7a5"} Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.230977 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="069cfa7aa15110f31674089d99f7e9f885fe54971e2a724133363c1f46e3f7a5" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.231063 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f3b0-account-create-update-5xhdn" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.295078 4846 scope.go:117] "RemoveContainer" containerID="daf627fdb7650538b5227b20b9c3661cc4f32bff6d3c41d2a44219cc248b6742" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.310099 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.349255 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.406466 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:30:39 crc kubenswrapper[4846]: E0202 12:30:39.406873 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="203788b8-50cf-402c-afb3-a546b4753707" containerName="mariadb-database-create" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.406885 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="203788b8-50cf-402c-afb3-a546b4753707" containerName="mariadb-database-create" Feb 02 12:30:39 crc kubenswrapper[4846]: E0202 12:30:39.406904 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77c4e938-e657-40a2-9d0d-7261afc00a38" containerName="mariadb-account-create-update" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.406910 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="77c4e938-e657-40a2-9d0d-7261afc00a38" containerName="mariadb-account-create-update" Feb 02 12:30:39 crc kubenswrapper[4846]: E0202 12:30:39.406922 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" containerName="glance-log" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.406928 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" containerName="glance-log" Feb 02 12:30:39 crc kubenswrapper[4846]: E0202 12:30:39.406941 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdb0f9a1-5ddc-45ba-af84-87dc5e2af531" containerName="mariadb-account-create-update" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.406947 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdb0f9a1-5ddc-45ba-af84-87dc5e2af531" containerName="mariadb-account-create-update" Feb 02 12:30:39 crc kubenswrapper[4846]: E0202 12:30:39.406960 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd7f59a-b591-4a5c-b183-7006ae325ed8" containerName="mariadb-database-create" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.406966 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd7f59a-b591-4a5c-b183-7006ae325ed8" containerName="mariadb-database-create" Feb 02 12:30:39 crc kubenswrapper[4846]: E0202 12:30:39.406986 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c146f388-9b5f-4340-a68d-4a6f32d634cc" containerName="mariadb-database-create" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.406994 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c146f388-9b5f-4340-a68d-4a6f32d634cc" containerName="mariadb-database-create" Feb 02 12:30:39 crc kubenswrapper[4846]: E0202 12:30:39.407010 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="037eb6da-9f07-4b72-9efe-5000aaa83b94" containerName="mariadb-account-create-update" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.407024 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="037eb6da-9f07-4b72-9efe-5000aaa83b94" containerName="mariadb-account-create-update" Feb 02 12:30:39 crc kubenswrapper[4846]: E0202 12:30:39.407047 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" containerName="glance-httpd" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.407053 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" containerName="glance-httpd" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.407294 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="203788b8-50cf-402c-afb3-a546b4753707" containerName="mariadb-database-create" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.407315 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="acd7f59a-b591-4a5c-b183-7006ae325ed8" containerName="mariadb-database-create" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.407323 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="77c4e938-e657-40a2-9d0d-7261afc00a38" containerName="mariadb-account-create-update" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.407334 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" containerName="glance-log" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.407346 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdb0f9a1-5ddc-45ba-af84-87dc5e2af531" containerName="mariadb-account-create-update" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.407356 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" containerName="glance-httpd" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.407371 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="037eb6da-9f07-4b72-9efe-5000aaa83b94" containerName="mariadb-account-create-update" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.407380 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c146f388-9b5f-4340-a68d-4a6f32d634cc" containerName="mariadb-database-create" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.408422 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.414898 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.415256 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.419975 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.511991 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d51c1b58-d44f-44cd-a513-af4fb272c0a7" path="/var/lib/kubelet/pods/d51c1b58-d44f-44cd-a513-af4fb272c0a7/volumes" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.555492 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.555552 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.555668 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.555718 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.555764 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-logs\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.555806 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.555842 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhmqx\" (UniqueName: \"kubernetes.io/projected/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-kube-api-access-dhmqx\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.555899 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.657346 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhmqx\" (UniqueName: \"kubernetes.io/projected/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-kube-api-access-dhmqx\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.657417 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.657461 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.657505 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.657558 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.657596 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.657660 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-logs\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.657698 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.658263 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.662029 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.662217 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.662746 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-logs\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.667417 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.668526 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.669467 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.678492 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhmqx\" (UniqueName: \"kubernetes.io/projected/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-kube-api-access-dhmqx\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.688211 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " pod="openstack/glance-default-internal-api-0" Feb 02 12:30:39 crc kubenswrapper[4846]: I0202 12:30:39.741970 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.310296 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"271eeb3d-6d83-49c6-9d80-75b2276a304a","Type":"ContainerStarted","Data":"a3bf2b4a419820c2d36fe369e8dac05f3052d401b7c0bdcc51a1cdf750ef3a24"} Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.313116 4846 generic.go:334] "Generic (PLEG): container finished" podID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" containerID="a7c5dc480cc9c268bfef70458e359db736dad2aa52a7110af88fdcd9bdce0c4a" exitCode=0 Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.313179 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5564fdb9d4-xt6f7" event={"ID":"447fc9d5-7789-43d6-8d11-d5ef5b00c371","Type":"ContainerDied","Data":"a7c5dc480cc9c268bfef70458e359db736dad2aa52a7110af88fdcd9bdce0c4a"} Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.313207 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5564fdb9d4-xt6f7" event={"ID":"447fc9d5-7789-43d6-8d11-d5ef5b00c371","Type":"ContainerDied","Data":"134b847080e6eb8962ce3f55225f92a0323d0bac5e6436cc287b32e457441961"} Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.313224 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="134b847080e6eb8962ce3f55225f92a0323d0bac5e6436cc287b32e457441961" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.332864 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd738edb-655a-40fc-937f-cdfbbd5d18db","Type":"ContainerStarted","Data":"04102fe16cf5633a4483e6d0f0a72eb749fefdc8ecc151fd5d15355d963aa402"} Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.399318 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.427695 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.478518 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-config\") pod \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.478579 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-httpd-config\") pod \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.478689 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-ovndb-tls-certs\") pod \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.478857 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-combined-ca-bundle\") pod \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.478884 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvm6k\" (UniqueName: \"kubernetes.io/projected/447fc9d5-7789-43d6-8d11-d5ef5b00c371-kube-api-access-tvm6k\") pod \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\" (UID: \"447fc9d5-7789-43d6-8d11-d5ef5b00c371\") " Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.488477 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "447fc9d5-7789-43d6-8d11-d5ef5b00c371" (UID: "447fc9d5-7789-43d6-8d11-d5ef5b00c371"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.490028 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/447fc9d5-7789-43d6-8d11-d5ef5b00c371-kube-api-access-tvm6k" (OuterVolumeSpecName: "kube-api-access-tvm6k") pod "447fc9d5-7789-43d6-8d11-d5ef5b00c371" (UID: "447fc9d5-7789-43d6-8d11-d5ef5b00c371"). InnerVolumeSpecName "kube-api-access-tvm6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.534120 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-config" (OuterVolumeSpecName: "config") pod "447fc9d5-7789-43d6-8d11-d5ef5b00c371" (UID: "447fc9d5-7789-43d6-8d11-d5ef5b00c371"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.539098 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "447fc9d5-7789-43d6-8d11-d5ef5b00c371" (UID: "447fc9d5-7789-43d6-8d11-d5ef5b00c371"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.559784 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "447fc9d5-7789-43d6-8d11-d5ef5b00c371" (UID: "447fc9d5-7789-43d6-8d11-d5ef5b00c371"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.586750 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.587434 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.587463 4846 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.587477 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/447fc9d5-7789-43d6-8d11-d5ef5b00c371-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:40 crc kubenswrapper[4846]: I0202 12:30:40.587487 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvm6k\" (UniqueName: \"kubernetes.io/projected/447fc9d5-7789-43d6-8d11-d5ef5b00c371-kube-api-access-tvm6k\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:41 crc kubenswrapper[4846]: I0202 12:30:41.369413 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"271eeb3d-6d83-49c6-9d80-75b2276a304a","Type":"ContainerStarted","Data":"dbd55a0c1904df42f4b3eb5b57e2aa4fadf58f7e576128bd91743f3ff542702f"} Feb 02 12:30:41 crc kubenswrapper[4846]: I0202 12:30:41.374146 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9","Type":"ContainerStarted","Data":"7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686"} Feb 02 12:30:41 crc kubenswrapper[4846]: I0202 12:30:41.374182 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9","Type":"ContainerStarted","Data":"4bbaef0fedc86edabb38183c4145e745121eaa5ceefb10d69f194519fd2e79eb"} Feb 02 12:30:41 crc kubenswrapper[4846]: I0202 12:30:41.389512 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5564fdb9d4-xt6f7" Feb 02 12:30:41 crc kubenswrapper[4846]: I0202 12:30:41.389707 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd738edb-655a-40fc-937f-cdfbbd5d18db","Type":"ContainerStarted","Data":"dfbf79fdd516504fe08b30c2a8d4a02c0a822265755aca5ca2510788866ae88b"} Feb 02 12:30:41 crc kubenswrapper[4846]: I0202 12:30:41.407452 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.407433525 podStartE2EDuration="4.407433525s" podCreationTimestamp="2026-02-02 12:30:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:30:41.394159835 +0000 UTC m=+1272.622746708" watchObservedRunningTime="2026-02-02 12:30:41.407433525 +0000 UTC m=+1272.636020388" Feb 02 12:30:41 crc kubenswrapper[4846]: I0202 12:30:41.460355 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5564fdb9d4-xt6f7"] Feb 02 12:30:41 crc kubenswrapper[4846]: I0202 12:30:41.470268 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5564fdb9d4-xt6f7"] Feb 02 12:30:42 crc kubenswrapper[4846]: I0202 12:30:42.402269 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9","Type":"ContainerStarted","Data":"265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f"} Feb 02 12:30:42 crc kubenswrapper[4846]: I0202 12:30:42.430516 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.43050057 podStartE2EDuration="3.43050057s" podCreationTimestamp="2026-02-02 12:30:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:30:42.427414593 +0000 UTC m=+1273.656001456" watchObservedRunningTime="2026-02-02 12:30:42.43050057 +0000 UTC m=+1273.659087433" Feb 02 12:30:42 crc kubenswrapper[4846]: I0202 12:30:42.806004 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:42 crc kubenswrapper[4846]: I0202 12:30:42.817112 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:30:42 crc kubenswrapper[4846]: I0202 12:30:42.899954 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7b686584c4-zpsks"] Feb 02 12:30:42 crc kubenswrapper[4846]: I0202 12:30:42.900260 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7b686584c4-zpsks" podUID="a8dce541-0c00-476e-8efe-fea46f5fd26b" containerName="placement-log" containerID="cri-o://07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376" gracePeriod=30 Feb 02 12:30:42 crc kubenswrapper[4846]: I0202 12:30:42.900441 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7b686584c4-zpsks" podUID="a8dce541-0c00-476e-8efe-fea46f5fd26b" containerName="placement-api" containerID="cri-o://493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb" gracePeriod=30 Feb 02 12:30:43 crc kubenswrapper[4846]: I0202 12:30:43.413782 4846 generic.go:334] "Generic (PLEG): container finished" podID="a8dce541-0c00-476e-8efe-fea46f5fd26b" containerID="07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376" exitCode=143 Feb 02 12:30:43 crc kubenswrapper[4846]: I0202 12:30:43.413857 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b686584c4-zpsks" event={"ID":"a8dce541-0c00-476e-8efe-fea46f5fd26b","Type":"ContainerDied","Data":"07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376"} Feb 02 12:30:43 crc kubenswrapper[4846]: I0202 12:30:43.436301 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" path="/var/lib/kubelet/pods/447fc9d5-7789-43d6-8d11-d5ef5b00c371/volumes" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.387894 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g5kjm"] Feb 02 12:30:44 crc kubenswrapper[4846]: E0202 12:30:44.388890 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" containerName="neutron-api" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.388914 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" containerName="neutron-api" Feb 02 12:30:44 crc kubenswrapper[4846]: E0202 12:30:44.389948 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" containerName="neutron-httpd" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.390006 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" containerName="neutron-httpd" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.390788 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" containerName="neutron-api" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.390848 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="447fc9d5-7789-43d6-8d11-d5ef5b00c371" containerName="neutron-httpd" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.391976 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.396393 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pkshr" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.396840 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.397116 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.402223 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g5kjm"] Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.432790 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd738edb-655a-40fc-937f-cdfbbd5d18db","Type":"ContainerStarted","Data":"4d57a7e0b87f5992937c999de2d1144c544067617683b8753f5908886195c623"} Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.432936 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.432971 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="ceilometer-central-agent" containerID="cri-o://bfda7d36305988bf593cee08d1f38c56c7cdb6fb74b7377ce314cd04037dadec" gracePeriod=30 Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.433002 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="proxy-httpd" containerID="cri-o://4d57a7e0b87f5992937c999de2d1144c544067617683b8753f5908886195c623" gracePeriod=30 Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.433090 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="sg-core" containerID="cri-o://dfbf79fdd516504fe08b30c2a8d4a02c0a822265755aca5ca2510788866ae88b" gracePeriod=30 Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.433136 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="ceilometer-notification-agent" containerID="cri-o://04102fe16cf5633a4483e6d0f0a72eb749fefdc8ecc151fd5d15355d963aa402" gracePeriod=30 Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.461279 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.609155342 podStartE2EDuration="8.461258698s" podCreationTimestamp="2026-02-02 12:30:36 +0000 UTC" firstStartedPulling="2026-02-02 12:30:37.901996958 +0000 UTC m=+1269.130583821" lastFinishedPulling="2026-02-02 12:30:43.754100314 +0000 UTC m=+1274.982687177" observedRunningTime="2026-02-02 12:30:44.452685825 +0000 UTC m=+1275.681272708" watchObservedRunningTime="2026-02-02 12:30:44.461258698 +0000 UTC m=+1275.689845561" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.464788 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-scripts\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.464928 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.465068 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-config-data\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.465108 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpgcn\" (UniqueName: \"kubernetes.io/projected/db0a7be1-2077-4c1b-967c-4b767e5d6a51-kube-api-access-tpgcn\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.566290 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpgcn\" (UniqueName: \"kubernetes.io/projected/db0a7be1-2077-4c1b-967c-4b767e5d6a51-kube-api-access-tpgcn\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.566333 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-scripts\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.566404 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.566499 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-config-data\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.571574 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.572257 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-config-data\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.572735 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-scripts\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.583770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpgcn\" (UniqueName: \"kubernetes.io/projected/db0a7be1-2077-4c1b-967c-4b767e5d6a51-kube-api-access-tpgcn\") pod \"nova-cell0-conductor-db-sync-g5kjm\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:44 crc kubenswrapper[4846]: I0202 12:30:44.711112 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:30:45 crc kubenswrapper[4846]: I0202 12:30:45.175445 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g5kjm"] Feb 02 12:30:45 crc kubenswrapper[4846]: W0202 12:30:45.175925 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb0a7be1_2077_4c1b_967c_4b767e5d6a51.slice/crio-90e57db732a86b43630ae65b8adc9a3761e0cb5e2e2a026189d239b54c9e8818 WatchSource:0}: Error finding container 90e57db732a86b43630ae65b8adc9a3761e0cb5e2e2a026189d239b54c9e8818: Status 404 returned error can't find the container with id 90e57db732a86b43630ae65b8adc9a3761e0cb5e2e2a026189d239b54c9e8818 Feb 02 12:30:45 crc kubenswrapper[4846]: I0202 12:30:45.451478 4846 generic.go:334] "Generic (PLEG): container finished" podID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerID="4d57a7e0b87f5992937c999de2d1144c544067617683b8753f5908886195c623" exitCode=0 Feb 02 12:30:45 crc kubenswrapper[4846]: I0202 12:30:45.451514 4846 generic.go:334] "Generic (PLEG): container finished" podID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerID="dfbf79fdd516504fe08b30c2a8d4a02c0a822265755aca5ca2510788866ae88b" exitCode=2 Feb 02 12:30:45 crc kubenswrapper[4846]: I0202 12:30:45.451523 4846 generic.go:334] "Generic (PLEG): container finished" podID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerID="04102fe16cf5633a4483e6d0f0a72eb749fefdc8ecc151fd5d15355d963aa402" exitCode=0 Feb 02 12:30:45 crc kubenswrapper[4846]: I0202 12:30:45.451579 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd738edb-655a-40fc-937f-cdfbbd5d18db","Type":"ContainerDied","Data":"4d57a7e0b87f5992937c999de2d1144c544067617683b8753f5908886195c623"} Feb 02 12:30:45 crc kubenswrapper[4846]: I0202 12:30:45.451637 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd738edb-655a-40fc-937f-cdfbbd5d18db","Type":"ContainerDied","Data":"dfbf79fdd516504fe08b30c2a8d4a02c0a822265755aca5ca2510788866ae88b"} Feb 02 12:30:45 crc kubenswrapper[4846]: I0202 12:30:45.451653 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd738edb-655a-40fc-937f-cdfbbd5d18db","Type":"ContainerDied","Data":"04102fe16cf5633a4483e6d0f0a72eb749fefdc8ecc151fd5d15355d963aa402"} Feb 02 12:30:45 crc kubenswrapper[4846]: I0202 12:30:45.452662 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g5kjm" event={"ID":"db0a7be1-2077-4c1b-967c-4b767e5d6a51","Type":"ContainerStarted","Data":"90e57db732a86b43630ae65b8adc9a3761e0cb5e2e2a026189d239b54c9e8818"} Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.454474 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.462813 4846 generic.go:334] "Generic (PLEG): container finished" podID="a8dce541-0c00-476e-8efe-fea46f5fd26b" containerID="493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb" exitCode=0 Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.462863 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b686584c4-zpsks" event={"ID":"a8dce541-0c00-476e-8efe-fea46f5fd26b","Type":"ContainerDied","Data":"493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb"} Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.462888 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b686584c4-zpsks" event={"ID":"a8dce541-0c00-476e-8efe-fea46f5fd26b","Type":"ContainerDied","Data":"1febc541411b1d75707a3e5c2074281449c1ae411202f330f81cbeb589253295"} Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.462898 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b686584c4-zpsks" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.462913 4846 scope.go:117] "RemoveContainer" containerID="493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.496899 4846 scope.go:117] "RemoveContainer" containerID="07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.501450 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-combined-ca-bundle\") pod \"a8dce541-0c00-476e-8efe-fea46f5fd26b\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.501539 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-internal-tls-certs\") pod \"a8dce541-0c00-476e-8efe-fea46f5fd26b\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.501600 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8dce541-0c00-476e-8efe-fea46f5fd26b-logs\") pod \"a8dce541-0c00-476e-8efe-fea46f5fd26b\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.501794 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-config-data\") pod \"a8dce541-0c00-476e-8efe-fea46f5fd26b\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.501846 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-scripts\") pod \"a8dce541-0c00-476e-8efe-fea46f5fd26b\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.501879 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-public-tls-certs\") pod \"a8dce541-0c00-476e-8efe-fea46f5fd26b\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.501905 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tc8x9\" (UniqueName: \"kubernetes.io/projected/a8dce541-0c00-476e-8efe-fea46f5fd26b-kube-api-access-tc8x9\") pod \"a8dce541-0c00-476e-8efe-fea46f5fd26b\" (UID: \"a8dce541-0c00-476e-8efe-fea46f5fd26b\") " Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.502396 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8dce541-0c00-476e-8efe-fea46f5fd26b-logs" (OuterVolumeSpecName: "logs") pod "a8dce541-0c00-476e-8efe-fea46f5fd26b" (UID: "a8dce541-0c00-476e-8efe-fea46f5fd26b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.503360 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8dce541-0c00-476e-8efe-fea46f5fd26b-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.509013 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-scripts" (OuterVolumeSpecName: "scripts") pod "a8dce541-0c00-476e-8efe-fea46f5fd26b" (UID: "a8dce541-0c00-476e-8efe-fea46f5fd26b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.517595 4846 scope.go:117] "RemoveContainer" containerID="493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb" Feb 02 12:30:46 crc kubenswrapper[4846]: E0202 12:30:46.518241 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb\": container with ID starting with 493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb not found: ID does not exist" containerID="493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.518266 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb"} err="failed to get container status \"493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb\": rpc error: code = NotFound desc = could not find container \"493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb\": container with ID starting with 493a9ecc12100b86a78afeddb1fc64c6f883b4faef30dcc0d9121202381904bb not found: ID does not exist" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.518285 4846 scope.go:117] "RemoveContainer" containerID="07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376" Feb 02 12:30:46 crc kubenswrapper[4846]: E0202 12:30:46.518534 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376\": container with ID starting with 07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376 not found: ID does not exist" containerID="07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.518557 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376"} err="failed to get container status \"07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376\": rpc error: code = NotFound desc = could not find container \"07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376\": container with ID starting with 07f65ec772b28fd1a727a5ba744eb8c2feb27b26b779e6febb7e2ea00d339376 not found: ID does not exist" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.525410 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8dce541-0c00-476e-8efe-fea46f5fd26b-kube-api-access-tc8x9" (OuterVolumeSpecName: "kube-api-access-tc8x9") pod "a8dce541-0c00-476e-8efe-fea46f5fd26b" (UID: "a8dce541-0c00-476e-8efe-fea46f5fd26b"). InnerVolumeSpecName "kube-api-access-tc8x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.567429 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-config-data" (OuterVolumeSpecName: "config-data") pod "a8dce541-0c00-476e-8efe-fea46f5fd26b" (UID: "a8dce541-0c00-476e-8efe-fea46f5fd26b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.569790 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8dce541-0c00-476e-8efe-fea46f5fd26b" (UID: "a8dce541-0c00-476e-8efe-fea46f5fd26b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.606735 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tc8x9\" (UniqueName: \"kubernetes.io/projected/a8dce541-0c00-476e-8efe-fea46f5fd26b-kube-api-access-tc8x9\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.606775 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.606788 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.606801 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.617368 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a8dce541-0c00-476e-8efe-fea46f5fd26b" (UID: "a8dce541-0c00-476e-8efe-fea46f5fd26b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.658741 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a8dce541-0c00-476e-8efe-fea46f5fd26b" (UID: "a8dce541-0c00-476e-8efe-fea46f5fd26b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.708642 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.708678 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8dce541-0c00-476e-8efe-fea46f5fd26b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.805910 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7b686584c4-zpsks"] Feb 02 12:30:46 crc kubenswrapper[4846]: I0202 12:30:46.815755 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7b686584c4-zpsks"] Feb 02 12:30:47 crc kubenswrapper[4846]: I0202 12:30:47.437010 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8dce541-0c00-476e-8efe-fea46f5fd26b" path="/var/lib/kubelet/pods/a8dce541-0c00-476e-8efe-fea46f5fd26b/volumes" Feb 02 12:30:47 crc kubenswrapper[4846]: I0202 12:30:47.650779 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 02 12:30:47 crc kubenswrapper[4846]: I0202 12:30:47.650868 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 02 12:30:47 crc kubenswrapper[4846]: I0202 12:30:47.691416 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 02 12:30:47 crc kubenswrapper[4846]: I0202 12:30:47.697425 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 02 12:30:48 crc kubenswrapper[4846]: I0202 12:30:48.487674 4846 generic.go:334] "Generic (PLEG): container finished" podID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerID="bfda7d36305988bf593cee08d1f38c56c7cdb6fb74b7377ce314cd04037dadec" exitCode=0 Feb 02 12:30:48 crc kubenswrapper[4846]: I0202 12:30:48.489017 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd738edb-655a-40fc-937f-cdfbbd5d18db","Type":"ContainerDied","Data":"bfda7d36305988bf593cee08d1f38c56c7cdb6fb74b7377ce314cd04037dadec"} Feb 02 12:30:48 crc kubenswrapper[4846]: I0202 12:30:48.489058 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 02 12:30:48 crc kubenswrapper[4846]: I0202 12:30:48.489182 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 02 12:30:49 crc kubenswrapper[4846]: I0202 12:30:49.743243 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:49 crc kubenswrapper[4846]: I0202 12:30:49.743297 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:49 crc kubenswrapper[4846]: I0202 12:30:49.792401 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:49 crc kubenswrapper[4846]: I0202 12:30:49.830993 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:50 crc kubenswrapper[4846]: I0202 12:30:50.506723 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 12:30:50 crc kubenswrapper[4846]: I0202 12:30:50.507028 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 12:30:50 crc kubenswrapper[4846]: I0202 12:30:50.509734 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:50 crc kubenswrapper[4846]: I0202 12:30:50.509764 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:50 crc kubenswrapper[4846]: I0202 12:30:50.766227 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 02 12:30:50 crc kubenswrapper[4846]: I0202 12:30:50.869332 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 02 12:30:52 crc kubenswrapper[4846]: I0202 12:30:52.524340 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 12:30:52 crc kubenswrapper[4846]: I0202 12:30:52.524679 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 12:30:52 crc kubenswrapper[4846]: I0202 12:30:52.626351 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:52 crc kubenswrapper[4846]: I0202 12:30:52.628012 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.886962 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.969501 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75xfn\" (UniqueName: \"kubernetes.io/projected/cd738edb-655a-40fc-937f-cdfbbd5d18db-kube-api-access-75xfn\") pod \"cd738edb-655a-40fc-937f-cdfbbd5d18db\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.969847 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-run-httpd\") pod \"cd738edb-655a-40fc-937f-cdfbbd5d18db\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.969898 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-sg-core-conf-yaml\") pod \"cd738edb-655a-40fc-937f-cdfbbd5d18db\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.969944 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-log-httpd\") pod \"cd738edb-655a-40fc-937f-cdfbbd5d18db\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.969961 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-combined-ca-bundle\") pod \"cd738edb-655a-40fc-937f-cdfbbd5d18db\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.970016 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-scripts\") pod \"cd738edb-655a-40fc-937f-cdfbbd5d18db\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.970049 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-config-data\") pod \"cd738edb-655a-40fc-937f-cdfbbd5d18db\" (UID: \"cd738edb-655a-40fc-937f-cdfbbd5d18db\") " Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.978519 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cd738edb-655a-40fc-937f-cdfbbd5d18db" (UID: "cd738edb-655a-40fc-937f-cdfbbd5d18db"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.983067 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cd738edb-655a-40fc-937f-cdfbbd5d18db" (UID: "cd738edb-655a-40fc-937f-cdfbbd5d18db"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.991398 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-scripts" (OuterVolumeSpecName: "scripts") pod "cd738edb-655a-40fc-937f-cdfbbd5d18db" (UID: "cd738edb-655a-40fc-937f-cdfbbd5d18db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:53 crc kubenswrapper[4846]: I0202 12:30:53.991961 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd738edb-655a-40fc-937f-cdfbbd5d18db-kube-api-access-75xfn" (OuterVolumeSpecName: "kube-api-access-75xfn") pod "cd738edb-655a-40fc-937f-cdfbbd5d18db" (UID: "cd738edb-655a-40fc-937f-cdfbbd5d18db"). InnerVolumeSpecName "kube-api-access-75xfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.034373 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cd738edb-655a-40fc-937f-cdfbbd5d18db" (UID: "cd738edb-655a-40fc-937f-cdfbbd5d18db"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.073389 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.073414 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75xfn\" (UniqueName: \"kubernetes.io/projected/cd738edb-655a-40fc-937f-cdfbbd5d18db-kube-api-access-75xfn\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.073425 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.073434 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.073441 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd738edb-655a-40fc-937f-cdfbbd5d18db-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.085459 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd738edb-655a-40fc-937f-cdfbbd5d18db" (UID: "cd738edb-655a-40fc-937f-cdfbbd5d18db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.121743 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-config-data" (OuterVolumeSpecName: "config-data") pod "cd738edb-655a-40fc-937f-cdfbbd5d18db" (UID: "cd738edb-655a-40fc-937f-cdfbbd5d18db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.175542 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.175572 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd738edb-655a-40fc-937f-cdfbbd5d18db-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.542445 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g5kjm" event={"ID":"db0a7be1-2077-4c1b-967c-4b767e5d6a51","Type":"ContainerStarted","Data":"3c7bbc02055a0ecada54268b37ac8bc22b7506ea3e81485f4bcfe67d74919ccc"} Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.547795 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.549699 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd738edb-655a-40fc-937f-cdfbbd5d18db","Type":"ContainerDied","Data":"a7e61f6bc2901eace0444585c513b94fb5b384f895be35d589561d145e317a41"} Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.549749 4846 scope.go:117] "RemoveContainer" containerID="4d57a7e0b87f5992937c999de2d1144c544067617683b8753f5908886195c623" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.571255 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-g5kjm" podStartSLOduration=1.973981282 podStartE2EDuration="10.571236209s" podCreationTimestamp="2026-02-02 12:30:44 +0000 UTC" firstStartedPulling="2026-02-02 12:30:45.178225446 +0000 UTC m=+1276.406812319" lastFinishedPulling="2026-02-02 12:30:53.775480383 +0000 UTC m=+1285.004067246" observedRunningTime="2026-02-02 12:30:54.565103116 +0000 UTC m=+1285.793689989" watchObservedRunningTime="2026-02-02 12:30:54.571236209 +0000 UTC m=+1285.799823072" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.608041 4846 scope.go:117] "RemoveContainer" containerID="dfbf79fdd516504fe08b30c2a8d4a02c0a822265755aca5ca2510788866ae88b" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.616357 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.628689 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.635278 4846 scope.go:117] "RemoveContainer" containerID="04102fe16cf5633a4483e6d0f0a72eb749fefdc8ecc151fd5d15355d963aa402" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.658217 4846 scope.go:117] "RemoveContainer" containerID="bfda7d36305988bf593cee08d1f38c56c7cdb6fb74b7377ce314cd04037dadec" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.664128 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:54 crc kubenswrapper[4846]: E0202 12:30:54.664592 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="sg-core" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.664611 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="sg-core" Feb 02 12:30:54 crc kubenswrapper[4846]: E0202 12:30:54.664661 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8dce541-0c00-476e-8efe-fea46f5fd26b" containerName="placement-api" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.664671 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8dce541-0c00-476e-8efe-fea46f5fd26b" containerName="placement-api" Feb 02 12:30:54 crc kubenswrapper[4846]: E0202 12:30:54.664684 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="ceilometer-central-agent" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.664692 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="ceilometer-central-agent" Feb 02 12:30:54 crc kubenswrapper[4846]: E0202 12:30:54.664700 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="proxy-httpd" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.664709 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="proxy-httpd" Feb 02 12:30:54 crc kubenswrapper[4846]: E0202 12:30:54.664727 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="ceilometer-notification-agent" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.664735 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="ceilometer-notification-agent" Feb 02 12:30:54 crc kubenswrapper[4846]: E0202 12:30:54.664755 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8dce541-0c00-476e-8efe-fea46f5fd26b" containerName="placement-log" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.664763 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8dce541-0c00-476e-8efe-fea46f5fd26b" containerName="placement-log" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.664959 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8dce541-0c00-476e-8efe-fea46f5fd26b" containerName="placement-log" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.664969 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8dce541-0c00-476e-8efe-fea46f5fd26b" containerName="placement-api" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.664996 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="ceilometer-central-agent" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.665007 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="proxy-httpd" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.665020 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="sg-core" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.665032 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" containerName="ceilometer-notification-agent" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.668255 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.670575 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.670772 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.673992 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.786898 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-log-httpd\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.787226 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-scripts\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.787346 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6tpn\" (UniqueName: \"kubernetes.io/projected/12a752e3-9dbd-4d56-abd6-d4c58f463f86-kube-api-access-x6tpn\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.787950 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.788020 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.788121 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-run-httpd\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.788153 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-config-data\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.889962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.890010 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.890075 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-run-httpd\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.890107 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-config-data\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.890151 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-log-httpd\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.890180 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-scripts\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.890238 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6tpn\" (UniqueName: \"kubernetes.io/projected/12a752e3-9dbd-4d56-abd6-d4c58f463f86-kube-api-access-x6tpn\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.890737 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-run-httpd\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.891068 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-log-httpd\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.894000 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.894676 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-config-data\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.898370 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-scripts\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.902198 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.908084 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6tpn\" (UniqueName: \"kubernetes.io/projected/12a752e3-9dbd-4d56-abd6-d4c58f463f86-kube-api-access-x6tpn\") pod \"ceilometer-0\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " pod="openstack/ceilometer-0" Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.946813 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:54 crc kubenswrapper[4846]: I0202 12:30:54.947882 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:30:55 crc kubenswrapper[4846]: I0202 12:30:55.436761 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd738edb-655a-40fc-937f-cdfbbd5d18db" path="/var/lib/kubelet/pods/cd738edb-655a-40fc-937f-cdfbbd5d18db/volumes" Feb 02 12:30:55 crc kubenswrapper[4846]: I0202 12:30:55.475219 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:30:55 crc kubenswrapper[4846]: W0202 12:30:55.479238 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12a752e3_9dbd_4d56_abd6_d4c58f463f86.slice/crio-3ba57056734467dd24d0daef45b1a5487c480b14d37a87c715ebdb6ba6179a41 WatchSource:0}: Error finding container 3ba57056734467dd24d0daef45b1a5487c480b14d37a87c715ebdb6ba6179a41: Status 404 returned error can't find the container with id 3ba57056734467dd24d0daef45b1a5487c480b14d37a87c715ebdb6ba6179a41 Feb 02 12:30:55 crc kubenswrapper[4846]: I0202 12:30:55.560863 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12a752e3-9dbd-4d56-abd6-d4c58f463f86","Type":"ContainerStarted","Data":"3ba57056734467dd24d0daef45b1a5487c480b14d37a87c715ebdb6ba6179a41"} Feb 02 12:30:56 crc kubenswrapper[4846]: I0202 12:30:56.571675 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12a752e3-9dbd-4d56-abd6-d4c58f463f86","Type":"ContainerStarted","Data":"0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0"} Feb 02 12:30:57 crc kubenswrapper[4846]: I0202 12:30:57.584287 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12a752e3-9dbd-4d56-abd6-d4c58f463f86","Type":"ContainerStarted","Data":"b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60"} Feb 02 12:30:57 crc kubenswrapper[4846]: I0202 12:30:57.585709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12a752e3-9dbd-4d56-abd6-d4c58f463f86","Type":"ContainerStarted","Data":"fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea"} Feb 02 12:31:00 crc kubenswrapper[4846]: I0202 12:31:00.479071 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:31:00 crc kubenswrapper[4846]: I0202 12:31:00.479735 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:31:00 crc kubenswrapper[4846]: I0202 12:31:00.615044 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12a752e3-9dbd-4d56-abd6-d4c58f463f86","Type":"ContainerStarted","Data":"d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9"} Feb 02 12:31:00 crc kubenswrapper[4846]: I0202 12:31:00.615183 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="ceilometer-central-agent" containerID="cri-o://0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0" gracePeriod=30 Feb 02 12:31:00 crc kubenswrapper[4846]: I0202 12:31:00.615561 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 12:31:00 crc kubenswrapper[4846]: I0202 12:31:00.615265 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="ceilometer-notification-agent" containerID="cri-o://fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea" gracePeriod=30 Feb 02 12:31:00 crc kubenswrapper[4846]: I0202 12:31:00.615266 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="sg-core" containerID="cri-o://b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60" gracePeriod=30 Feb 02 12:31:00 crc kubenswrapper[4846]: I0202 12:31:00.615292 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="proxy-httpd" containerID="cri-o://d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9" gracePeriod=30 Feb 02 12:31:00 crc kubenswrapper[4846]: I0202 12:31:00.648644 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.497157682 podStartE2EDuration="6.648596523s" podCreationTimestamp="2026-02-02 12:30:54 +0000 UTC" firstStartedPulling="2026-02-02 12:30:55.481692176 +0000 UTC m=+1286.710279039" lastFinishedPulling="2026-02-02 12:30:59.633131017 +0000 UTC m=+1290.861717880" observedRunningTime="2026-02-02 12:31:00.646192254 +0000 UTC m=+1291.874779167" watchObservedRunningTime="2026-02-02 12:31:00.648596523 +0000 UTC m=+1291.877183406" Feb 02 12:31:01 crc kubenswrapper[4846]: I0202 12:31:01.636384 4846 generic.go:334] "Generic (PLEG): container finished" podID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerID="d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9" exitCode=0 Feb 02 12:31:01 crc kubenswrapper[4846]: I0202 12:31:01.636419 4846 generic.go:334] "Generic (PLEG): container finished" podID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerID="b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60" exitCode=2 Feb 02 12:31:01 crc kubenswrapper[4846]: I0202 12:31:01.636436 4846 generic.go:334] "Generic (PLEG): container finished" podID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerID="fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea" exitCode=0 Feb 02 12:31:01 crc kubenswrapper[4846]: I0202 12:31:01.636459 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12a752e3-9dbd-4d56-abd6-d4c58f463f86","Type":"ContainerDied","Data":"d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9"} Feb 02 12:31:01 crc kubenswrapper[4846]: I0202 12:31:01.636488 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12a752e3-9dbd-4d56-abd6-d4c58f463f86","Type":"ContainerDied","Data":"b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60"} Feb 02 12:31:01 crc kubenswrapper[4846]: I0202 12:31:01.636500 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12a752e3-9dbd-4d56-abd6-d4c58f463f86","Type":"ContainerDied","Data":"fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea"} Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.180439 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.269906 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-sg-core-conf-yaml\") pod \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.270188 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-run-httpd\") pod \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.270360 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-log-httpd\") pod \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.270450 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6tpn\" (UniqueName: \"kubernetes.io/projected/12a752e3-9dbd-4d56-abd6-d4c58f463f86-kube-api-access-x6tpn\") pod \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.271135 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "12a752e3-9dbd-4d56-abd6-d4c58f463f86" (UID: "12a752e3-9dbd-4d56-abd6-d4c58f463f86"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.271353 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "12a752e3-9dbd-4d56-abd6-d4c58f463f86" (UID: "12a752e3-9dbd-4d56-abd6-d4c58f463f86"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.271588 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-combined-ca-bundle\") pod \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.271701 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-config-data\") pod \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.272328 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-scripts\") pod \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\" (UID: \"12a752e3-9dbd-4d56-abd6-d4c58f463f86\") " Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.273024 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.273081 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12a752e3-9dbd-4d56-abd6-d4c58f463f86-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.276675 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12a752e3-9dbd-4d56-abd6-d4c58f463f86-kube-api-access-x6tpn" (OuterVolumeSpecName: "kube-api-access-x6tpn") pod "12a752e3-9dbd-4d56-abd6-d4c58f463f86" (UID: "12a752e3-9dbd-4d56-abd6-d4c58f463f86"). InnerVolumeSpecName "kube-api-access-x6tpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.280866 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-scripts" (OuterVolumeSpecName: "scripts") pod "12a752e3-9dbd-4d56-abd6-d4c58f463f86" (UID: "12a752e3-9dbd-4d56-abd6-d4c58f463f86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.310219 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "12a752e3-9dbd-4d56-abd6-d4c58f463f86" (UID: "12a752e3-9dbd-4d56-abd6-d4c58f463f86"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.347938 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12a752e3-9dbd-4d56-abd6-d4c58f463f86" (UID: "12a752e3-9dbd-4d56-abd6-d4c58f463f86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.357897 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-config-data" (OuterVolumeSpecName: "config-data") pod "12a752e3-9dbd-4d56-abd6-d4c58f463f86" (UID: "12a752e3-9dbd-4d56-abd6-d4c58f463f86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.374314 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.374342 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6tpn\" (UniqueName: \"kubernetes.io/projected/12a752e3-9dbd-4d56-abd6-d4c58f463f86-kube-api-access-x6tpn\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.374352 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.374360 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.374369 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12a752e3-9dbd-4d56-abd6-d4c58f463f86-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.665856 4846 generic.go:334] "Generic (PLEG): container finished" podID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerID="0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0" exitCode=0 Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.665921 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.665924 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12a752e3-9dbd-4d56-abd6-d4c58f463f86","Type":"ContainerDied","Data":"0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0"} Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.666023 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12a752e3-9dbd-4d56-abd6-d4c58f463f86","Type":"ContainerDied","Data":"3ba57056734467dd24d0daef45b1a5487c480b14d37a87c715ebdb6ba6179a41"} Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.666039 4846 scope.go:117] "RemoveContainer" containerID="d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.668320 4846 generic.go:334] "Generic (PLEG): container finished" podID="db0a7be1-2077-4c1b-967c-4b767e5d6a51" containerID="3c7bbc02055a0ecada54268b37ac8bc22b7506ea3e81485f4bcfe67d74919ccc" exitCode=0 Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.668372 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g5kjm" event={"ID":"db0a7be1-2077-4c1b-967c-4b767e5d6a51","Type":"ContainerDied","Data":"3c7bbc02055a0ecada54268b37ac8bc22b7506ea3e81485f4bcfe67d74919ccc"} Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.698169 4846 scope.go:117] "RemoveContainer" containerID="b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.721011 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.738710 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.742611 4846 scope.go:117] "RemoveContainer" containerID="fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.758758 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:04 crc kubenswrapper[4846]: E0202 12:31:04.759130 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="ceilometer-central-agent" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.759147 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="ceilometer-central-agent" Feb 02 12:31:04 crc kubenswrapper[4846]: E0202 12:31:04.759165 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="sg-core" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.759172 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="sg-core" Feb 02 12:31:04 crc kubenswrapper[4846]: E0202 12:31:04.759186 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="proxy-httpd" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.759192 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="proxy-httpd" Feb 02 12:31:04 crc kubenswrapper[4846]: E0202 12:31:04.759203 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="ceilometer-notification-agent" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.759209 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="ceilometer-notification-agent" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.759361 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="proxy-httpd" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.759377 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="ceilometer-notification-agent" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.759387 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="ceilometer-central-agent" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.759404 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" containerName="sg-core" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.774534 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.777525 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.777586 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.781170 4846 scope.go:117] "RemoveContainer" containerID="0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.797656 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.818531 4846 scope.go:117] "RemoveContainer" containerID="d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9" Feb 02 12:31:04 crc kubenswrapper[4846]: E0202 12:31:04.819813 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9\": container with ID starting with d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9 not found: ID does not exist" containerID="d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.819845 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9"} err="failed to get container status \"d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9\": rpc error: code = NotFound desc = could not find container \"d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9\": container with ID starting with d92265ec1498948f55f9495ca9e07b01c3b54e4f52120aab5712553f7e9086d9 not found: ID does not exist" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.819864 4846 scope.go:117] "RemoveContainer" containerID="b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60" Feb 02 12:31:04 crc kubenswrapper[4846]: E0202 12:31:04.822312 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60\": container with ID starting with b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60 not found: ID does not exist" containerID="b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.822362 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60"} err="failed to get container status \"b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60\": rpc error: code = NotFound desc = could not find container \"b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60\": container with ID starting with b17f06933d1dbad8065a94590fd5a3fc3f5dfe4489404b7d901b2f083c20bf60 not found: ID does not exist" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.822419 4846 scope.go:117] "RemoveContainer" containerID="fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea" Feb 02 12:31:04 crc kubenswrapper[4846]: E0202 12:31:04.827795 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea\": container with ID starting with fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea not found: ID does not exist" containerID="fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.827838 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea"} err="failed to get container status \"fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea\": rpc error: code = NotFound desc = could not find container \"fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea\": container with ID starting with fa36b545d6ed11dcaf8ae0848b69139083fdc4abe89629b94d0366f11fddfaea not found: ID does not exist" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.827869 4846 scope.go:117] "RemoveContainer" containerID="0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0" Feb 02 12:31:04 crc kubenswrapper[4846]: E0202 12:31:04.831784 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0\": container with ID starting with 0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0 not found: ID does not exist" containerID="0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.831826 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0"} err="failed to get container status \"0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0\": rpc error: code = NotFound desc = could not find container \"0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0\": container with ID starting with 0284528084dedb43e185aad08f4a3e92b9e19bc1b861e11a8a798e7f4d91dfa0 not found: ID does not exist" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.883171 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9wn4\" (UniqueName: \"kubernetes.io/projected/1eeec13e-8226-4641-90c7-31cd475bbb19-kube-api-access-x9wn4\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.883212 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-scripts\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.883231 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-config-data\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.883364 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-log-httpd\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.883489 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-run-httpd\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.883732 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.883849 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.985163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9wn4\" (UniqueName: \"kubernetes.io/projected/1eeec13e-8226-4641-90c7-31cd475bbb19-kube-api-access-x9wn4\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.985204 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-scripts\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.985219 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-config-data\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.985239 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-log-httpd\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.985272 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-run-httpd\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.985323 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.985354 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.986464 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-run-httpd\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.986553 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-log-httpd\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.990415 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.991654 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.991882 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-scripts\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:04 crc kubenswrapper[4846]: I0202 12:31:04.994186 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-config-data\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:05 crc kubenswrapper[4846]: I0202 12:31:05.018692 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9wn4\" (UniqueName: \"kubernetes.io/projected/1eeec13e-8226-4641-90c7-31cd475bbb19-kube-api-access-x9wn4\") pod \"ceilometer-0\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " pod="openstack/ceilometer-0" Feb 02 12:31:05 crc kubenswrapper[4846]: I0202 12:31:05.092256 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:05 crc kubenswrapper[4846]: I0202 12:31:05.436802 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12a752e3-9dbd-4d56-abd6-d4c58f463f86" path="/var/lib/kubelet/pods/12a752e3-9dbd-4d56-abd6-d4c58f463f86/volumes" Feb 02 12:31:05 crc kubenswrapper[4846]: I0202 12:31:05.540257 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:05 crc kubenswrapper[4846]: I0202 12:31:05.677255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eeec13e-8226-4641-90c7-31cd475bbb19","Type":"ContainerStarted","Data":"d61b7a402e445c3380a656f4debd05e1df88a2c9a4e1ab1f385837473a1d9282"} Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.006272 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.105295 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-combined-ca-bundle\") pod \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.105678 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpgcn\" (UniqueName: \"kubernetes.io/projected/db0a7be1-2077-4c1b-967c-4b767e5d6a51-kube-api-access-tpgcn\") pod \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.105807 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-scripts\") pod \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.105921 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-config-data\") pod \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\" (UID: \"db0a7be1-2077-4c1b-967c-4b767e5d6a51\") " Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.109299 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db0a7be1-2077-4c1b-967c-4b767e5d6a51-kube-api-access-tpgcn" (OuterVolumeSpecName: "kube-api-access-tpgcn") pod "db0a7be1-2077-4c1b-967c-4b767e5d6a51" (UID: "db0a7be1-2077-4c1b-967c-4b767e5d6a51"). InnerVolumeSpecName "kube-api-access-tpgcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.109346 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-scripts" (OuterVolumeSpecName: "scripts") pod "db0a7be1-2077-4c1b-967c-4b767e5d6a51" (UID: "db0a7be1-2077-4c1b-967c-4b767e5d6a51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.130125 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db0a7be1-2077-4c1b-967c-4b767e5d6a51" (UID: "db0a7be1-2077-4c1b-967c-4b767e5d6a51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.131219 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-config-data" (OuterVolumeSpecName: "config-data") pod "db0a7be1-2077-4c1b-967c-4b767e5d6a51" (UID: "db0a7be1-2077-4c1b-967c-4b767e5d6a51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.207683 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.207714 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.207724 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0a7be1-2077-4c1b-967c-4b767e5d6a51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.207733 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpgcn\" (UniqueName: \"kubernetes.io/projected/db0a7be1-2077-4c1b-967c-4b767e5d6a51-kube-api-access-tpgcn\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.686938 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-g5kjm" event={"ID":"db0a7be1-2077-4c1b-967c-4b767e5d6a51","Type":"ContainerDied","Data":"90e57db732a86b43630ae65b8adc9a3761e0cb5e2e2a026189d239b54c9e8818"} Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.686979 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90e57db732a86b43630ae65b8adc9a3761e0cb5e2e2a026189d239b54c9e8818" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.686988 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-g5kjm" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.688565 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eeec13e-8226-4641-90c7-31cd475bbb19","Type":"ContainerStarted","Data":"777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be"} Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.798252 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 12:31:06 crc kubenswrapper[4846]: E0202 12:31:06.798904 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0a7be1-2077-4c1b-967c-4b767e5d6a51" containerName="nova-cell0-conductor-db-sync" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.798920 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0a7be1-2077-4c1b-967c-4b767e5d6a51" containerName="nova-cell0-conductor-db-sync" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.799125 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0a7be1-2077-4c1b-967c-4b767e5d6a51" containerName="nova-cell0-conductor-db-sync" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.799722 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.804459 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pkshr" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.804719 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.812229 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.929465 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.929522 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6xjw\" (UniqueName: \"kubernetes.io/projected/c7dd2f87-cea2-4575-ad0c-483944433a55-kube-api-access-f6xjw\") pod \"nova-cell0-conductor-0\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:06 crc kubenswrapper[4846]: I0202 12:31:06.929887 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.031187 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.031239 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6xjw\" (UniqueName: \"kubernetes.io/projected/c7dd2f87-cea2-4575-ad0c-483944433a55-kube-api-access-f6xjw\") pod \"nova-cell0-conductor-0\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.031323 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.034867 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.036152 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.047503 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6xjw\" (UniqueName: \"kubernetes.io/projected/c7dd2f87-cea2-4575-ad0c-483944433a55-kube-api-access-f6xjw\") pod \"nova-cell0-conductor-0\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.187233 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.624535 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.721788 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c7dd2f87-cea2-4575-ad0c-483944433a55","Type":"ContainerStarted","Data":"fa693a993222ca3fa10756cb1e809d64ccf4f9a8cd192182b044fb70d408eab3"} Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.725074 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eeec13e-8226-4641-90c7-31cd475bbb19","Type":"ContainerStarted","Data":"f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a"} Feb 02 12:31:07 crc kubenswrapper[4846]: I0202 12:31:07.725237 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eeec13e-8226-4641-90c7-31cd475bbb19","Type":"ContainerStarted","Data":"aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1"} Feb 02 12:31:08 crc kubenswrapper[4846]: I0202 12:31:08.747137 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c7dd2f87-cea2-4575-ad0c-483944433a55","Type":"ContainerStarted","Data":"40dea2331c92832ae0ee4b74bc35edbdedf985815110294b594bde6cb57fb223"} Feb 02 12:31:08 crc kubenswrapper[4846]: I0202 12:31:08.747512 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:08 crc kubenswrapper[4846]: I0202 12:31:08.769495 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.769478591 podStartE2EDuration="2.769478591s" podCreationTimestamp="2026-02-02 12:31:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:08.764513588 +0000 UTC m=+1299.993100491" watchObservedRunningTime="2026-02-02 12:31:08.769478591 +0000 UTC m=+1299.998065454" Feb 02 12:31:10 crc kubenswrapper[4846]: I0202 12:31:10.765277 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eeec13e-8226-4641-90c7-31cd475bbb19","Type":"ContainerStarted","Data":"e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102"} Feb 02 12:31:10 crc kubenswrapper[4846]: I0202 12:31:10.765973 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 12:31:10 crc kubenswrapper[4846]: I0202 12:31:10.790033 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.081686625 podStartE2EDuration="6.790011935s" podCreationTimestamp="2026-02-02 12:31:04 +0000 UTC" firstStartedPulling="2026-02-02 12:31:05.543369387 +0000 UTC m=+1296.771956250" lastFinishedPulling="2026-02-02 12:31:10.251694697 +0000 UTC m=+1301.480281560" observedRunningTime="2026-02-02 12:31:10.782750355 +0000 UTC m=+1302.011337218" watchObservedRunningTime="2026-02-02 12:31:10.790011935 +0000 UTC m=+1302.018598788" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.233076 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.747452 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-2z6vf"] Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.748792 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.751251 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.757833 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.759714 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2z6vf"] Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.841011 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-scripts\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.841304 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-config-data\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.841575 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz757\" (UniqueName: \"kubernetes.io/projected/28d3c741-e2ee-45b6-bb97-61485cf748be-kube-api-access-nz757\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.841714 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.875797 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.876927 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.887722 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.887999 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.943553 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx2zt\" (UniqueName: \"kubernetes.io/projected/f43c0043-4619-4cd8-93e0-60f8f24158aa-kube-api-access-kx2zt\") pod \"nova-scheduler-0\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.943635 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz757\" (UniqueName: \"kubernetes.io/projected/28d3c741-e2ee-45b6-bb97-61485cf748be-kube-api-access-nz757\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.943685 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-config-data\") pod \"nova-scheduler-0\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.943701 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.943753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-scripts\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.943798 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-config-data\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.943824 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.951461 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-scripts\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.952397 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.959464 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-config-data\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.980925 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz757\" (UniqueName: \"kubernetes.io/projected/28d3c741-e2ee-45b6-bb97-61485cf748be-kube-api-access-nz757\") pod \"nova-cell0-cell-mapping-2z6vf\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:12 crc kubenswrapper[4846]: I0202 12:31:12.999535 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.004414 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.013522 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.032764 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.050014 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx2zt\" (UniqueName: \"kubernetes.io/projected/f43c0043-4619-4cd8-93e0-60f8f24158aa-kube-api-access-kx2zt\") pod \"nova-scheduler-0\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.050092 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-config-data\") pod \"nova-scheduler-0\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.050196 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.061334 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.075064 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.093414 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-config-data\") pod \"nova-scheduler-0\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.101425 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.107361 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.111899 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.127711 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx2zt\" (UniqueName: \"kubernetes.io/projected/f43c0043-4619-4cd8-93e0-60f8f24158aa-kube-api-access-kx2zt\") pod \"nova-scheduler-0\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.146586 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.153582 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g8zl\" (UniqueName: \"kubernetes.io/projected/191cb963-11cf-4914-b82d-63e018e91ca8-kube-api-access-6g8zl\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.153741 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-config-data\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.153769 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/191cb963-11cf-4914-b82d-63e018e91ca8-logs\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.168163 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.211147 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.212857 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.215783 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.222364 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.224103 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.262029 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-2wdnq"] Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.264783 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.271058 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.271124 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-config-data\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.271150 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.271169 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/191cb963-11cf-4914-b82d-63e018e91ca8-logs\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.271191 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.271211 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjcsm\" (UniqueName: \"kubernetes.io/projected/c8c6ee44-38ac-42b1-a961-bf62820300fe-kube-api-access-cjcsm\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.271289 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g8zl\" (UniqueName: \"kubernetes.io/projected/191cb963-11cf-4914-b82d-63e018e91ca8-kube-api-access-6g8zl\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.276084 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/191cb963-11cf-4914-b82d-63e018e91ca8-logs\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.279420 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.283059 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-config-data\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.300466 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-2wdnq"] Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.312413 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g8zl\" (UniqueName: \"kubernetes.io/projected/191cb963-11cf-4914-b82d-63e018e91ca8-kube-api-access-6g8zl\") pod \"nova-metadata-0\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.374793 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-sb\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.374847 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.374922 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjcsm\" (UniqueName: \"kubernetes.io/projected/c8c6ee44-38ac-42b1-a961-bf62820300fe-kube-api-access-cjcsm\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.374942 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-config-data\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.374983 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-svc\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.375050 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-swift-storage-0\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.375079 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5c4p\" (UniqueName: \"kubernetes.io/projected/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-kube-api-access-v5c4p\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.375121 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsqw6\" (UniqueName: \"kubernetes.io/projected/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-kube-api-access-wsqw6\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.375163 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-config\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.375187 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-nb\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.375206 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-logs\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.375229 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.375253 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.386094 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.389208 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.399345 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjcsm\" (UniqueName: \"kubernetes.io/projected/c8c6ee44-38ac-42b1-a961-bf62820300fe-kube-api-access-cjcsm\") pod \"nova-cell1-novncproxy-0\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.477844 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-sb\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.477919 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-config-data\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.477975 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-svc\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.478045 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-swift-storage-0\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.478075 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5c4p\" (UniqueName: \"kubernetes.io/projected/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-kube-api-access-v5c4p\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.478120 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsqw6\" (UniqueName: \"kubernetes.io/projected/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-kube-api-access-wsqw6\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.478182 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-config\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.478212 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-nb\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.478235 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-logs\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.478267 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.479503 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-sb\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.480927 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-config\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.482421 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-config-data\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.484598 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.496940 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-logs\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.497884 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-nb\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.500529 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5c4p\" (UniqueName: \"kubernetes.io/projected/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-kube-api-access-v5c4p\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.501207 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-svc\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.501450 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-swift-storage-0\") pod \"dnsmasq-dns-868bc9dc59-2wdnq\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.501935 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsqw6\" (UniqueName: \"kubernetes.io/projected/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-kube-api-access-wsqw6\") pod \"nova-api-0\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.526495 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.554949 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.583344 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.630275 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.770380 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2z6vf"] Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.813112 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2z6vf" event={"ID":"28d3c741-e2ee-45b6-bb97-61485cf748be","Type":"ContainerStarted","Data":"469d6599e235c24c936e26ce5df66f911975e82a89df629774d83c547c453b7e"} Feb 02 12:31:13 crc kubenswrapper[4846]: I0202 12:31:13.882745 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:31:13 crc kubenswrapper[4846]: W0202 12:31:13.901026 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf43c0043_4619_4cd8_93e0_60f8f24158aa.slice/crio-60846ed56961951ebfd75eda6945b62ac0c244148def375511a279a43681d33b WatchSource:0}: Error finding container 60846ed56961951ebfd75eda6945b62ac0c244148def375511a279a43681d33b: Status 404 returned error can't find the container with id 60846ed56961951ebfd75eda6945b62ac0c244148def375511a279a43681d33b Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.070185 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7fjsq"] Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.073304 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.075554 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.077559 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.101568 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7fjsq"] Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.196295 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.196493 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcxx6\" (UniqueName: \"kubernetes.io/projected/db37c129-8a93-4f2b-a21b-abe8565d1fee-kube-api-access-jcxx6\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.196782 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-scripts\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.196994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-config-data\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.244324 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.255098 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:14 crc kubenswrapper[4846]: W0202 12:31:14.270900 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8cbab16_182d_434d_9f8b_dc80bbab9dbc.slice/crio-d34995f1ac21e1300eb7227e833a1efaa99b28091c381264d9305b11500afdea WatchSource:0}: Error finding container d34995f1ac21e1300eb7227e833a1efaa99b28091c381264d9305b11500afdea: Status 404 returned error can't find the container with id d34995f1ac21e1300eb7227e833a1efaa99b28091c381264d9305b11500afdea Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.298670 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcxx6\" (UniqueName: \"kubernetes.io/projected/db37c129-8a93-4f2b-a21b-abe8565d1fee-kube-api-access-jcxx6\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.298782 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-scripts\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.298822 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-config-data\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.298888 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.306928 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.307546 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-scripts\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.310394 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-config-data\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.317879 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcxx6\" (UniqueName: \"kubernetes.io/projected/db37c129-8a93-4f2b-a21b-abe8565d1fee-kube-api-access-jcxx6\") pod \"nova-cell1-conductor-db-sync-7fjsq\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.396906 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.421859 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:31:14 crc kubenswrapper[4846]: W0202 12:31:14.423324 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8c6ee44_38ac_42b1_a961_bf62820300fe.slice/crio-9a5b3b78631139800fcae9fbb404803a9bddc2db807347f7977949c631ab80fb WatchSource:0}: Error finding container 9a5b3b78631139800fcae9fbb404803a9bddc2db807347f7977949c631ab80fb: Status 404 returned error can't find the container with id 9a5b3b78631139800fcae9fbb404803a9bddc2db807347f7977949c631ab80fb Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.466136 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-2wdnq"] Feb 02 12:31:14 crc kubenswrapper[4846]: W0202 12:31:14.472535 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda87c7f72_7e31_4e8c_a3dc_70c574a6dc93.slice/crio-3a7ec4dbe23cb3a13b82504f0419ba04b33412d3410803ce690b398bdfea047f WatchSource:0}: Error finding container 3a7ec4dbe23cb3a13b82504f0419ba04b33412d3410803ce690b398bdfea047f: Status 404 returned error can't find the container with id 3a7ec4dbe23cb3a13b82504f0419ba04b33412d3410803ce690b398bdfea047f Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.824965 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8cbab16-182d-434d-9f8b-dc80bbab9dbc","Type":"ContainerStarted","Data":"d34995f1ac21e1300eb7227e833a1efaa99b28091c381264d9305b11500afdea"} Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.826579 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"191cb963-11cf-4914-b82d-63e018e91ca8","Type":"ContainerStarted","Data":"864004bdc0416f32cfbe55ccbd5b4876832e88a216280b48da50ae94842a5e79"} Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.828574 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f43c0043-4619-4cd8-93e0-60f8f24158aa","Type":"ContainerStarted","Data":"60846ed56961951ebfd75eda6945b62ac0c244148def375511a279a43681d33b"} Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.830517 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c8c6ee44-38ac-42b1-a961-bf62820300fe","Type":"ContainerStarted","Data":"9a5b3b78631139800fcae9fbb404803a9bddc2db807347f7977949c631ab80fb"} Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.832035 4846 generic.go:334] "Generic (PLEG): container finished" podID="a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" containerID="fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31" exitCode=0 Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.832136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" event={"ID":"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93","Type":"ContainerDied","Data":"fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31"} Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.832746 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" event={"ID":"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93","Type":"ContainerStarted","Data":"3a7ec4dbe23cb3a13b82504f0419ba04b33412d3410803ce690b398bdfea047f"} Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.833714 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2z6vf" event={"ID":"28d3c741-e2ee-45b6-bb97-61485cf748be","Type":"ContainerStarted","Data":"b21926f23c5b1a73584f4ed6528b202d36ba1f36ff760cbf9ed6a2eaf4aaf751"} Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.889726 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-2z6vf" podStartSLOduration=2.889707369 podStartE2EDuration="2.889707369s" podCreationTimestamp="2026-02-02 12:31:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:14.880704166 +0000 UTC m=+1306.109291049" watchObservedRunningTime="2026-02-02 12:31:14.889707369 +0000 UTC m=+1306.118294232" Feb 02 12:31:14 crc kubenswrapper[4846]: I0202 12:31:14.909993 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7fjsq"] Feb 02 12:31:15 crc kubenswrapper[4846]: I0202 12:31:15.847768 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7fjsq" event={"ID":"db37c129-8a93-4f2b-a21b-abe8565d1fee","Type":"ContainerStarted","Data":"1b9592c0a9b70ced628973c2dd7cfe317289df85e6586e3fd6bf1d7f4fd1f2d9"} Feb 02 12:31:15 crc kubenswrapper[4846]: I0202 12:31:15.848024 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7fjsq" event={"ID":"db37c129-8a93-4f2b-a21b-abe8565d1fee","Type":"ContainerStarted","Data":"d458f44accfd1b790304e10d4d829898d8c76a9dbaf1a2d778df3ffa6381868f"} Feb 02 12:31:15 crc kubenswrapper[4846]: I0202 12:31:15.850077 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" event={"ID":"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93","Type":"ContainerStarted","Data":"5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680"} Feb 02 12:31:15 crc kubenswrapper[4846]: I0202 12:31:15.850273 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:15 crc kubenswrapper[4846]: I0202 12:31:15.867078 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-7fjsq" podStartSLOduration=1.867063368 podStartE2EDuration="1.867063368s" podCreationTimestamp="2026-02-02 12:31:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:15.862251559 +0000 UTC m=+1307.090838442" watchObservedRunningTime="2026-02-02 12:31:15.867063368 +0000 UTC m=+1307.095650231" Feb 02 12:31:15 crc kubenswrapper[4846]: I0202 12:31:15.884999 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" podStartSLOduration=2.884979154 podStartE2EDuration="2.884979154s" podCreationTimestamp="2026-02-02 12:31:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:15.883029616 +0000 UTC m=+1307.111616509" watchObservedRunningTime="2026-02-02 12:31:15.884979154 +0000 UTC m=+1307.113566017" Feb 02 12:31:16 crc kubenswrapper[4846]: I0202 12:31:16.487396 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:16 crc kubenswrapper[4846]: I0202 12:31:16.538560 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:31:17 crc kubenswrapper[4846]: I0202 12:31:17.873205 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8cbab16-182d-434d-9f8b-dc80bbab9dbc","Type":"ContainerStarted","Data":"6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c"} Feb 02 12:31:17 crc kubenswrapper[4846]: I0202 12:31:17.880310 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"191cb963-11cf-4914-b82d-63e018e91ca8","Type":"ContainerStarted","Data":"e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d"} Feb 02 12:31:17 crc kubenswrapper[4846]: I0202 12:31:17.884766 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f43c0043-4619-4cd8-93e0-60f8f24158aa","Type":"ContainerStarted","Data":"a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e"} Feb 02 12:31:17 crc kubenswrapper[4846]: I0202 12:31:17.889859 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c8c6ee44-38ac-42b1-a961-bf62820300fe","Type":"ContainerStarted","Data":"a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034"} Feb 02 12:31:17 crc kubenswrapper[4846]: I0202 12:31:17.889956 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="c8c6ee44-38ac-42b1-a961-bf62820300fe" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034" gracePeriod=30 Feb 02 12:31:17 crc kubenswrapper[4846]: I0202 12:31:17.904917 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.381797598 podStartE2EDuration="5.904900813s" podCreationTimestamp="2026-02-02 12:31:12 +0000 UTC" firstStartedPulling="2026-02-02 12:31:13.905732547 +0000 UTC m=+1305.134319410" lastFinishedPulling="2026-02-02 12:31:17.428835762 +0000 UTC m=+1308.657422625" observedRunningTime="2026-02-02 12:31:17.899763225 +0000 UTC m=+1309.128350108" watchObservedRunningTime="2026-02-02 12:31:17.904900813 +0000 UTC m=+1309.133487666" Feb 02 12:31:17 crc kubenswrapper[4846]: I0202 12:31:17.940120 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.897109454 podStartE2EDuration="4.940097227s" podCreationTimestamp="2026-02-02 12:31:13 +0000 UTC" firstStartedPulling="2026-02-02 12:31:14.427133024 +0000 UTC m=+1305.655719877" lastFinishedPulling="2026-02-02 12:31:17.470120787 +0000 UTC m=+1308.698707650" observedRunningTime="2026-02-02 12:31:17.925321271 +0000 UTC m=+1309.153908134" watchObservedRunningTime="2026-02-02 12:31:17.940097227 +0000 UTC m=+1309.168684110" Feb 02 12:31:18 crc kubenswrapper[4846]: I0202 12:31:18.225268 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 02 12:31:18 crc kubenswrapper[4846]: I0202 12:31:18.556049 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:18 crc kubenswrapper[4846]: I0202 12:31:18.900077 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8cbab16-182d-434d-9f8b-dc80bbab9dbc","Type":"ContainerStarted","Data":"197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f"} Feb 02 12:31:18 crc kubenswrapper[4846]: I0202 12:31:18.901950 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"191cb963-11cf-4914-b82d-63e018e91ca8","Type":"ContainerStarted","Data":"2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404"} Feb 02 12:31:18 crc kubenswrapper[4846]: I0202 12:31:18.902038 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="191cb963-11cf-4914-b82d-63e018e91ca8" containerName="nova-metadata-log" containerID="cri-o://e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d" gracePeriod=30 Feb 02 12:31:18 crc kubenswrapper[4846]: I0202 12:31:18.902073 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="191cb963-11cf-4914-b82d-63e018e91ca8" containerName="nova-metadata-metadata" containerID="cri-o://2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404" gracePeriod=30 Feb 02 12:31:18 crc kubenswrapper[4846]: I0202 12:31:18.923841 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.771028582 podStartE2EDuration="5.923825795s" podCreationTimestamp="2026-02-02 12:31:13 +0000 UTC" firstStartedPulling="2026-02-02 12:31:14.275799603 +0000 UTC m=+1305.504386466" lastFinishedPulling="2026-02-02 12:31:17.428596816 +0000 UTC m=+1308.657183679" observedRunningTime="2026-02-02 12:31:18.920026131 +0000 UTC m=+1310.148613014" watchObservedRunningTime="2026-02-02 12:31:18.923825795 +0000 UTC m=+1310.152412658" Feb 02 12:31:18 crc kubenswrapper[4846]: I0202 12:31:18.964997 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.802465484 podStartE2EDuration="6.964973628s" podCreationTimestamp="2026-02-02 12:31:12 +0000 UTC" firstStartedPulling="2026-02-02 12:31:14.265651331 +0000 UTC m=+1305.494238194" lastFinishedPulling="2026-02-02 12:31:17.428159475 +0000 UTC m=+1308.656746338" observedRunningTime="2026-02-02 12:31:18.946676333 +0000 UTC m=+1310.175263216" watchObservedRunningTime="2026-02-02 12:31:18.964973628 +0000 UTC m=+1310.193560491" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.512683 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.634237 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g8zl\" (UniqueName: \"kubernetes.io/projected/191cb963-11cf-4914-b82d-63e018e91ca8-kube-api-access-6g8zl\") pod \"191cb963-11cf-4914-b82d-63e018e91ca8\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.634433 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/191cb963-11cf-4914-b82d-63e018e91ca8-logs\") pod \"191cb963-11cf-4914-b82d-63e018e91ca8\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.634566 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-combined-ca-bundle\") pod \"191cb963-11cf-4914-b82d-63e018e91ca8\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.634708 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/191cb963-11cf-4914-b82d-63e018e91ca8-logs" (OuterVolumeSpecName: "logs") pod "191cb963-11cf-4914-b82d-63e018e91ca8" (UID: "191cb963-11cf-4914-b82d-63e018e91ca8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.634769 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-config-data\") pod \"191cb963-11cf-4914-b82d-63e018e91ca8\" (UID: \"191cb963-11cf-4914-b82d-63e018e91ca8\") " Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.635433 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/191cb963-11cf-4914-b82d-63e018e91ca8-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.644972 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/191cb963-11cf-4914-b82d-63e018e91ca8-kube-api-access-6g8zl" (OuterVolumeSpecName: "kube-api-access-6g8zl") pod "191cb963-11cf-4914-b82d-63e018e91ca8" (UID: "191cb963-11cf-4914-b82d-63e018e91ca8"). InnerVolumeSpecName "kube-api-access-6g8zl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.664409 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-config-data" (OuterVolumeSpecName: "config-data") pod "191cb963-11cf-4914-b82d-63e018e91ca8" (UID: "191cb963-11cf-4914-b82d-63e018e91ca8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.677317 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "191cb963-11cf-4914-b82d-63e018e91ca8" (UID: "191cb963-11cf-4914-b82d-63e018e91ca8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.737080 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.737115 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/191cb963-11cf-4914-b82d-63e018e91ca8-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.737128 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g8zl\" (UniqueName: \"kubernetes.io/projected/191cb963-11cf-4914-b82d-63e018e91ca8-kube-api-access-6g8zl\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.913166 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.913183 4846 generic.go:334] "Generic (PLEG): container finished" podID="191cb963-11cf-4914-b82d-63e018e91ca8" containerID="2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404" exitCode=0 Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.913213 4846 generic.go:334] "Generic (PLEG): container finished" podID="191cb963-11cf-4914-b82d-63e018e91ca8" containerID="e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d" exitCode=143 Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.913161 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"191cb963-11cf-4914-b82d-63e018e91ca8","Type":"ContainerDied","Data":"2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404"} Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.913651 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"191cb963-11cf-4914-b82d-63e018e91ca8","Type":"ContainerDied","Data":"e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d"} Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.913669 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"191cb963-11cf-4914-b82d-63e018e91ca8","Type":"ContainerDied","Data":"864004bdc0416f32cfbe55ccbd5b4876832e88a216280b48da50ae94842a5e79"} Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.913689 4846 scope.go:117] "RemoveContainer" containerID="2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.941239 4846 scope.go:117] "RemoveContainer" containerID="e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.950746 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.965837 4846 scope.go:117] "RemoveContainer" containerID="2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404" Feb 02 12:31:19 crc kubenswrapper[4846]: E0202 12:31:19.966374 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404\": container with ID starting with 2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404 not found: ID does not exist" containerID="2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.966414 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404"} err="failed to get container status \"2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404\": rpc error: code = NotFound desc = could not find container \"2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404\": container with ID starting with 2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404 not found: ID does not exist" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.966444 4846 scope.go:117] "RemoveContainer" containerID="e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d" Feb 02 12:31:19 crc kubenswrapper[4846]: E0202 12:31:19.968919 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d\": container with ID starting with e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d not found: ID does not exist" containerID="e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.968952 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d"} err="failed to get container status \"e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d\": rpc error: code = NotFound desc = could not find container \"e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d\": container with ID starting with e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d not found: ID does not exist" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.968975 4846 scope.go:117] "RemoveContainer" containerID="2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.970910 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404"} err="failed to get container status \"2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404\": rpc error: code = NotFound desc = could not find container \"2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404\": container with ID starting with 2337dcefc75bc61518b250a42a7621088392cc3690453416aed95a5cd6ba9404 not found: ID does not exist" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.970955 4846 scope.go:117] "RemoveContainer" containerID="e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.973330 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d"} err="failed to get container status \"e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d\": rpc error: code = NotFound desc = could not find container \"e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d\": container with ID starting with e0a822a4c2162dd35ee83419728b820a4b581fd86b9110938d3239f71d61ad6d not found: ID does not exist" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.974404 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.986777 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:19 crc kubenswrapper[4846]: E0202 12:31:19.987312 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="191cb963-11cf-4914-b82d-63e018e91ca8" containerName="nova-metadata-log" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.987337 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="191cb963-11cf-4914-b82d-63e018e91ca8" containerName="nova-metadata-log" Feb 02 12:31:19 crc kubenswrapper[4846]: E0202 12:31:19.987354 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="191cb963-11cf-4914-b82d-63e018e91ca8" containerName="nova-metadata-metadata" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.987361 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="191cb963-11cf-4914-b82d-63e018e91ca8" containerName="nova-metadata-metadata" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.987537 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="191cb963-11cf-4914-b82d-63e018e91ca8" containerName="nova-metadata-log" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.987557 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="191cb963-11cf-4914-b82d-63e018e91ca8" containerName="nova-metadata-metadata" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.989384 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.996146 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 02 12:31:19 crc kubenswrapper[4846]: I0202 12:31:19.996611 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.016354 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.144039 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.144196 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5c2b07-dd84-4f83-96a1-1921583726cf-logs\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.144368 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-config-data\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.144489 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.144535 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6nrk\" (UniqueName: \"kubernetes.io/projected/8c5c2b07-dd84-4f83-96a1-1921583726cf-kube-api-access-h6nrk\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.246777 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.246886 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5c2b07-dd84-4f83-96a1-1921583726cf-logs\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.246934 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-config-data\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.246968 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.246996 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6nrk\" (UniqueName: \"kubernetes.io/projected/8c5c2b07-dd84-4f83-96a1-1921583726cf-kube-api-access-h6nrk\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.248530 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5c2b07-dd84-4f83-96a1-1921583726cf-logs\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.252748 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.252893 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-config-data\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.253288 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.271243 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6nrk\" (UniqueName: \"kubernetes.io/projected/8c5c2b07-dd84-4f83-96a1-1921583726cf-kube-api-access-h6nrk\") pod \"nova-metadata-0\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.330460 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.771117 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:20 crc kubenswrapper[4846]: W0202 12:31:20.781122 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c5c2b07_dd84_4f83_96a1_1921583726cf.slice/crio-693274a3deb94cb39d2cc015c8f54032b0d6aab12f95cb0a36620dd5d78471f9 WatchSource:0}: Error finding container 693274a3deb94cb39d2cc015c8f54032b0d6aab12f95cb0a36620dd5d78471f9: Status 404 returned error can't find the container with id 693274a3deb94cb39d2cc015c8f54032b0d6aab12f95cb0a36620dd5d78471f9 Feb 02 12:31:20 crc kubenswrapper[4846]: I0202 12:31:20.922852 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c5c2b07-dd84-4f83-96a1-1921583726cf","Type":"ContainerStarted","Data":"693274a3deb94cb39d2cc015c8f54032b0d6aab12f95cb0a36620dd5d78471f9"} Feb 02 12:31:21 crc kubenswrapper[4846]: I0202 12:31:21.458654 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="191cb963-11cf-4914-b82d-63e018e91ca8" path="/var/lib/kubelet/pods/191cb963-11cf-4914-b82d-63e018e91ca8/volumes" Feb 02 12:31:21 crc kubenswrapper[4846]: I0202 12:31:21.938393 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c5c2b07-dd84-4f83-96a1-1921583726cf","Type":"ContainerStarted","Data":"d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6"} Feb 02 12:31:21 crc kubenswrapper[4846]: I0202 12:31:21.938448 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c5c2b07-dd84-4f83-96a1-1921583726cf","Type":"ContainerStarted","Data":"a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc"} Feb 02 12:31:21 crc kubenswrapper[4846]: I0202 12:31:21.995345 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.995315937 podStartE2EDuration="2.995315937s" podCreationTimestamp="2026-02-02 12:31:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:21.984004096 +0000 UTC m=+1313.212590979" watchObservedRunningTime="2026-02-02 12:31:21.995315937 +0000 UTC m=+1313.223902810" Feb 02 12:31:22 crc kubenswrapper[4846]: I0202 12:31:22.948412 4846 generic.go:334] "Generic (PLEG): container finished" podID="28d3c741-e2ee-45b6-bb97-61485cf748be" containerID="b21926f23c5b1a73584f4ed6528b202d36ba1f36ff760cbf9ed6a2eaf4aaf751" exitCode=0 Feb 02 12:31:22 crc kubenswrapper[4846]: I0202 12:31:22.948540 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2z6vf" event={"ID":"28d3c741-e2ee-45b6-bb97-61485cf748be","Type":"ContainerDied","Data":"b21926f23c5b1a73584f4ed6528b202d36ba1f36ff760cbf9ed6a2eaf4aaf751"} Feb 02 12:31:22 crc kubenswrapper[4846]: I0202 12:31:22.952274 4846 generic.go:334] "Generic (PLEG): container finished" podID="db37c129-8a93-4f2b-a21b-abe8565d1fee" containerID="1b9592c0a9b70ced628973c2dd7cfe317289df85e6586e3fd6bf1d7f4fd1f2d9" exitCode=0 Feb 02 12:31:22 crc kubenswrapper[4846]: I0202 12:31:22.952662 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7fjsq" event={"ID":"db37c129-8a93-4f2b-a21b-abe8565d1fee","Type":"ContainerDied","Data":"1b9592c0a9b70ced628973c2dd7cfe317289df85e6586e3fd6bf1d7f4fd1f2d9"} Feb 02 12:31:23 crc kubenswrapper[4846]: I0202 12:31:23.225106 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 02 12:31:23 crc kubenswrapper[4846]: I0202 12:31:23.258688 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 02 12:31:23 crc kubenswrapper[4846]: I0202 12:31:23.584949 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 12:31:23 crc kubenswrapper[4846]: I0202 12:31:23.585000 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 12:31:23 crc kubenswrapper[4846]: I0202 12:31:23.631661 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:31:23 crc kubenswrapper[4846]: I0202 12:31:23.686071 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-849fd69845-gdnm8"] Feb 02 12:31:23 crc kubenswrapper[4846]: I0202 12:31:23.686342 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" podUID="910f6926-083b-489e-b89c-7f8a46a1966e" containerName="dnsmasq-dns" containerID="cri-o://8100cf88b62535f5fde3d24535b341bccc58bbc5c406f5b2996fb84c5d013d72" gracePeriod=10 Feb 02 12:31:23 crc kubenswrapper[4846]: I0202 12:31:23.965450 4846 generic.go:334] "Generic (PLEG): container finished" podID="910f6926-083b-489e-b89c-7f8a46a1966e" containerID="8100cf88b62535f5fde3d24535b341bccc58bbc5c406f5b2996fb84c5d013d72" exitCode=0 Feb 02 12:31:23 crc kubenswrapper[4846]: I0202 12:31:23.965572 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" event={"ID":"910f6926-083b-489e-b89c-7f8a46a1966e","Type":"ContainerDied","Data":"8100cf88b62535f5fde3d24535b341bccc58bbc5c406f5b2996fb84c5d013d72"} Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.018222 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.242192 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.334994 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-nb\") pod \"910f6926-083b-489e-b89c-7f8a46a1966e\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.335041 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-svc\") pod \"910f6926-083b-489e-b89c-7f8a46a1966e\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.335121 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-sb\") pod \"910f6926-083b-489e-b89c-7f8a46a1966e\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.335167 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-swift-storage-0\") pod \"910f6926-083b-489e-b89c-7f8a46a1966e\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.335200 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5jss\" (UniqueName: \"kubernetes.io/projected/910f6926-083b-489e-b89c-7f8a46a1966e-kube-api-access-h5jss\") pod \"910f6926-083b-489e-b89c-7f8a46a1966e\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.335293 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-config\") pod \"910f6926-083b-489e-b89c-7f8a46a1966e\" (UID: \"910f6926-083b-489e-b89c-7f8a46a1966e\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.384463 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/910f6926-083b-489e-b89c-7f8a46a1966e-kube-api-access-h5jss" (OuterVolumeSpecName: "kube-api-access-h5jss") pod "910f6926-083b-489e-b89c-7f8a46a1966e" (UID: "910f6926-083b-489e-b89c-7f8a46a1966e"). InnerVolumeSpecName "kube-api-access-h5jss". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.416020 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "910f6926-083b-489e-b89c-7f8a46a1966e" (UID: "910f6926-083b-489e-b89c-7f8a46a1966e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.430943 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-config" (OuterVolumeSpecName: "config") pod "910f6926-083b-489e-b89c-7f8a46a1966e" (UID: "910f6926-083b-489e-b89c-7f8a46a1966e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.436049 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "910f6926-083b-489e-b89c-7f8a46a1966e" (UID: "910f6926-083b-489e-b89c-7f8a46a1966e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.437403 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5jss\" (UniqueName: \"kubernetes.io/projected/910f6926-083b-489e-b89c-7f8a46a1966e-kube-api-access-h5jss\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.437421 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.437429 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.437443 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.455316 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "910f6926-083b-489e-b89c-7f8a46a1966e" (UID: "910f6926-083b-489e-b89c-7f8a46a1966e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.482320 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "910f6926-083b-489e-b89c-7f8a46a1966e" (UID: "910f6926-083b-489e-b89c-7f8a46a1966e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.488931 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.498889 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.539345 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.539367 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/910f6926-083b-489e-b89c-7f8a46a1966e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.641193 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz757\" (UniqueName: \"kubernetes.io/projected/28d3c741-e2ee-45b6-bb97-61485cf748be-kube-api-access-nz757\") pod \"28d3c741-e2ee-45b6-bb97-61485cf748be\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.641404 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-combined-ca-bundle\") pod \"28d3c741-e2ee-45b6-bb97-61485cf748be\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.641436 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-config-data\") pod \"db37c129-8a93-4f2b-a21b-abe8565d1fee\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.641497 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-scripts\") pod \"db37c129-8a93-4f2b-a21b-abe8565d1fee\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.641562 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-scripts\") pod \"28d3c741-e2ee-45b6-bb97-61485cf748be\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.641659 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcxx6\" (UniqueName: \"kubernetes.io/projected/db37c129-8a93-4f2b-a21b-abe8565d1fee-kube-api-access-jcxx6\") pod \"db37c129-8a93-4f2b-a21b-abe8565d1fee\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.641719 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-combined-ca-bundle\") pod \"db37c129-8a93-4f2b-a21b-abe8565d1fee\" (UID: \"db37c129-8a93-4f2b-a21b-abe8565d1fee\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.641753 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-config-data\") pod \"28d3c741-e2ee-45b6-bb97-61485cf748be\" (UID: \"28d3c741-e2ee-45b6-bb97-61485cf748be\") " Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.645356 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db37c129-8a93-4f2b-a21b-abe8565d1fee-kube-api-access-jcxx6" (OuterVolumeSpecName: "kube-api-access-jcxx6") pod "db37c129-8a93-4f2b-a21b-abe8565d1fee" (UID: "db37c129-8a93-4f2b-a21b-abe8565d1fee"). InnerVolumeSpecName "kube-api-access-jcxx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.648451 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-scripts" (OuterVolumeSpecName: "scripts") pod "db37c129-8a93-4f2b-a21b-abe8565d1fee" (UID: "db37c129-8a93-4f2b-a21b-abe8565d1fee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.649727 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-scripts" (OuterVolumeSpecName: "scripts") pod "28d3c741-e2ee-45b6-bb97-61485cf748be" (UID: "28d3c741-e2ee-45b6-bb97-61485cf748be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.649989 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28d3c741-e2ee-45b6-bb97-61485cf748be-kube-api-access-nz757" (OuterVolumeSpecName: "kube-api-access-nz757") pod "28d3c741-e2ee-45b6-bb97-61485cf748be" (UID: "28d3c741-e2ee-45b6-bb97-61485cf748be"). InnerVolumeSpecName "kube-api-access-nz757". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.668216 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.668544 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.671574 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28d3c741-e2ee-45b6-bb97-61485cf748be" (UID: "28d3c741-e2ee-45b6-bb97-61485cf748be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.682914 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-config-data" (OuterVolumeSpecName: "config-data") pod "28d3c741-e2ee-45b6-bb97-61485cf748be" (UID: "28d3c741-e2ee-45b6-bb97-61485cf748be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.685870 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-config-data" (OuterVolumeSpecName: "config-data") pod "db37c129-8a93-4f2b-a21b-abe8565d1fee" (UID: "db37c129-8a93-4f2b-a21b-abe8565d1fee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.690854 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db37c129-8a93-4f2b-a21b-abe8565d1fee" (UID: "db37c129-8a93-4f2b-a21b-abe8565d1fee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.744727 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz757\" (UniqueName: \"kubernetes.io/projected/28d3c741-e2ee-45b6-bb97-61485cf748be-kube-api-access-nz757\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.745005 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.745125 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.745236 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.745377 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.745491 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcxx6\" (UniqueName: \"kubernetes.io/projected/db37c129-8a93-4f2b-a21b-abe8565d1fee-kube-api-access-jcxx6\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.745767 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db37c129-8a93-4f2b-a21b-abe8565d1fee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.745897 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28d3c741-e2ee-45b6-bb97-61485cf748be-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.981140 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2z6vf" event={"ID":"28d3c741-e2ee-45b6-bb97-61485cf748be","Type":"ContainerDied","Data":"469d6599e235c24c936e26ce5df66f911975e82a89df629774d83c547c453b7e"} Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.981493 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="469d6599e235c24c936e26ce5df66f911975e82a89df629774d83c547c453b7e" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.981146 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2z6vf" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.988837 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.988854 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849fd69845-gdnm8" event={"ID":"910f6926-083b-489e-b89c-7f8a46a1966e","Type":"ContainerDied","Data":"3ac2f45ab6d6663923341dbe2d10412efb0c4a33702b91b1797ab6a418c5934e"} Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.988913 4846 scope.go:117] "RemoveContainer" containerID="8100cf88b62535f5fde3d24535b341bccc58bbc5c406f5b2996fb84c5d013d72" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.992812 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7fjsq" Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.993592 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7fjsq" event={"ID":"db37c129-8a93-4f2b-a21b-abe8565d1fee","Type":"ContainerDied","Data":"d458f44accfd1b790304e10d4d829898d8c76a9dbaf1a2d778df3ffa6381868f"} Feb 02 12:31:24 crc kubenswrapper[4846]: I0202 12:31:24.993666 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d458f44accfd1b790304e10d4d829898d8c76a9dbaf1a2d778df3ffa6381868f" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.049967 4846 scope.go:117] "RemoveContainer" containerID="bc9070cb63fe7e9b5243f71356b106983b303c752a2ace197a2a53a26bcccd20" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.110181 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-849fd69845-gdnm8"] Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.122593 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-849fd69845-gdnm8"] Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.164326 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 12:31:25 crc kubenswrapper[4846]: E0202 12:31:25.165015 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="910f6926-083b-489e-b89c-7f8a46a1966e" containerName="dnsmasq-dns" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.165109 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="910f6926-083b-489e-b89c-7f8a46a1966e" containerName="dnsmasq-dns" Feb 02 12:31:25 crc kubenswrapper[4846]: E0202 12:31:25.165247 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="910f6926-083b-489e-b89c-7f8a46a1966e" containerName="init" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.165324 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="910f6926-083b-489e-b89c-7f8a46a1966e" containerName="init" Feb 02 12:31:25 crc kubenswrapper[4846]: E0202 12:31:25.165422 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db37c129-8a93-4f2b-a21b-abe8565d1fee" containerName="nova-cell1-conductor-db-sync" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.165528 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="db37c129-8a93-4f2b-a21b-abe8565d1fee" containerName="nova-cell1-conductor-db-sync" Feb 02 12:31:25 crc kubenswrapper[4846]: E0202 12:31:25.165625 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d3c741-e2ee-45b6-bb97-61485cf748be" containerName="nova-manage" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.165728 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d3c741-e2ee-45b6-bb97-61485cf748be" containerName="nova-manage" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.166327 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="910f6926-083b-489e-b89c-7f8a46a1966e" containerName="dnsmasq-dns" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.166435 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="db37c129-8a93-4f2b-a21b-abe8565d1fee" containerName="nova-cell1-conductor-db-sync" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.166519 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d3c741-e2ee-45b6-bb97-61485cf748be" containerName="nova-manage" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.167334 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.179090 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.195885 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.208259 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.208504 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerName="nova-api-log" containerID="cri-o://6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c" gracePeriod=30 Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.209039 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerName="nova-api-api" containerID="cri-o://197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f" gracePeriod=30 Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.229890 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.256177 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.256436 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c5c2b07-dd84-4f83-96a1-1921583726cf" containerName="nova-metadata-log" containerID="cri-o://a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc" gracePeriod=30 Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.256938 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c5c2b07-dd84-4f83-96a1-1921583726cf" containerName="nova-metadata-metadata" containerID="cri-o://d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6" gracePeriod=30 Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.259269 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.259465 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.259594 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpqfm\" (UniqueName: \"kubernetes.io/projected/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-kube-api-access-cpqfm\") pod \"nova-cell1-conductor-0\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.336846 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.336920 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.361713 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpqfm\" (UniqueName: \"kubernetes.io/projected/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-kube-api-access-cpqfm\") pod \"nova-cell1-conductor-0\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.361838 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.361956 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.367673 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.367809 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.388592 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpqfm\" (UniqueName: \"kubernetes.io/projected/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-kube-api-access-cpqfm\") pod \"nova-cell1-conductor-0\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.441499 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="910f6926-083b-489e-b89c-7f8a46a1966e" path="/var/lib/kubelet/pods/910f6926-083b-489e-b89c-7f8a46a1966e/volumes" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.493846 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.847958 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.978198 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-combined-ca-bundle\") pod \"8c5c2b07-dd84-4f83-96a1-1921583726cf\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.978300 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-nova-metadata-tls-certs\") pod \"8c5c2b07-dd84-4f83-96a1-1921583726cf\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.978341 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5c2b07-dd84-4f83-96a1-1921583726cf-logs\") pod \"8c5c2b07-dd84-4f83-96a1-1921583726cf\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.978383 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6nrk\" (UniqueName: \"kubernetes.io/projected/8c5c2b07-dd84-4f83-96a1-1921583726cf-kube-api-access-h6nrk\") pod \"8c5c2b07-dd84-4f83-96a1-1921583726cf\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.978478 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-config-data\") pod \"8c5c2b07-dd84-4f83-96a1-1921583726cf\" (UID: \"8c5c2b07-dd84-4f83-96a1-1921583726cf\") " Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.979298 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c5c2b07-dd84-4f83-96a1-1921583726cf-logs" (OuterVolumeSpecName: "logs") pod "8c5c2b07-dd84-4f83-96a1-1921583726cf" (UID: "8c5c2b07-dd84-4f83-96a1-1921583726cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:31:25 crc kubenswrapper[4846]: I0202 12:31:25.988056 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c5c2b07-dd84-4f83-96a1-1921583726cf-kube-api-access-h6nrk" (OuterVolumeSpecName: "kube-api-access-h6nrk") pod "8c5c2b07-dd84-4f83-96a1-1921583726cf" (UID: "8c5c2b07-dd84-4f83-96a1-1921583726cf"). InnerVolumeSpecName "kube-api-access-h6nrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.031909 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c5c2b07-dd84-4f83-96a1-1921583726cf" containerID="d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6" exitCode=0 Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.031960 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c5c2b07-dd84-4f83-96a1-1921583726cf" containerID="a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc" exitCode=143 Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.032079 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c5c2b07-dd84-4f83-96a1-1921583726cf","Type":"ContainerDied","Data":"d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6"} Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.032130 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c5c2b07-dd84-4f83-96a1-1921583726cf","Type":"ContainerDied","Data":"a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc"} Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.032146 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c5c2b07-dd84-4f83-96a1-1921583726cf","Type":"ContainerDied","Data":"693274a3deb94cb39d2cc015c8f54032b0d6aab12f95cb0a36620dd5d78471f9"} Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.032175 4846 scope.go:117] "RemoveContainer" containerID="d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.032377 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.035851 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c5c2b07-dd84-4f83-96a1-1921583726cf" (UID: "8c5c2b07-dd84-4f83-96a1-1921583726cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.043394 4846 generic.go:334] "Generic (PLEG): container finished" podID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerID="6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c" exitCode=143 Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.043818 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f43c0043-4619-4cd8-93e0-60f8f24158aa" containerName="nova-scheduler-scheduler" containerID="cri-o://a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e" gracePeriod=30 Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.044202 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8cbab16-182d-434d-9f8b-dc80bbab9dbc","Type":"ContainerDied","Data":"6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c"} Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.049844 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8c5c2b07-dd84-4f83-96a1-1921583726cf" (UID: "8c5c2b07-dd84-4f83-96a1-1921583726cf"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.081254 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.083101 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.083143 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c5c2b07-dd84-4f83-96a1-1921583726cf-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.083156 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6nrk\" (UniqueName: \"kubernetes.io/projected/8c5c2b07-dd84-4f83-96a1-1921583726cf-kube-api-access-h6nrk\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.085600 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-config-data" (OuterVolumeSpecName: "config-data") pod "8c5c2b07-dd84-4f83-96a1-1921583726cf" (UID: "8c5c2b07-dd84-4f83-96a1-1921583726cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.103720 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.185473 4846 scope.go:117] "RemoveContainer" containerID="a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.187246 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5c2b07-dd84-4f83-96a1-1921583726cf-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.215701 4846 scope.go:117] "RemoveContainer" containerID="d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6" Feb 02 12:31:26 crc kubenswrapper[4846]: E0202 12:31:26.216483 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6\": container with ID starting with d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6 not found: ID does not exist" containerID="d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.216549 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6"} err="failed to get container status \"d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6\": rpc error: code = NotFound desc = could not find container \"d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6\": container with ID starting with d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6 not found: ID does not exist" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.216588 4846 scope.go:117] "RemoveContainer" containerID="a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc" Feb 02 12:31:26 crc kubenswrapper[4846]: E0202 12:31:26.217048 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc\": container with ID starting with a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc not found: ID does not exist" containerID="a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.217087 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc"} err="failed to get container status \"a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc\": rpc error: code = NotFound desc = could not find container \"a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc\": container with ID starting with a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc not found: ID does not exist" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.217111 4846 scope.go:117] "RemoveContainer" containerID="d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.217421 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6"} err="failed to get container status \"d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6\": rpc error: code = NotFound desc = could not find container \"d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6\": container with ID starting with d0f5068863850f28bc850e3999b77a13815e0f8732b303cf579a1eb258414ac6 not found: ID does not exist" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.217448 4846 scope.go:117] "RemoveContainer" containerID="a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.217927 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc"} err="failed to get container status \"a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc\": rpc error: code = NotFound desc = could not find container \"a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc\": container with ID starting with a8ef0775fb7db4982cabf0b223acb4495c5bcc8b3c05b7c1c7fc6ee3c4937bbc not found: ID does not exist" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.401718 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.410081 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.427966 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:26 crc kubenswrapper[4846]: E0202 12:31:26.428437 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c5c2b07-dd84-4f83-96a1-1921583726cf" containerName="nova-metadata-log" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.428463 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c5c2b07-dd84-4f83-96a1-1921583726cf" containerName="nova-metadata-log" Feb 02 12:31:26 crc kubenswrapper[4846]: E0202 12:31:26.428507 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c5c2b07-dd84-4f83-96a1-1921583726cf" containerName="nova-metadata-metadata" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.428516 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c5c2b07-dd84-4f83-96a1-1921583726cf" containerName="nova-metadata-metadata" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.428852 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c5c2b07-dd84-4f83-96a1-1921583726cf" containerName="nova-metadata-log" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.428890 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c5c2b07-dd84-4f83-96a1-1921583726cf" containerName="nova-metadata-metadata" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.430189 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.433907 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.433955 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.446286 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.493202 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/053e0de3-f1a2-451d-9f01-8f26cc67fd92-logs\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.493586 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-config-data\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.493743 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.493873 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.493965 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hwzn\" (UniqueName: \"kubernetes.io/projected/053e0de3-f1a2-451d-9f01-8f26cc67fd92-kube-api-access-8hwzn\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.596911 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-config-data\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.597011 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.597091 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.597126 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hwzn\" (UniqueName: \"kubernetes.io/projected/053e0de3-f1a2-451d-9f01-8f26cc67fd92-kube-api-access-8hwzn\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.597164 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/053e0de3-f1a2-451d-9f01-8f26cc67fd92-logs\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.597598 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/053e0de3-f1a2-451d-9f01-8f26cc67fd92-logs\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.602471 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.602828 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.605834 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-config-data\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.623238 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hwzn\" (UniqueName: \"kubernetes.io/projected/053e0de3-f1a2-451d-9f01-8f26cc67fd92-kube-api-access-8hwzn\") pod \"nova-metadata-0\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " pod="openstack/nova-metadata-0" Feb 02 12:31:26 crc kubenswrapper[4846]: I0202 12:31:26.749302 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:31:27 crc kubenswrapper[4846]: I0202 12:31:27.063673 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d5f78273-c1e2-452f-8bcb-a51c9bc238e4","Type":"ContainerStarted","Data":"b89c77647ebfc096292368b4c128270a751702e9c4f3ecf26b326d816dff5fbc"} Feb 02 12:31:27 crc kubenswrapper[4846]: I0202 12:31:27.063719 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d5f78273-c1e2-452f-8bcb-a51c9bc238e4","Type":"ContainerStarted","Data":"8a05e0a9e87160b63b680e8161dfe05acfe7a085f7a20c66e6b6cf52a774d301"} Feb 02 12:31:27 crc kubenswrapper[4846]: I0202 12:31:27.063833 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:27 crc kubenswrapper[4846]: I0202 12:31:27.092349 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.092326678 podStartE2EDuration="2.092326678s" podCreationTimestamp="2026-02-02 12:31:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:27.082523914 +0000 UTC m=+1318.311110777" watchObservedRunningTime="2026-02-02 12:31:27.092326678 +0000 UTC m=+1318.320913541" Feb 02 12:31:27 crc kubenswrapper[4846]: I0202 12:31:27.262133 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:31:27 crc kubenswrapper[4846]: I0202 12:31:27.435610 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c5c2b07-dd84-4f83-96a1-1921583726cf" path="/var/lib/kubelet/pods/8c5c2b07-dd84-4f83-96a1-1921583726cf/volumes" Feb 02 12:31:28 crc kubenswrapper[4846]: I0202 12:31:28.074997 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"053e0de3-f1a2-451d-9f01-8f26cc67fd92","Type":"ContainerStarted","Data":"7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021"} Feb 02 12:31:28 crc kubenswrapper[4846]: I0202 12:31:28.075243 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"053e0de3-f1a2-451d-9f01-8f26cc67fd92","Type":"ContainerStarted","Data":"ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b"} Feb 02 12:31:28 crc kubenswrapper[4846]: I0202 12:31:28.075255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"053e0de3-f1a2-451d-9f01-8f26cc67fd92","Type":"ContainerStarted","Data":"9eed17d9c7dd77001d75ff7544c20c5736721d23ee18f3567c3f2b1307ffe755"} Feb 02 12:31:28 crc kubenswrapper[4846]: I0202 12:31:28.097780 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.097759114 podStartE2EDuration="2.097759114s" podCreationTimestamp="2026-02-02 12:31:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:28.091675942 +0000 UTC m=+1319.320262825" watchObservedRunningTime="2026-02-02 12:31:28.097759114 +0000 UTC m=+1319.326345967" Feb 02 12:31:28 crc kubenswrapper[4846]: E0202 12:31:28.226451 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 12:31:28 crc kubenswrapper[4846]: E0202 12:31:28.228035 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 12:31:28 crc kubenswrapper[4846]: E0202 12:31:28.229462 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 12:31:28 crc kubenswrapper[4846]: E0202 12:31:28.229508 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f43c0043-4619-4cd8-93e0-60f8f24158aa" containerName="nova-scheduler-scheduler" Feb 02 12:31:29 crc kubenswrapper[4846]: I0202 12:31:29.507838 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:31:29 crc kubenswrapper[4846]: I0202 12:31:29.577485 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx2zt\" (UniqueName: \"kubernetes.io/projected/f43c0043-4619-4cd8-93e0-60f8f24158aa-kube-api-access-kx2zt\") pod \"f43c0043-4619-4cd8-93e0-60f8f24158aa\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " Feb 02 12:31:29 crc kubenswrapper[4846]: I0202 12:31:29.577538 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-config-data\") pod \"f43c0043-4619-4cd8-93e0-60f8f24158aa\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " Feb 02 12:31:29 crc kubenswrapper[4846]: I0202 12:31:29.577574 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-combined-ca-bundle\") pod \"f43c0043-4619-4cd8-93e0-60f8f24158aa\" (UID: \"f43c0043-4619-4cd8-93e0-60f8f24158aa\") " Feb 02 12:31:29 crc kubenswrapper[4846]: I0202 12:31:29.584179 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f43c0043-4619-4cd8-93e0-60f8f24158aa-kube-api-access-kx2zt" (OuterVolumeSpecName: "kube-api-access-kx2zt") pod "f43c0043-4619-4cd8-93e0-60f8f24158aa" (UID: "f43c0043-4619-4cd8-93e0-60f8f24158aa"). InnerVolumeSpecName "kube-api-access-kx2zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:29 crc kubenswrapper[4846]: I0202 12:31:29.603565 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f43c0043-4619-4cd8-93e0-60f8f24158aa" (UID: "f43c0043-4619-4cd8-93e0-60f8f24158aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:29 crc kubenswrapper[4846]: I0202 12:31:29.605098 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-config-data" (OuterVolumeSpecName: "config-data") pod "f43c0043-4619-4cd8-93e0-60f8f24158aa" (UID: "f43c0043-4619-4cd8-93e0-60f8f24158aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:29 crc kubenswrapper[4846]: I0202 12:31:29.679850 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx2zt\" (UniqueName: \"kubernetes.io/projected/f43c0043-4619-4cd8-93e0-60f8f24158aa-kube-api-access-kx2zt\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:29 crc kubenswrapper[4846]: I0202 12:31:29.680485 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:29 crc kubenswrapper[4846]: I0202 12:31:29.680528 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43c0043-4619-4cd8-93e0-60f8f24158aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.102335 4846 generic.go:334] "Generic (PLEG): container finished" podID="f43c0043-4619-4cd8-93e0-60f8f24158aa" containerID="a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e" exitCode=0 Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.102395 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f43c0043-4619-4cd8-93e0-60f8f24158aa","Type":"ContainerDied","Data":"a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e"} Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.102423 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f43c0043-4619-4cd8-93e0-60f8f24158aa","Type":"ContainerDied","Data":"60846ed56961951ebfd75eda6945b62ac0c244148def375511a279a43681d33b"} Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.102441 4846 scope.go:117] "RemoveContainer" containerID="a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.102610 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.133568 4846 scope.go:117] "RemoveContainer" containerID="a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e" Feb 02 12:31:30 crc kubenswrapper[4846]: E0202 12:31:30.134044 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e\": container with ID starting with a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e not found: ID does not exist" containerID="a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.134100 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e"} err="failed to get container status \"a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e\": rpc error: code = NotFound desc = could not find container \"a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e\": container with ID starting with a2087fc530ed30198895b90b73a51ded78ace6d988c2e120215ce76ee7060a0e not found: ID does not exist" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.140084 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.150972 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.162415 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:31:30 crc kubenswrapper[4846]: E0202 12:31:30.162852 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43c0043-4619-4cd8-93e0-60f8f24158aa" containerName="nova-scheduler-scheduler" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.162870 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43c0043-4619-4cd8-93e0-60f8f24158aa" containerName="nova-scheduler-scheduler" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.163046 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43c0043-4619-4cd8-93e0-60f8f24158aa" containerName="nova-scheduler-scheduler" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.163860 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.165612 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.173549 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.289521 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-config-data\") pod \"nova-scheduler-0\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.289577 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.289782 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkd8q\" (UniqueName: \"kubernetes.io/projected/c3207510-5457-46af-b2ea-05b33efc61fb-kube-api-access-fkd8q\") pod \"nova-scheduler-0\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.391377 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkd8q\" (UniqueName: \"kubernetes.io/projected/c3207510-5457-46af-b2ea-05b33efc61fb-kube-api-access-fkd8q\") pod \"nova-scheduler-0\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.391475 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-config-data\") pod \"nova-scheduler-0\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.391506 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.395449 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.395690 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-config-data\") pod \"nova-scheduler-0\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.414728 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkd8q\" (UniqueName: \"kubernetes.io/projected/c3207510-5457-46af-b2ea-05b33efc61fb-kube-api-access-fkd8q\") pod \"nova-scheduler-0\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " pod="openstack/nova-scheduler-0" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.479398 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.479466 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.479550 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.480332 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5342c3479165e81b2ca97736d850f5b8bc8a41ed04e9666168601ccf61d9277c"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.480401 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://5342c3479165e81b2ca97736d850f5b8bc8a41ed04e9666168601ccf61d9277c" gracePeriod=600 Feb 02 12:31:30 crc kubenswrapper[4846]: I0202 12:31:30.497324 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.086474 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.106065 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.114126 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="5342c3479165e81b2ca97736d850f5b8bc8a41ed04e9666168601ccf61d9277c" exitCode=0 Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.114194 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"5342c3479165e81b2ca97736d850f5b8bc8a41ed04e9666168601ccf61d9277c"} Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.114251 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b"} Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.114273 4846 scope.go:117] "RemoveContainer" containerID="d4f60d595c3f3559b0b02374c951ace4cd6f20670671fc67c4b4653981928977" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.116150 4846 generic.go:334] "Generic (PLEG): container finished" podID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerID="197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f" exitCode=0 Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.116207 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8cbab16-182d-434d-9f8b-dc80bbab9dbc","Type":"ContainerDied","Data":"197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f"} Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.116228 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c8cbab16-182d-434d-9f8b-dc80bbab9dbc","Type":"ContainerDied","Data":"d34995f1ac21e1300eb7227e833a1efaa99b28091c381264d9305b11500afdea"} Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.116279 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.120968 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c3207510-5457-46af-b2ea-05b33efc61fb","Type":"ContainerStarted","Data":"ad144c609948fc4b457c6a8013010d9fe6baa30f993db447fcefd81150db17fb"} Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.172895 4846 scope.go:117] "RemoveContainer" containerID="197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.202682 4846 scope.go:117] "RemoveContainer" containerID="6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.206287 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-logs\") pod \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.206327 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-combined-ca-bundle\") pod \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.206468 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-config-data\") pod \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.206599 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsqw6\" (UniqueName: \"kubernetes.io/projected/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-kube-api-access-wsqw6\") pod \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\" (UID: \"c8cbab16-182d-434d-9f8b-dc80bbab9dbc\") " Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.206872 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-logs" (OuterVolumeSpecName: "logs") pod "c8cbab16-182d-434d-9f8b-dc80bbab9dbc" (UID: "c8cbab16-182d-434d-9f8b-dc80bbab9dbc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.207281 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.210827 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-kube-api-access-wsqw6" (OuterVolumeSpecName: "kube-api-access-wsqw6") pod "c8cbab16-182d-434d-9f8b-dc80bbab9dbc" (UID: "c8cbab16-182d-434d-9f8b-dc80bbab9dbc"). InnerVolumeSpecName "kube-api-access-wsqw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.237730 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-config-data" (OuterVolumeSpecName: "config-data") pod "c8cbab16-182d-434d-9f8b-dc80bbab9dbc" (UID: "c8cbab16-182d-434d-9f8b-dc80bbab9dbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.238129 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8cbab16-182d-434d-9f8b-dc80bbab9dbc" (UID: "c8cbab16-182d-434d-9f8b-dc80bbab9dbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.245837 4846 scope.go:117] "RemoveContainer" containerID="197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f" Feb 02 12:31:31 crc kubenswrapper[4846]: E0202 12:31:31.246414 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f\": container with ID starting with 197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f not found: ID does not exist" containerID="197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.246463 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f"} err="failed to get container status \"197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f\": rpc error: code = NotFound desc = could not find container \"197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f\": container with ID starting with 197391934dced15d707bd82285aab4fdeeda4793ada2a9e768f0cdff1973982f not found: ID does not exist" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.246490 4846 scope.go:117] "RemoveContainer" containerID="6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c" Feb 02 12:31:31 crc kubenswrapper[4846]: E0202 12:31:31.247221 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c\": container with ID starting with 6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c not found: ID does not exist" containerID="6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.247248 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c"} err="failed to get container status \"6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c\": rpc error: code = NotFound desc = could not find container \"6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c\": container with ID starting with 6edeb889a0f40a7d39a3479e82a4997eb0946d88592c8b9e612d51154f1eac9c not found: ID does not exist" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.308649 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsqw6\" (UniqueName: \"kubernetes.io/projected/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-kube-api-access-wsqw6\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.308683 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.308693 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cbab16-182d-434d-9f8b-dc80bbab9dbc-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.437410 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f43c0043-4619-4cd8-93e0-60f8f24158aa" path="/var/lib/kubelet/pods/f43c0043-4619-4cd8-93e0-60f8f24158aa/volumes" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.485045 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.505947 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.515351 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:31 crc kubenswrapper[4846]: E0202 12:31:31.523888 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerName="nova-api-log" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.523929 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerName="nova-api-log" Feb 02 12:31:31 crc kubenswrapper[4846]: E0202 12:31:31.524010 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerName="nova-api-api" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.524018 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerName="nova-api-api" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.524563 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerName="nova-api-api" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.524577 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" containerName="nova-api-log" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.525916 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.529171 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.532469 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.613490 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.613666 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-config-data\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.613733 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v66gr\" (UniqueName: \"kubernetes.io/projected/22b0ef8e-3708-4a24-8be4-f71edc795bc5-kube-api-access-v66gr\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.613799 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22b0ef8e-3708-4a24-8be4-f71edc795bc5-logs\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.715219 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-config-data\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.715323 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v66gr\" (UniqueName: \"kubernetes.io/projected/22b0ef8e-3708-4a24-8be4-f71edc795bc5-kube-api-access-v66gr\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.715402 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22b0ef8e-3708-4a24-8be4-f71edc795bc5-logs\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.715547 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.715865 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22b0ef8e-3708-4a24-8be4-f71edc795bc5-logs\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.721305 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.722723 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-config-data\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.736475 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v66gr\" (UniqueName: \"kubernetes.io/projected/22b0ef8e-3708-4a24-8be4-f71edc795bc5-kube-api-access-v66gr\") pod \"nova-api-0\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " pod="openstack/nova-api-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.751093 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.751139 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 12:31:31 crc kubenswrapper[4846]: I0202 12:31:31.851189 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:31:32 crc kubenswrapper[4846]: I0202 12:31:32.135325 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c3207510-5457-46af-b2ea-05b33efc61fb","Type":"ContainerStarted","Data":"42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c"} Feb 02 12:31:32 crc kubenswrapper[4846]: I0202 12:31:32.161276 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.161259209 podStartE2EDuration="2.161259209s" podCreationTimestamp="2026-02-02 12:31:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:32.151319132 +0000 UTC m=+1323.379906005" watchObservedRunningTime="2026-02-02 12:31:32.161259209 +0000 UTC m=+1323.389846062" Feb 02 12:31:32 crc kubenswrapper[4846]: I0202 12:31:32.317704 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:33 crc kubenswrapper[4846]: I0202 12:31:33.144893 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22b0ef8e-3708-4a24-8be4-f71edc795bc5","Type":"ContainerStarted","Data":"0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71"} Feb 02 12:31:33 crc kubenswrapper[4846]: I0202 12:31:33.145273 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22b0ef8e-3708-4a24-8be4-f71edc795bc5","Type":"ContainerStarted","Data":"7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad"} Feb 02 12:31:33 crc kubenswrapper[4846]: I0202 12:31:33.145472 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22b0ef8e-3708-4a24-8be4-f71edc795bc5","Type":"ContainerStarted","Data":"113b72b061ff1e94e3265a6db031c6697f679249f7bbdd04fd338e69f650787f"} Feb 02 12:31:33 crc kubenswrapper[4846]: I0202 12:31:33.166649 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.166607884 podStartE2EDuration="2.166607884s" podCreationTimestamp="2026-02-02 12:31:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:33.162125873 +0000 UTC m=+1324.390712736" watchObservedRunningTime="2026-02-02 12:31:33.166607884 +0000 UTC m=+1324.395194767" Feb 02 12:31:33 crc kubenswrapper[4846]: I0202 12:31:33.439414 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8cbab16-182d-434d-9f8b-dc80bbab9dbc" path="/var/lib/kubelet/pods/c8cbab16-182d-434d-9f8b-dc80bbab9dbc/volumes" Feb 02 12:31:35 crc kubenswrapper[4846]: I0202 12:31:35.101587 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 02 12:31:35 crc kubenswrapper[4846]: I0202 12:31:35.497413 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 02 12:31:35 crc kubenswrapper[4846]: I0202 12:31:35.524095 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 02 12:31:36 crc kubenswrapper[4846]: I0202 12:31:36.751021 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 12:31:36 crc kubenswrapper[4846]: I0202 12:31:36.751372 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 12:31:37 crc kubenswrapper[4846]: I0202 12:31:37.764876 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 12:31:37 crc kubenswrapper[4846]: I0202 12:31:37.764895 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 12:31:38 crc kubenswrapper[4846]: I0202 12:31:38.720937 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:31:38 crc kubenswrapper[4846]: I0202 12:31:38.721408 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="8c0a429c-a50e-493c-958e-61d8a3a404c6" containerName="kube-state-metrics" containerID="cri-o://1d67d357e4081d3bfaa9f34fd59e7809db1eb45b38eb505b82cd78ccaa91a725" gracePeriod=30 Feb 02 12:31:39 crc kubenswrapper[4846]: I0202 12:31:39.206189 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c0a429c-a50e-493c-958e-61d8a3a404c6" containerID="1d67d357e4081d3bfaa9f34fd59e7809db1eb45b38eb505b82cd78ccaa91a725" exitCode=2 Feb 02 12:31:39 crc kubenswrapper[4846]: I0202 12:31:39.206297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8c0a429c-a50e-493c-958e-61d8a3a404c6","Type":"ContainerDied","Data":"1d67d357e4081d3bfaa9f34fd59e7809db1eb45b38eb505b82cd78ccaa91a725"} Feb 02 12:31:39 crc kubenswrapper[4846]: I0202 12:31:39.206522 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8c0a429c-a50e-493c-958e-61d8a3a404c6","Type":"ContainerDied","Data":"58310c62e7ca162822f3972aa3f1805f863e8fe7b7d161842fffff5af17b05f9"} Feb 02 12:31:39 crc kubenswrapper[4846]: I0202 12:31:39.206540 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58310c62e7ca162822f3972aa3f1805f863e8fe7b7d161842fffff5af17b05f9" Feb 02 12:31:39 crc kubenswrapper[4846]: I0202 12:31:39.247171 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 12:31:39 crc kubenswrapper[4846]: I0202 12:31:39.354347 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4mn2\" (UniqueName: \"kubernetes.io/projected/8c0a429c-a50e-493c-958e-61d8a3a404c6-kube-api-access-q4mn2\") pod \"8c0a429c-a50e-493c-958e-61d8a3a404c6\" (UID: \"8c0a429c-a50e-493c-958e-61d8a3a404c6\") " Feb 02 12:31:39 crc kubenswrapper[4846]: I0202 12:31:39.360088 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c0a429c-a50e-493c-958e-61d8a3a404c6-kube-api-access-q4mn2" (OuterVolumeSpecName: "kube-api-access-q4mn2") pod "8c0a429c-a50e-493c-958e-61d8a3a404c6" (UID: "8c0a429c-a50e-493c-958e-61d8a3a404c6"). InnerVolumeSpecName "kube-api-access-q4mn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:39 crc kubenswrapper[4846]: I0202 12:31:39.456641 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4mn2\" (UniqueName: \"kubernetes.io/projected/8c0a429c-a50e-493c-958e-61d8a3a404c6-kube-api-access-q4mn2\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.213517 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.233347 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.248305 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.257365 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:31:40 crc kubenswrapper[4846]: E0202 12:31:40.257861 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0a429c-a50e-493c-958e-61d8a3a404c6" containerName="kube-state-metrics" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.257884 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0a429c-a50e-493c-958e-61d8a3a404c6" containerName="kube-state-metrics" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.258135 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c0a429c-a50e-493c-958e-61d8a3a404c6" containerName="kube-state-metrics" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.258909 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.261209 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.261846 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.269117 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.375843 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.375902 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.375940 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.376061 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbjp4\" (UniqueName: \"kubernetes.io/projected/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-api-access-hbjp4\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.477687 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.478229 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.478555 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.479021 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbjp4\" (UniqueName: \"kubernetes.io/projected/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-api-access-hbjp4\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.484475 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.485276 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.485491 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.497667 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.499380 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbjp4\" (UniqueName: \"kubernetes.io/projected/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-api-access-hbjp4\") pod \"kube-state-metrics-0\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " pod="openstack/kube-state-metrics-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.530957 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.573759 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.574053 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="ceilometer-central-agent" containerID="cri-o://777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be" gracePeriod=30 Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.574263 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="proxy-httpd" containerID="cri-o://e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102" gracePeriod=30 Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.574340 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="ceilometer-notification-agent" containerID="cri-o://aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1" gracePeriod=30 Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.574493 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="sg-core" containerID="cri-o://f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a" gracePeriod=30 Feb 02 12:31:40 crc kubenswrapper[4846]: I0202 12:31:40.584811 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.067689 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.223892 4846 generic.go:334] "Generic (PLEG): container finished" podID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerID="e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102" exitCode=0 Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.223930 4846 generic.go:334] "Generic (PLEG): container finished" podID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerID="f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a" exitCode=2 Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.223940 4846 generic.go:334] "Generic (PLEG): container finished" podID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerID="777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be" exitCode=0 Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.223953 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eeec13e-8226-4641-90c7-31cd475bbb19","Type":"ContainerDied","Data":"e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102"} Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.223997 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eeec13e-8226-4641-90c7-31cd475bbb19","Type":"ContainerDied","Data":"f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a"} Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.224009 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eeec13e-8226-4641-90c7-31cd475bbb19","Type":"ContainerDied","Data":"777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be"} Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.225512 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9a52309a-c6e4-481a-9fba-69de4c57e6b1","Type":"ContainerStarted","Data":"72249adb66feaa7262b83c820c56d294b09b1b83a53e9d1b6f4f6ee1495786c2"} Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.260143 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.438424 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c0a429c-a50e-493c-958e-61d8a3a404c6" path="/var/lib/kubelet/pods/8c0a429c-a50e-493c-958e-61d8a3a404c6/volumes" Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.852678 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 12:31:41 crc kubenswrapper[4846]: I0202 12:31:41.852723 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 12:31:42 crc kubenswrapper[4846]: I0202 12:31:42.240172 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9a52309a-c6e4-481a-9fba-69de4c57e6b1","Type":"ContainerStarted","Data":"218c4aa156faa3e69e561faabd794bbb67a15863d0a9d773a8b44e2e31cb4a4c"} Feb 02 12:31:42 crc kubenswrapper[4846]: I0202 12:31:42.240544 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 02 12:31:42 crc kubenswrapper[4846]: I0202 12:31:42.268321 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.9218644889999998 podStartE2EDuration="2.268304969s" podCreationTimestamp="2026-02-02 12:31:40 +0000 UTC" firstStartedPulling="2026-02-02 12:31:41.071199588 +0000 UTC m=+1332.299786451" lastFinishedPulling="2026-02-02 12:31:41.417640048 +0000 UTC m=+1332.646226931" observedRunningTime="2026-02-02 12:31:42.253223254 +0000 UTC m=+1333.481810117" watchObservedRunningTime="2026-02-02 12:31:42.268304969 +0000 UTC m=+1333.496891832" Feb 02 12:31:42 crc kubenswrapper[4846]: I0202 12:31:42.934852 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 12:31:42 crc kubenswrapper[4846]: I0202 12:31:42.934856 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.257423 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.259239 4846 generic.go:334] "Generic (PLEG): container finished" podID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerID="aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1" exitCode=0 Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.260103 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eeec13e-8226-4641-90c7-31cd475bbb19","Type":"ContainerDied","Data":"aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1"} Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.260135 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1eeec13e-8226-4641-90c7-31cd475bbb19","Type":"ContainerDied","Data":"d61b7a402e445c3380a656f4debd05e1df88a2c9a4e1ab1f385837473a1d9282"} Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.260153 4846 scope.go:117] "RemoveContainer" containerID="e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.284397 4846 scope.go:117] "RemoveContainer" containerID="f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.308185 4846 scope.go:117] "RemoveContainer" containerID="aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.344779 4846 scope.go:117] "RemoveContainer" containerID="777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.362401 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-config-data\") pod \"1eeec13e-8226-4641-90c7-31cd475bbb19\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.362492 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9wn4\" (UniqueName: \"kubernetes.io/projected/1eeec13e-8226-4641-90c7-31cd475bbb19-kube-api-access-x9wn4\") pod \"1eeec13e-8226-4641-90c7-31cd475bbb19\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.362521 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-scripts\") pod \"1eeec13e-8226-4641-90c7-31cd475bbb19\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.362554 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-run-httpd\") pod \"1eeec13e-8226-4641-90c7-31cd475bbb19\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.362634 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-sg-core-conf-yaml\") pod \"1eeec13e-8226-4641-90c7-31cd475bbb19\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.362817 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-log-httpd\") pod \"1eeec13e-8226-4641-90c7-31cd475bbb19\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.362849 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-combined-ca-bundle\") pod \"1eeec13e-8226-4641-90c7-31cd475bbb19\" (UID: \"1eeec13e-8226-4641-90c7-31cd475bbb19\") " Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.363113 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1eeec13e-8226-4641-90c7-31cd475bbb19" (UID: "1eeec13e-8226-4641-90c7-31cd475bbb19"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.363468 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1eeec13e-8226-4641-90c7-31cd475bbb19" (UID: "1eeec13e-8226-4641-90c7-31cd475bbb19"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.363842 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.363874 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1eeec13e-8226-4641-90c7-31cd475bbb19-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.368451 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-scripts" (OuterVolumeSpecName: "scripts") pod "1eeec13e-8226-4641-90c7-31cd475bbb19" (UID: "1eeec13e-8226-4641-90c7-31cd475bbb19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.370318 4846 scope.go:117] "RemoveContainer" containerID="e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.372979 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eeec13e-8226-4641-90c7-31cd475bbb19-kube-api-access-x9wn4" (OuterVolumeSpecName: "kube-api-access-x9wn4") pod "1eeec13e-8226-4641-90c7-31cd475bbb19" (UID: "1eeec13e-8226-4641-90c7-31cd475bbb19"). InnerVolumeSpecName "kube-api-access-x9wn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:44 crc kubenswrapper[4846]: E0202 12:31:44.373289 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102\": container with ID starting with e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102 not found: ID does not exist" containerID="e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.373316 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102"} err="failed to get container status \"e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102\": rpc error: code = NotFound desc = could not find container \"e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102\": container with ID starting with e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102 not found: ID does not exist" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.373337 4846 scope.go:117] "RemoveContainer" containerID="f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a" Feb 02 12:31:44 crc kubenswrapper[4846]: E0202 12:31:44.373949 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a\": container with ID starting with f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a not found: ID does not exist" containerID="f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.373978 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a"} err="failed to get container status \"f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a\": rpc error: code = NotFound desc = could not find container \"f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a\": container with ID starting with f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a not found: ID does not exist" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.373993 4846 scope.go:117] "RemoveContainer" containerID="aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1" Feb 02 12:31:44 crc kubenswrapper[4846]: E0202 12:31:44.374736 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1\": container with ID starting with aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1 not found: ID does not exist" containerID="aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.374759 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1"} err="failed to get container status \"aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1\": rpc error: code = NotFound desc = could not find container \"aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1\": container with ID starting with aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1 not found: ID does not exist" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.374774 4846 scope.go:117] "RemoveContainer" containerID="777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be" Feb 02 12:31:44 crc kubenswrapper[4846]: E0202 12:31:44.375965 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be\": container with ID starting with 777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be not found: ID does not exist" containerID="777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.376019 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be"} err="failed to get container status \"777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be\": rpc error: code = NotFound desc = could not find container \"777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be\": container with ID starting with 777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be not found: ID does not exist" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.405739 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1eeec13e-8226-4641-90c7-31cd475bbb19" (UID: "1eeec13e-8226-4641-90c7-31cd475bbb19"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.444604 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1eeec13e-8226-4641-90c7-31cd475bbb19" (UID: "1eeec13e-8226-4641-90c7-31cd475bbb19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.465273 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.465311 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9wn4\" (UniqueName: \"kubernetes.io/projected/1eeec13e-8226-4641-90c7-31cd475bbb19-kube-api-access-x9wn4\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.465324 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.465333 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.476713 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-config-data" (OuterVolumeSpecName: "config-data") pod "1eeec13e-8226-4641-90c7-31cd475bbb19" (UID: "1eeec13e-8226-4641-90c7-31cd475bbb19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:44 crc kubenswrapper[4846]: I0202 12:31:44.567527 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1eeec13e-8226-4641-90c7-31cd475bbb19-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.269097 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.301745 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.310801 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.322914 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:45 crc kubenswrapper[4846]: E0202 12:31:45.323289 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="proxy-httpd" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.323306 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="proxy-httpd" Feb 02 12:31:45 crc kubenswrapper[4846]: E0202 12:31:45.323325 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="ceilometer-central-agent" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.323332 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="ceilometer-central-agent" Feb 02 12:31:45 crc kubenswrapper[4846]: E0202 12:31:45.323354 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="sg-core" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.323360 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="sg-core" Feb 02 12:31:45 crc kubenswrapper[4846]: E0202 12:31:45.323379 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="ceilometer-notification-agent" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.323386 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="ceilometer-notification-agent" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.323547 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="proxy-httpd" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.323559 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="sg-core" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.323587 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="ceilometer-central-agent" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.323598 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" containerName="ceilometer-notification-agent" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.325861 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.329179 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.329394 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.329427 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.334817 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.398100 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8479m\" (UniqueName: \"kubernetes.io/projected/1a088a6d-026e-4424-9257-886c929c41f1-kube-api-access-8479m\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.398197 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-run-httpd\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.398259 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.398292 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.398336 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.398390 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-scripts\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.398537 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-log-httpd\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.398593 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-config-data\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.434094 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eeec13e-8226-4641-90c7-31cd475bbb19" path="/var/lib/kubelet/pods/1eeec13e-8226-4641-90c7-31cd475bbb19/volumes" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.499856 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.499913 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.499950 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-scripts\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.500013 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-log-httpd\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.500072 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-config-data\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.500125 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8479m\" (UniqueName: \"kubernetes.io/projected/1a088a6d-026e-4424-9257-886c929c41f1-kube-api-access-8479m\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.500165 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-run-httpd\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.500189 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.501077 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-log-httpd\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.501333 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-run-httpd\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.505586 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-scripts\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.505697 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.506164 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.507382 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-config-data\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.508449 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.517658 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8479m\" (UniqueName: \"kubernetes.io/projected/1a088a6d-026e-4424-9257-886c929c41f1-kube-api-access-8479m\") pod \"ceilometer-0\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " pod="openstack/ceilometer-0" Feb 02 12:31:45 crc kubenswrapper[4846]: I0202 12:31:45.646212 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:46 crc kubenswrapper[4846]: I0202 12:31:46.100965 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:46 crc kubenswrapper[4846]: I0202 12:31:46.278075 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a088a6d-026e-4424-9257-886c929c41f1","Type":"ContainerStarted","Data":"c7bc95c35193a4a713499aef29810751d4b1f376f2e8397993b1a14693e4f0ee"} Feb 02 12:31:46 crc kubenswrapper[4846]: I0202 12:31:46.757276 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 12:31:46 crc kubenswrapper[4846]: I0202 12:31:46.760334 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 12:31:46 crc kubenswrapper[4846]: I0202 12:31:46.773906 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 12:31:47 crc kubenswrapper[4846]: I0202 12:31:47.292080 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a088a6d-026e-4424-9257-886c929c41f1","Type":"ContainerStarted","Data":"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891"} Feb 02 12:31:47 crc kubenswrapper[4846]: I0202 12:31:47.297493 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 12:31:48 crc kubenswrapper[4846]: E0202 12:31:48.180852 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eeec13e_8226_4641_90c7_31cd475bbb19.slice/crio-conmon-f2e6f9f384b2a74361337c3d961e0ca4e94caef222da09d8088214545e6fad1a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eeec13e_8226_4641_90c7_31cd475bbb19.slice/crio-conmon-e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eeec13e_8226_4641_90c7_31cd475bbb19.slice/crio-conmon-aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eeec13e_8226_4641_90c7_31cd475bbb19.slice/crio-e59b31132e29390c34a0278e49509e58a846959cdb56bfd9e5e842d514675102.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eeec13e_8226_4641_90c7_31cd475bbb19.slice/crio-d61b7a402e445c3380a656f4debd05e1df88a2c9a4e1ab1f385837473a1d9282\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eeec13e_8226_4641_90c7_31cd475bbb19.slice/crio-conmon-777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eeec13e_8226_4641_90c7_31cd475bbb19.slice/crio-777ae547d4b770780ca629206ac4f80e48edc232821a9ee3af267bbc241753be.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8c6ee44_38ac_42b1_a961_bf62820300fe.slice/crio-conmon-a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eeec13e_8226_4641_90c7_31cd475bbb19.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8c6ee44_38ac_42b1_a961_bf62820300fe.slice/crio-a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eeec13e_8226_4641_90c7_31cd475bbb19.slice/crio-aa61511fb970c878d280168cb48b196a46d6ef305f2c0afc4977bf1f42a4eef1.scope\": RecentStats: unable to find data in memory cache]" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.296173 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.301552 4846 generic.go:334] "Generic (PLEG): container finished" podID="c8c6ee44-38ac-42b1-a961-bf62820300fe" containerID="a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034" exitCode=137 Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.301665 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.301609 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c8c6ee44-38ac-42b1-a961-bf62820300fe","Type":"ContainerDied","Data":"a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034"} Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.301858 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c8c6ee44-38ac-42b1-a961-bf62820300fe","Type":"ContainerDied","Data":"9a5b3b78631139800fcae9fbb404803a9bddc2db807347f7977949c631ab80fb"} Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.301935 4846 scope.go:117] "RemoveContainer" containerID="a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.310391 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a088a6d-026e-4424-9257-886c929c41f1","Type":"ContainerStarted","Data":"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d"} Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.310450 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a088a6d-026e-4424-9257-886c929c41f1","Type":"ContainerStarted","Data":"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b"} Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.330655 4846 scope.go:117] "RemoveContainer" containerID="a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034" Feb 02 12:31:48 crc kubenswrapper[4846]: E0202 12:31:48.331188 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034\": container with ID starting with a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034 not found: ID does not exist" containerID="a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.332529 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034"} err="failed to get container status \"a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034\": rpc error: code = NotFound desc = could not find container \"a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034\": container with ID starting with a647726dc54528c6bedf61a8bab8a5ff54cfaf3646284927fafc80ab6bb4e034 not found: ID does not exist" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.354331 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjcsm\" (UniqueName: \"kubernetes.io/projected/c8c6ee44-38ac-42b1-a961-bf62820300fe-kube-api-access-cjcsm\") pod \"c8c6ee44-38ac-42b1-a961-bf62820300fe\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.354394 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-config-data\") pod \"c8c6ee44-38ac-42b1-a961-bf62820300fe\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.354421 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-combined-ca-bundle\") pod \"c8c6ee44-38ac-42b1-a961-bf62820300fe\" (UID: \"c8c6ee44-38ac-42b1-a961-bf62820300fe\") " Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.370998 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c6ee44-38ac-42b1-a961-bf62820300fe-kube-api-access-cjcsm" (OuterVolumeSpecName: "kube-api-access-cjcsm") pod "c8c6ee44-38ac-42b1-a961-bf62820300fe" (UID: "c8c6ee44-38ac-42b1-a961-bf62820300fe"). InnerVolumeSpecName "kube-api-access-cjcsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.384930 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8c6ee44-38ac-42b1-a961-bf62820300fe" (UID: "c8c6ee44-38ac-42b1-a961-bf62820300fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.394335 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-config-data" (OuterVolumeSpecName: "config-data") pod "c8c6ee44-38ac-42b1-a961-bf62820300fe" (UID: "c8c6ee44-38ac-42b1-a961-bf62820300fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.455776 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjcsm\" (UniqueName: \"kubernetes.io/projected/c8c6ee44-38ac-42b1-a961-bf62820300fe-kube-api-access-cjcsm\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.456095 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.456106 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c6ee44-38ac-42b1-a961-bf62820300fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.637250 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.649632 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.667899 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:31:48 crc kubenswrapper[4846]: E0202 12:31:48.668336 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c6ee44-38ac-42b1-a961-bf62820300fe" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.668350 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c6ee44-38ac-42b1-a961-bf62820300fe" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.668607 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c6ee44-38ac-42b1-a961-bf62820300fe" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.669511 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.675820 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.675844 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.675990 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.683930 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.762099 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfbtg\" (UniqueName: \"kubernetes.io/projected/fa1b6480-161c-433d-975b-83079f2c979a-kube-api-access-hfbtg\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.762197 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.762262 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.762292 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.762397 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.863328 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.863428 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfbtg\" (UniqueName: \"kubernetes.io/projected/fa1b6480-161c-433d-975b-83079f2c979a-kube-api-access-hfbtg\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.863454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.863493 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.863516 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.870398 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.870561 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.871011 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.871530 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.885316 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfbtg\" (UniqueName: \"kubernetes.io/projected/fa1b6480-161c-433d-975b-83079f2c979a-kube-api-access-hfbtg\") pod \"nova-cell1-novncproxy-0\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:48 crc kubenswrapper[4846]: I0202 12:31:48.992045 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:49 crc kubenswrapper[4846]: I0202 12:31:49.434488 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8c6ee44-38ac-42b1-a961-bf62820300fe" path="/var/lib/kubelet/pods/c8c6ee44-38ac-42b1-a961-bf62820300fe/volumes" Feb 02 12:31:49 crc kubenswrapper[4846]: I0202 12:31:49.494586 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:31:49 crc kubenswrapper[4846]: W0202 12:31:49.497346 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa1b6480_161c_433d_975b_83079f2c979a.slice/crio-aa453eed133ed193ddd2753ec424f485f7ac743d86340f632580643174b56da3 WatchSource:0}: Error finding container aa453eed133ed193ddd2753ec424f485f7ac743d86340f632580643174b56da3: Status 404 returned error can't find the container with id aa453eed133ed193ddd2753ec424f485f7ac743d86340f632580643174b56da3 Feb 02 12:31:50 crc kubenswrapper[4846]: I0202 12:31:50.336598 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fa1b6480-161c-433d-975b-83079f2c979a","Type":"ContainerStarted","Data":"c50d7d730873e682af6be480671b01b55a5ad561b1a9bbf213963fb50c01e8d1"} Feb 02 12:31:50 crc kubenswrapper[4846]: I0202 12:31:50.337288 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fa1b6480-161c-433d-975b-83079f2c979a","Type":"ContainerStarted","Data":"aa453eed133ed193ddd2753ec424f485f7ac743d86340f632580643174b56da3"} Feb 02 12:31:50 crc kubenswrapper[4846]: I0202 12:31:50.339984 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a088a6d-026e-4424-9257-886c929c41f1","Type":"ContainerStarted","Data":"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70"} Feb 02 12:31:50 crc kubenswrapper[4846]: I0202 12:31:50.341168 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 12:31:50 crc kubenswrapper[4846]: I0202 12:31:50.378175 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.378155043 podStartE2EDuration="2.378155043s" podCreationTimestamp="2026-02-02 12:31:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:50.364121554 +0000 UTC m=+1341.592708437" watchObservedRunningTime="2026-02-02 12:31:50.378155043 +0000 UTC m=+1341.606741906" Feb 02 12:31:50 crc kubenswrapper[4846]: I0202 12:31:50.403115 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.410960712 podStartE2EDuration="5.403096773s" podCreationTimestamp="2026-02-02 12:31:45 +0000 UTC" firstStartedPulling="2026-02-02 12:31:46.108413872 +0000 UTC m=+1337.337000735" lastFinishedPulling="2026-02-02 12:31:50.100549933 +0000 UTC m=+1341.329136796" observedRunningTime="2026-02-02 12:31:50.391125945 +0000 UTC m=+1341.619712808" watchObservedRunningTime="2026-02-02 12:31:50.403096773 +0000 UTC m=+1341.631683636" Feb 02 12:31:50 crc kubenswrapper[4846]: I0202 12:31:50.600669 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 02 12:31:51 crc kubenswrapper[4846]: I0202 12:31:51.855663 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 12:31:51 crc kubenswrapper[4846]: I0202 12:31:51.856405 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 12:31:51 crc kubenswrapper[4846]: I0202 12:31:51.860709 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 12:31:51 crc kubenswrapper[4846]: I0202 12:31:51.865016 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.363443 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.366751 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.559925 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-c48q4"] Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.561987 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.576312 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-c48q4"] Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.662367 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-sb\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.662725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-config\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.662750 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-swift-storage-0\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.662783 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-nb\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.662812 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-svc\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.662841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbmtk\" (UniqueName: \"kubernetes.io/projected/b7187614-7a06-41a1-9835-d309ceb1ac9f-kube-api-access-lbmtk\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.764774 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-sb\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.764856 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-config\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.764883 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-swift-storage-0\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.764935 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-nb\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.764965 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-svc\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.765049 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbmtk\" (UniqueName: \"kubernetes.io/projected/b7187614-7a06-41a1-9835-d309ceb1ac9f-kube-api-access-lbmtk\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.765840 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-config\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.765967 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-nb\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.766012 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-svc\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.766137 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-swift-storage-0\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.766610 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-sb\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.793580 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbmtk\" (UniqueName: \"kubernetes.io/projected/b7187614-7a06-41a1-9835-d309ceb1ac9f-kube-api-access-lbmtk\") pod \"dnsmasq-dns-74b99f9475-c48q4\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:52 crc kubenswrapper[4846]: I0202 12:31:52.899427 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:53 crc kubenswrapper[4846]: I0202 12:31:53.452332 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-c48q4"] Feb 02 12:31:53 crc kubenswrapper[4846]: I0202 12:31:53.992596 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:54 crc kubenswrapper[4846]: I0202 12:31:54.383595 4846 generic.go:334] "Generic (PLEG): container finished" podID="b7187614-7a06-41a1-9835-d309ceb1ac9f" containerID="af94415cb893888a9711b1b30d691047d195ad2f337e8d9c9b9a3cdc61e8c9db" exitCode=0 Feb 02 12:31:54 crc kubenswrapper[4846]: I0202 12:31:54.383724 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" event={"ID":"b7187614-7a06-41a1-9835-d309ceb1ac9f","Type":"ContainerDied","Data":"af94415cb893888a9711b1b30d691047d195ad2f337e8d9c9b9a3cdc61e8c9db"} Feb 02 12:31:54 crc kubenswrapper[4846]: I0202 12:31:54.383767 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" event={"ID":"b7187614-7a06-41a1-9835-d309ceb1ac9f","Type":"ContainerStarted","Data":"acb55e106bbee719c2bd0ceb2492ed79d97054ab454e95d754bd7535ef260c97"} Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.288724 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.289242 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="ceilometer-central-agent" containerID="cri-o://df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891" gracePeriod=30 Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.289690 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="ceilometer-notification-agent" containerID="cri-o://b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b" gracePeriod=30 Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.289713 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="sg-core" containerID="cri-o://8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d" gracePeriod=30 Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.289713 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="proxy-httpd" containerID="cri-o://363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70" gracePeriod=30 Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.398645 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" event={"ID":"b7187614-7a06-41a1-9835-d309ceb1ac9f","Type":"ContainerStarted","Data":"e9afbbe94cd0b66d2701c6bfe73f4de7c7680ba854f612982eafd6a0db29d1a5"} Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.399806 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.422278 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" podStartSLOduration=3.4222641879999998 podStartE2EDuration="3.422264188s" podCreationTimestamp="2026-02-02 12:31:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:31:55.419470258 +0000 UTC m=+1346.648057131" watchObservedRunningTime="2026-02-02 12:31:55.422264188 +0000 UTC m=+1346.650851051" Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.580887 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.581291 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerName="nova-api-log" containerID="cri-o://7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad" gracePeriod=30 Feb 02 12:31:55 crc kubenswrapper[4846]: I0202 12:31:55.581366 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerName="nova-api-api" containerID="cri-o://0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71" gracePeriod=30 Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.067223 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.134256 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-config-data\") pod \"1a088a6d-026e-4424-9257-886c929c41f1\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.134726 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-ceilometer-tls-certs\") pod \"1a088a6d-026e-4424-9257-886c929c41f1\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.134761 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-sg-core-conf-yaml\") pod \"1a088a6d-026e-4424-9257-886c929c41f1\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.134853 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-combined-ca-bundle\") pod \"1a088a6d-026e-4424-9257-886c929c41f1\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.134947 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-scripts\") pod \"1a088a6d-026e-4424-9257-886c929c41f1\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.135277 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-log-httpd\") pod \"1a088a6d-026e-4424-9257-886c929c41f1\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.135351 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8479m\" (UniqueName: \"kubernetes.io/projected/1a088a6d-026e-4424-9257-886c929c41f1-kube-api-access-8479m\") pod \"1a088a6d-026e-4424-9257-886c929c41f1\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.135403 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-run-httpd\") pod \"1a088a6d-026e-4424-9257-886c929c41f1\" (UID: \"1a088a6d-026e-4424-9257-886c929c41f1\") " Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.135530 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a088a6d-026e-4424-9257-886c929c41f1" (UID: "1a088a6d-026e-4424-9257-886c929c41f1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.135923 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.136381 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a088a6d-026e-4424-9257-886c929c41f1" (UID: "1a088a6d-026e-4424-9257-886c929c41f1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.158769 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-scripts" (OuterVolumeSpecName: "scripts") pod "1a088a6d-026e-4424-9257-886c929c41f1" (UID: "1a088a6d-026e-4424-9257-886c929c41f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.161779 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a088a6d-026e-4424-9257-886c929c41f1-kube-api-access-8479m" (OuterVolumeSpecName: "kube-api-access-8479m") pod "1a088a6d-026e-4424-9257-886c929c41f1" (UID: "1a088a6d-026e-4424-9257-886c929c41f1"). InnerVolumeSpecName "kube-api-access-8479m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.191300 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1a088a6d-026e-4424-9257-886c929c41f1" (UID: "1a088a6d-026e-4424-9257-886c929c41f1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.225723 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1a088a6d-026e-4424-9257-886c929c41f1" (UID: "1a088a6d-026e-4424-9257-886c929c41f1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.235891 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a088a6d-026e-4424-9257-886c929c41f1" (UID: "1a088a6d-026e-4424-9257-886c929c41f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.237765 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.237801 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.237814 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.237825 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.237852 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8479m\" (UniqueName: \"kubernetes.io/projected/1a088a6d-026e-4424-9257-886c929c41f1-kube-api-access-8479m\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.237864 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a088a6d-026e-4424-9257-886c929c41f1-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.292378 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-config-data" (OuterVolumeSpecName: "config-data") pod "1a088a6d-026e-4424-9257-886c929c41f1" (UID: "1a088a6d-026e-4424-9257-886c929c41f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.340158 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a088a6d-026e-4424-9257-886c929c41f1-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.408703 4846 generic.go:334] "Generic (PLEG): container finished" podID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerID="7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad" exitCode=143 Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.408776 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22b0ef8e-3708-4a24-8be4-f71edc795bc5","Type":"ContainerDied","Data":"7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad"} Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.411878 4846 generic.go:334] "Generic (PLEG): container finished" podID="1a088a6d-026e-4424-9257-886c929c41f1" containerID="363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70" exitCode=0 Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.411901 4846 generic.go:334] "Generic (PLEG): container finished" podID="1a088a6d-026e-4424-9257-886c929c41f1" containerID="8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d" exitCode=2 Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.411913 4846 generic.go:334] "Generic (PLEG): container finished" podID="1a088a6d-026e-4424-9257-886c929c41f1" containerID="b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b" exitCode=0 Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.411919 4846 generic.go:334] "Generic (PLEG): container finished" podID="1a088a6d-026e-4424-9257-886c929c41f1" containerID="df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891" exitCode=0 Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.411948 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.411970 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a088a6d-026e-4424-9257-886c929c41f1","Type":"ContainerDied","Data":"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70"} Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.412016 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a088a6d-026e-4424-9257-886c929c41f1","Type":"ContainerDied","Data":"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d"} Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.412027 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a088a6d-026e-4424-9257-886c929c41f1","Type":"ContainerDied","Data":"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b"} Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.412037 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a088a6d-026e-4424-9257-886c929c41f1","Type":"ContainerDied","Data":"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891"} Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.412045 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a088a6d-026e-4424-9257-886c929c41f1","Type":"ContainerDied","Data":"c7bc95c35193a4a713499aef29810751d4b1f376f2e8397993b1a14693e4f0ee"} Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.412063 4846 scope.go:117] "RemoveContainer" containerID="363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.500483 4846 scope.go:117] "RemoveContainer" containerID="8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.507665 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.524205 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.536765 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:56 crc kubenswrapper[4846]: E0202 12:31:56.537115 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="proxy-httpd" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.537133 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="proxy-httpd" Feb 02 12:31:56 crc kubenswrapper[4846]: E0202 12:31:56.537155 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="ceilometer-notification-agent" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.537161 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="ceilometer-notification-agent" Feb 02 12:31:56 crc kubenswrapper[4846]: E0202 12:31:56.537184 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="sg-core" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.537190 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="sg-core" Feb 02 12:31:56 crc kubenswrapper[4846]: E0202 12:31:56.537203 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="ceilometer-central-agent" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.537210 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="ceilometer-central-agent" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.537373 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="ceilometer-notification-agent" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.537396 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="sg-core" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.537408 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="ceilometer-central-agent" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.537419 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a088a6d-026e-4424-9257-886c929c41f1" containerName="proxy-httpd" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.538869 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.563304 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.564130 4846 scope.go:117] "RemoveContainer" containerID="b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.564356 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.564645 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.572447 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.598459 4846 scope.go:117] "RemoveContainer" containerID="df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.621919 4846 scope.go:117] "RemoveContainer" containerID="363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70" Feb 02 12:31:56 crc kubenswrapper[4846]: E0202 12:31:56.622711 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70\": container with ID starting with 363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70 not found: ID does not exist" containerID="363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.622762 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70"} err="failed to get container status \"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70\": rpc error: code = NotFound desc = could not find container \"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70\": container with ID starting with 363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70 not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.622781 4846 scope.go:117] "RemoveContainer" containerID="8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d" Feb 02 12:31:56 crc kubenswrapper[4846]: E0202 12:31:56.627122 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d\": container with ID starting with 8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d not found: ID does not exist" containerID="8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.627207 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d"} err="failed to get container status \"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d\": rpc error: code = NotFound desc = could not find container \"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d\": container with ID starting with 8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.627239 4846 scope.go:117] "RemoveContainer" containerID="b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b" Feb 02 12:31:56 crc kubenswrapper[4846]: E0202 12:31:56.632373 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b\": container with ID starting with b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b not found: ID does not exist" containerID="b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.633453 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b"} err="failed to get container status \"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b\": rpc error: code = NotFound desc = could not find container \"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b\": container with ID starting with b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.633599 4846 scope.go:117] "RemoveContainer" containerID="df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891" Feb 02 12:31:56 crc kubenswrapper[4846]: E0202 12:31:56.634539 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891\": container with ID starting with df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891 not found: ID does not exist" containerID="df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.634609 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891"} err="failed to get container status \"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891\": rpc error: code = NotFound desc = could not find container \"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891\": container with ID starting with df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891 not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.634654 4846 scope.go:117] "RemoveContainer" containerID="363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.635288 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70"} err="failed to get container status \"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70\": rpc error: code = NotFound desc = could not find container \"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70\": container with ID starting with 363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70 not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.635429 4846 scope.go:117] "RemoveContainer" containerID="8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.635994 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d"} err="failed to get container status \"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d\": rpc error: code = NotFound desc = could not find container \"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d\": container with ID starting with 8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.636047 4846 scope.go:117] "RemoveContainer" containerID="b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.636542 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b"} err="failed to get container status \"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b\": rpc error: code = NotFound desc = could not find container \"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b\": container with ID starting with b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.636572 4846 scope.go:117] "RemoveContainer" containerID="df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.636995 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891"} err="failed to get container status \"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891\": rpc error: code = NotFound desc = could not find container \"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891\": container with ID starting with df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891 not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.637099 4846 scope.go:117] "RemoveContainer" containerID="363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.637727 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70"} err="failed to get container status \"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70\": rpc error: code = NotFound desc = could not find container \"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70\": container with ID starting with 363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70 not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.637752 4846 scope.go:117] "RemoveContainer" containerID="8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.638025 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d"} err="failed to get container status \"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d\": rpc error: code = NotFound desc = could not find container \"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d\": container with ID starting with 8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.638063 4846 scope.go:117] "RemoveContainer" containerID="b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.638281 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b"} err="failed to get container status \"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b\": rpc error: code = NotFound desc = could not find container \"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b\": container with ID starting with b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.638295 4846 scope.go:117] "RemoveContainer" containerID="df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.639201 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891"} err="failed to get container status \"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891\": rpc error: code = NotFound desc = could not find container \"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891\": container with ID starting with df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891 not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.639226 4846 scope.go:117] "RemoveContainer" containerID="363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.639420 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70"} err="failed to get container status \"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70\": rpc error: code = NotFound desc = could not find container \"363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70\": container with ID starting with 363f24e20fccdc48d0a08e78f20c8ca61df7c00334a53650409f50852bc99a70 not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.639437 4846 scope.go:117] "RemoveContainer" containerID="8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.639720 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d"} err="failed to get container status \"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d\": rpc error: code = NotFound desc = could not find container \"8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d\": container with ID starting with 8840d79c6be07d74cb51e3b79d2bac0e98e80d2af8c4d536ab6d97f5e442533d not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.639843 4846 scope.go:117] "RemoveContainer" containerID="b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.640224 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b"} err="failed to get container status \"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b\": rpc error: code = NotFound desc = could not find container \"b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b\": container with ID starting with b6bcf8237f0e6578ef86f05986d551e08387cc569841917e8e17226c1e947f7b not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.640252 4846 scope.go:117] "RemoveContainer" containerID="df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.640509 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891"} err="failed to get container status \"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891\": rpc error: code = NotFound desc = could not find container \"df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891\": container with ID starting with df8a68c2be6cd495ae031b58222abb12abe568d4ad1d4628c6275f1a94728891 not found: ID does not exist" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.668198 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-scripts\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.668325 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.668394 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.668591 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-config-data\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.668608 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-log-httpd\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.668820 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-run-httpd\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.668969 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rpf4\" (UniqueName: \"kubernetes.io/projected/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-kube-api-access-9rpf4\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.669067 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.771166 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rpf4\" (UniqueName: \"kubernetes.io/projected/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-kube-api-access-9rpf4\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.771225 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.771311 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-scripts\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.771409 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.771465 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.771494 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-log-httpd\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.771516 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-config-data\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.771541 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-run-httpd\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.772087 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-run-httpd\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.772143 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-log-httpd\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.787940 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-scripts\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.790690 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-config-data\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.794544 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.799300 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.804167 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.838804 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rpf4\" (UniqueName: \"kubernetes.io/projected/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-kube-api-access-9rpf4\") pod \"ceilometer-0\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " pod="openstack/ceilometer-0" Feb 02 12:31:56 crc kubenswrapper[4846]: I0202 12:31:56.894233 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:31:57 crc kubenswrapper[4846]: I0202 12:31:57.383766 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:57 crc kubenswrapper[4846]: W0202 12:31:57.386456 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95ab55a6_e416_4f23_831e_3fb81fc1aaf5.slice/crio-336b01a63d37bb7addac84e930b878d872c5678e01db8e8420a0e2a934bd4e6a WatchSource:0}: Error finding container 336b01a63d37bb7addac84e930b878d872c5678e01db8e8420a0e2a934bd4e6a: Status 404 returned error can't find the container with id 336b01a63d37bb7addac84e930b878d872c5678e01db8e8420a0e2a934bd4e6a Feb 02 12:31:57 crc kubenswrapper[4846]: I0202 12:31:57.452744 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a088a6d-026e-4424-9257-886c929c41f1" path="/var/lib/kubelet/pods/1a088a6d-026e-4424-9257-886c929c41f1/volumes" Feb 02 12:31:57 crc kubenswrapper[4846]: I0202 12:31:57.453998 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95ab55a6-e416-4f23-831e-3fb81fc1aaf5","Type":"ContainerStarted","Data":"336b01a63d37bb7addac84e930b878d872c5678e01db8e8420a0e2a934bd4e6a"} Feb 02 12:31:57 crc kubenswrapper[4846]: I0202 12:31:57.587546 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:31:58 crc kubenswrapper[4846]: I0202 12:31:58.457259 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95ab55a6-e416-4f23-831e-3fb81fc1aaf5","Type":"ContainerStarted","Data":"a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c"} Feb 02 12:31:58 crc kubenswrapper[4846]: I0202 12:31:58.992603 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.018082 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.218502 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.314492 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-combined-ca-bundle\") pod \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.314714 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22b0ef8e-3708-4a24-8be4-f71edc795bc5-logs\") pod \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.314844 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-config-data\") pod \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.314915 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v66gr\" (UniqueName: \"kubernetes.io/projected/22b0ef8e-3708-4a24-8be4-f71edc795bc5-kube-api-access-v66gr\") pod \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\" (UID: \"22b0ef8e-3708-4a24-8be4-f71edc795bc5\") " Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.316585 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22b0ef8e-3708-4a24-8be4-f71edc795bc5-logs" (OuterVolumeSpecName: "logs") pod "22b0ef8e-3708-4a24-8be4-f71edc795bc5" (UID: "22b0ef8e-3708-4a24-8be4-f71edc795bc5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.320864 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22b0ef8e-3708-4a24-8be4-f71edc795bc5-kube-api-access-v66gr" (OuterVolumeSpecName: "kube-api-access-v66gr") pod "22b0ef8e-3708-4a24-8be4-f71edc795bc5" (UID: "22b0ef8e-3708-4a24-8be4-f71edc795bc5"). InnerVolumeSpecName "kube-api-access-v66gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.356118 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-config-data" (OuterVolumeSpecName: "config-data") pod "22b0ef8e-3708-4a24-8be4-f71edc795bc5" (UID: "22b0ef8e-3708-4a24-8be4-f71edc795bc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.356584 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22b0ef8e-3708-4a24-8be4-f71edc795bc5" (UID: "22b0ef8e-3708-4a24-8be4-f71edc795bc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.417461 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22b0ef8e-3708-4a24-8be4-f71edc795bc5-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.417495 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.417505 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v66gr\" (UniqueName: \"kubernetes.io/projected/22b0ef8e-3708-4a24-8be4-f71edc795bc5-kube-api-access-v66gr\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.417515 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22b0ef8e-3708-4a24-8be4-f71edc795bc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.470141 4846 generic.go:334] "Generic (PLEG): container finished" podID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerID="0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71" exitCode=0 Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.470195 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22b0ef8e-3708-4a24-8be4-f71edc795bc5","Type":"ContainerDied","Data":"0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71"} Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.470218 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22b0ef8e-3708-4a24-8be4-f71edc795bc5","Type":"ContainerDied","Data":"113b72b061ff1e94e3265a6db031c6697f679249f7bbdd04fd338e69f650787f"} Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.470234 4846 scope.go:117] "RemoveContainer" containerID="0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.470310 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.486153 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95ab55a6-e416-4f23-831e-3fb81fc1aaf5","Type":"ContainerStarted","Data":"ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb"} Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.529469 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.546199 4846 scope.go:117] "RemoveContainer" containerID="7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.582899 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.596842 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.600845 4846 scope.go:117] "RemoveContainer" containerID="0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71" Feb 02 12:31:59 crc kubenswrapper[4846]: E0202 12:31:59.601430 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71\": container with ID starting with 0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71 not found: ID does not exist" containerID="0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.601475 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71"} err="failed to get container status \"0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71\": rpc error: code = NotFound desc = could not find container \"0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71\": container with ID starting with 0d7d0aa9eb77f2c18e51f9bd0b5fda0077142d75dcd4d2ca13b3927688c14d71 not found: ID does not exist" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.601504 4846 scope.go:117] "RemoveContainer" containerID="7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad" Feb 02 12:31:59 crc kubenswrapper[4846]: E0202 12:31:59.601850 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad\": container with ID starting with 7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad not found: ID does not exist" containerID="7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.601890 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad"} err="failed to get container status \"7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad\": rpc error: code = NotFound desc = could not find container \"7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad\": container with ID starting with 7708cef3a5679ea22c093aa28e919117206cd6782a0088a395ea8d1dda1b5fad not found: ID does not exist" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.626874 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:59 crc kubenswrapper[4846]: E0202 12:31:59.627990 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerName="nova-api-api" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.628009 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerName="nova-api-api" Feb 02 12:31:59 crc kubenswrapper[4846]: E0202 12:31:59.628042 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerName="nova-api-log" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.628049 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerName="nova-api-log" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.629676 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerName="nova-api-log" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.629741 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" containerName="nova-api-api" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.633652 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.637458 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.637729 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.638558 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.640180 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.748633 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f604b0-a17e-4da4-b12d-32a71bd3c687-logs\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.748748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-config-data\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.748855 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6xtk\" (UniqueName: \"kubernetes.io/projected/a0f604b0-a17e-4da4-b12d-32a71bd3c687-kube-api-access-n6xtk\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.748898 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.748929 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.748958 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-public-tls-certs\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.772996 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-gmb9m"] Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.774612 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.778813 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.779173 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.787803 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gmb9m"] Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.850315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6xtk\" (UniqueName: \"kubernetes.io/projected/a0f604b0-a17e-4da4-b12d-32a71bd3c687-kube-api-access-n6xtk\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.850369 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.850391 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.850409 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.850435 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-public-tls-certs\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.850457 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f604b0-a17e-4da4-b12d-32a71bd3c687-logs\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.850495 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-scripts\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.850529 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-config-data\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.850563 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-config-data\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.850598 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gphf\" (UniqueName: \"kubernetes.io/projected/a7696a86-1c3e-4740-a2ce-ed1a45896127-kube-api-access-8gphf\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.851007 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f604b0-a17e-4da4-b12d-32a71bd3c687-logs\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.859435 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-public-tls-certs\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.861203 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.861749 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.877132 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-config-data\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.878003 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6xtk\" (UniqueName: \"kubernetes.io/projected/a0f604b0-a17e-4da4-b12d-32a71bd3c687-kube-api-access-n6xtk\") pod \"nova-api-0\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " pod="openstack/nova-api-0" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.952134 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gphf\" (UniqueName: \"kubernetes.io/projected/a7696a86-1c3e-4740-a2ce-ed1a45896127-kube-api-access-8gphf\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.952295 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.952421 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-scripts\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.952532 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-config-data\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.956540 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-scripts\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.956738 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-config-data\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.957253 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.973272 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gphf\" (UniqueName: \"kubernetes.io/projected/a7696a86-1c3e-4740-a2ce-ed1a45896127-kube-api-access-8gphf\") pod \"nova-cell1-cell-mapping-gmb9m\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:31:59 crc kubenswrapper[4846]: I0202 12:31:59.998134 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:32:00 crc kubenswrapper[4846]: I0202 12:32:00.123810 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:32:00 crc kubenswrapper[4846]: I0202 12:32:00.502171 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:32:00 crc kubenswrapper[4846]: I0202 12:32:00.534452 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95ab55a6-e416-4f23-831e-3fb81fc1aaf5","Type":"ContainerStarted","Data":"e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0"} Feb 02 12:32:00 crc kubenswrapper[4846]: I0202 12:32:00.610926 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gmb9m"] Feb 02 12:32:00 crc kubenswrapper[4846]: W0202 12:32:00.620853 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7696a86_1c3e_4740_a2ce_ed1a45896127.slice/crio-2424c630d43ce691d372bfb0e516b61cc2c911d5246b4b8ad3856468f4b3bd57 WatchSource:0}: Error finding container 2424c630d43ce691d372bfb0e516b61cc2c911d5246b4b8ad3856468f4b3bd57: Status 404 returned error can't find the container with id 2424c630d43ce691d372bfb0e516b61cc2c911d5246b4b8ad3856468f4b3bd57 Feb 02 12:32:01 crc kubenswrapper[4846]: I0202 12:32:01.434315 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22b0ef8e-3708-4a24-8be4-f71edc795bc5" path="/var/lib/kubelet/pods/22b0ef8e-3708-4a24-8be4-f71edc795bc5/volumes" Feb 02 12:32:01 crc kubenswrapper[4846]: I0202 12:32:01.560979 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0f604b0-a17e-4da4-b12d-32a71bd3c687","Type":"ContainerStarted","Data":"fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332"} Feb 02 12:32:01 crc kubenswrapper[4846]: I0202 12:32:01.562111 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0f604b0-a17e-4da4-b12d-32a71bd3c687","Type":"ContainerStarted","Data":"514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b"} Feb 02 12:32:01 crc kubenswrapper[4846]: I0202 12:32:01.562654 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0f604b0-a17e-4da4-b12d-32a71bd3c687","Type":"ContainerStarted","Data":"bf65f70f5c576003b41f29420eee613805ec9f0ed6aeea5308c682092b1ba70b"} Feb 02 12:32:01 crc kubenswrapper[4846]: I0202 12:32:01.564307 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gmb9m" event={"ID":"a7696a86-1c3e-4740-a2ce-ed1a45896127","Type":"ContainerStarted","Data":"e4a716613f696100470f5f89eb0a597105b79f131836ad346870b04f41dfc79d"} Feb 02 12:32:01 crc kubenswrapper[4846]: I0202 12:32:01.564983 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gmb9m" event={"ID":"a7696a86-1c3e-4740-a2ce-ed1a45896127","Type":"ContainerStarted","Data":"2424c630d43ce691d372bfb0e516b61cc2c911d5246b4b8ad3856468f4b3bd57"} Feb 02 12:32:01 crc kubenswrapper[4846]: I0202 12:32:01.609304 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-gmb9m" podStartSLOduration=2.609284327 podStartE2EDuration="2.609284327s" podCreationTimestamp="2026-02-02 12:31:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:32:01.606231971 +0000 UTC m=+1352.834818834" watchObservedRunningTime="2026-02-02 12:32:01.609284327 +0000 UTC m=+1352.837871190" Feb 02 12:32:01 crc kubenswrapper[4846]: I0202 12:32:01.611260 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.611251706 podStartE2EDuration="2.611251706s" podCreationTimestamp="2026-02-02 12:31:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:32:01.589343961 +0000 UTC m=+1352.817930824" watchObservedRunningTime="2026-02-02 12:32:01.611251706 +0000 UTC m=+1352.839838569" Feb 02 12:32:02 crc kubenswrapper[4846]: I0202 12:32:02.900793 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:32:02 crc kubenswrapper[4846]: I0202 12:32:02.962506 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-2wdnq"] Feb 02 12:32:02 crc kubenswrapper[4846]: I0202 12:32:02.962735 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" podUID="a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" containerName="dnsmasq-dns" containerID="cri-o://5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680" gracePeriod=10 Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.585199 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.586571 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95ab55a6-e416-4f23-831e-3fb81fc1aaf5","Type":"ContainerStarted","Data":"d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1"} Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.586864 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="ceilometer-central-agent" containerID="cri-o://a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c" gracePeriod=30 Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.586878 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="sg-core" containerID="cri-o://e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0" gracePeriod=30 Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.586892 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="proxy-httpd" containerID="cri-o://d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1" gracePeriod=30 Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.586915 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="ceilometer-notification-agent" containerID="cri-o://ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb" gracePeriod=30 Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.586944 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.593144 4846 generic.go:334] "Generic (PLEG): container finished" podID="a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" containerID="5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680" exitCode=0 Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.593215 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.593225 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" event={"ID":"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93","Type":"ContainerDied","Data":"5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680"} Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.593299 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-868bc9dc59-2wdnq" event={"ID":"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93","Type":"ContainerDied","Data":"3a7ec4dbe23cb3a13b82504f0419ba04b33412d3410803ce690b398bdfea047f"} Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.593325 4846 scope.go:117] "RemoveContainer" containerID="5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.630261 4846 scope.go:117] "RemoveContainer" containerID="fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.630819 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5c4p\" (UniqueName: \"kubernetes.io/projected/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-kube-api-access-v5c4p\") pod \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.630904 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-config\") pod \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.630925 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-sb\") pod \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.631102 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-swift-storage-0\") pod \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.631124 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-nb\") pod \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.631143 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-svc\") pod \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\" (UID: \"a87c7f72-7e31-4e8c-a3dc-70c574a6dc93\") " Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.645690 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-kube-api-access-v5c4p" (OuterVolumeSpecName: "kube-api-access-v5c4p") pod "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" (UID: "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93"). InnerVolumeSpecName "kube-api-access-v5c4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.661818 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.588573057 podStartE2EDuration="7.661796485s" podCreationTimestamp="2026-02-02 12:31:56 +0000 UTC" firstStartedPulling="2026-02-02 12:31:57.3903904 +0000 UTC m=+1348.618977263" lastFinishedPulling="2026-02-02 12:32:02.463613818 +0000 UTC m=+1353.692200691" observedRunningTime="2026-02-02 12:32:03.633106703 +0000 UTC m=+1354.861693576" watchObservedRunningTime="2026-02-02 12:32:03.661796485 +0000 UTC m=+1354.890383348" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.732898 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5c4p\" (UniqueName: \"kubernetes.io/projected/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-kube-api-access-v5c4p\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.733848 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" (UID: "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.734114 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-config" (OuterVolumeSpecName: "config") pod "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" (UID: "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.747450 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" (UID: "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.748384 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" (UID: "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.756465 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" (UID: "a87c7f72-7e31-4e8c-a3dc-70c574a6dc93"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.835188 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.835234 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.835248 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.835260 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.835272 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.838106 4846 scope.go:117] "RemoveContainer" containerID="5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680" Feb 02 12:32:03 crc kubenswrapper[4846]: E0202 12:32:03.838680 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680\": container with ID starting with 5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680 not found: ID does not exist" containerID="5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.838739 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680"} err="failed to get container status \"5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680\": rpc error: code = NotFound desc = could not find container \"5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680\": container with ID starting with 5cb755eed93bd5f08c614a1f94a64a99655261b53e6026d5f1f7a2810d015680 not found: ID does not exist" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.838779 4846 scope.go:117] "RemoveContainer" containerID="fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31" Feb 02 12:32:03 crc kubenswrapper[4846]: E0202 12:32:03.839107 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31\": container with ID starting with fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31 not found: ID does not exist" containerID="fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.839139 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31"} err="failed to get container status \"fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31\": rpc error: code = NotFound desc = could not find container \"fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31\": container with ID starting with fd5c1ba7096af267f1f1460f8d0bbb99ec2ae2e29b4f29784eccba9fbdfcee31 not found: ID does not exist" Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.937458 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-2wdnq"] Feb 02 12:32:03 crc kubenswrapper[4846]: I0202 12:32:03.952147 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-868bc9dc59-2wdnq"] Feb 02 12:32:04 crc kubenswrapper[4846]: I0202 12:32:04.606386 4846 generic.go:334] "Generic (PLEG): container finished" podID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerID="d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1" exitCode=0 Feb 02 12:32:04 crc kubenswrapper[4846]: I0202 12:32:04.606727 4846 generic.go:334] "Generic (PLEG): container finished" podID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerID="e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0" exitCode=2 Feb 02 12:32:04 crc kubenswrapper[4846]: I0202 12:32:04.606743 4846 generic.go:334] "Generic (PLEG): container finished" podID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerID="ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb" exitCode=0 Feb 02 12:32:04 crc kubenswrapper[4846]: I0202 12:32:04.606454 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95ab55a6-e416-4f23-831e-3fb81fc1aaf5","Type":"ContainerDied","Data":"d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1"} Feb 02 12:32:04 crc kubenswrapper[4846]: I0202 12:32:04.606780 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95ab55a6-e416-4f23-831e-3fb81fc1aaf5","Type":"ContainerDied","Data":"e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0"} Feb 02 12:32:04 crc kubenswrapper[4846]: I0202 12:32:04.606795 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95ab55a6-e416-4f23-831e-3fb81fc1aaf5","Type":"ContainerDied","Data":"ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb"} Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.116242 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.159134 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-config-data\") pod \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.159239 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-combined-ca-bundle\") pod \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.159302 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-sg-core-conf-yaml\") pod \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.159513 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-log-httpd\") pod \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.159609 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rpf4\" (UniqueName: \"kubernetes.io/projected/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-kube-api-access-9rpf4\") pod \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.159724 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-ceilometer-tls-certs\") pod \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.159850 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-run-httpd\") pod \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.159899 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-scripts\") pod \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.160099 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "95ab55a6-e416-4f23-831e-3fb81fc1aaf5" (UID: "95ab55a6-e416-4f23-831e-3fb81fc1aaf5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.160250 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "95ab55a6-e416-4f23-831e-3fb81fc1aaf5" (UID: "95ab55a6-e416-4f23-831e-3fb81fc1aaf5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.160806 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.160843 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.164969 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-kube-api-access-9rpf4" (OuterVolumeSpecName: "kube-api-access-9rpf4") pod "95ab55a6-e416-4f23-831e-3fb81fc1aaf5" (UID: "95ab55a6-e416-4f23-831e-3fb81fc1aaf5"). InnerVolumeSpecName "kube-api-access-9rpf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.169724 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-scripts" (OuterVolumeSpecName: "scripts") pod "95ab55a6-e416-4f23-831e-3fb81fc1aaf5" (UID: "95ab55a6-e416-4f23-831e-3fb81fc1aaf5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.189817 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "95ab55a6-e416-4f23-831e-3fb81fc1aaf5" (UID: "95ab55a6-e416-4f23-831e-3fb81fc1aaf5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.233015 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "95ab55a6-e416-4f23-831e-3fb81fc1aaf5" (UID: "95ab55a6-e416-4f23-831e-3fb81fc1aaf5"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.261686 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95ab55a6-e416-4f23-831e-3fb81fc1aaf5" (UID: "95ab55a6-e416-4f23-831e-3fb81fc1aaf5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.262056 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-combined-ca-bundle\") pod \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\" (UID: \"95ab55a6-e416-4f23-831e-3fb81fc1aaf5\") " Feb 02 12:32:05 crc kubenswrapper[4846]: W0202 12:32:05.262172 4846 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/95ab55a6-e416-4f23-831e-3fb81fc1aaf5/volumes/kubernetes.io~secret/combined-ca-bundle Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.262190 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95ab55a6-e416-4f23-831e-3fb81fc1aaf5" (UID: "95ab55a6-e416-4f23-831e-3fb81fc1aaf5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.262819 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.262842 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.262853 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rpf4\" (UniqueName: \"kubernetes.io/projected/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-kube-api-access-9rpf4\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.262864 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.262873 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.265647 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-config-data" (OuterVolumeSpecName: "config-data") pod "95ab55a6-e416-4f23-831e-3fb81fc1aaf5" (UID: "95ab55a6-e416-4f23-831e-3fb81fc1aaf5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.364754 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95ab55a6-e416-4f23-831e-3fb81fc1aaf5-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.439272 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" path="/var/lib/kubelet/pods/a87c7f72-7e31-4e8c-a3dc-70c574a6dc93/volumes" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.623580 4846 generic.go:334] "Generic (PLEG): container finished" podID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerID="a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c" exitCode=0 Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.623938 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95ab55a6-e416-4f23-831e-3fb81fc1aaf5","Type":"ContainerDied","Data":"a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c"} Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.623965 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"95ab55a6-e416-4f23-831e-3fb81fc1aaf5","Type":"ContainerDied","Data":"336b01a63d37bb7addac84e930b878d872c5678e01db8e8420a0e2a934bd4e6a"} Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.623982 4846 scope.go:117] "RemoveContainer" containerID="d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.624094 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.650853 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.678040 4846 scope.go:117] "RemoveContainer" containerID="e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.701674 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.715369 4846 scope.go:117] "RemoveContainer" containerID="ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.727181 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:32:05 crc kubenswrapper[4846]: E0202 12:32:05.727811 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" containerName="init" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.727832 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" containerName="init" Feb 02 12:32:05 crc kubenswrapper[4846]: E0202 12:32:05.727884 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="ceilometer-notification-agent" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.727892 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="ceilometer-notification-agent" Feb 02 12:32:05 crc kubenswrapper[4846]: E0202 12:32:05.727910 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="ceilometer-central-agent" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.727916 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="ceilometer-central-agent" Feb 02 12:32:05 crc kubenswrapper[4846]: E0202 12:32:05.727925 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="sg-core" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.727931 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="sg-core" Feb 02 12:32:05 crc kubenswrapper[4846]: E0202 12:32:05.727946 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="proxy-httpd" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.727951 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="proxy-httpd" Feb 02 12:32:05 crc kubenswrapper[4846]: E0202 12:32:05.727960 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" containerName="dnsmasq-dns" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.727966 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" containerName="dnsmasq-dns" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.728133 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="ceilometer-central-agent" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.728150 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="sg-core" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.728167 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="ceilometer-notification-agent" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.728179 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" containerName="proxy-httpd" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.728196 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87c7f72-7e31-4e8c-a3dc-70c574a6dc93" containerName="dnsmasq-dns" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.729782 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.735898 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.736144 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.738563 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.738773 4846 scope.go:117] "RemoveContainer" containerID="a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.743051 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.768068 4846 scope.go:117] "RemoveContainer" containerID="d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1" Feb 02 12:32:05 crc kubenswrapper[4846]: E0202 12:32:05.768476 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1\": container with ID starting with d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1 not found: ID does not exist" containerID="d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.768517 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1"} err="failed to get container status \"d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1\": rpc error: code = NotFound desc = could not find container \"d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1\": container with ID starting with d0302ed87418eafe2d50f313cad68035ea68b4a0945b30d716ced3ae84891af1 not found: ID does not exist" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.768543 4846 scope.go:117] "RemoveContainer" containerID="e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0" Feb 02 12:32:05 crc kubenswrapper[4846]: E0202 12:32:05.769013 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0\": container with ID starting with e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0 not found: ID does not exist" containerID="e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.769032 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0"} err="failed to get container status \"e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0\": rpc error: code = NotFound desc = could not find container \"e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0\": container with ID starting with e1fe8462d636976e024cf0681b3f58805bd4fae3e1d0ba7851395432f83fb2b0 not found: ID does not exist" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.769045 4846 scope.go:117] "RemoveContainer" containerID="ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb" Feb 02 12:32:05 crc kubenswrapper[4846]: E0202 12:32:05.769266 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb\": container with ID starting with ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb not found: ID does not exist" containerID="ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.769374 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb"} err="failed to get container status \"ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb\": rpc error: code = NotFound desc = could not find container \"ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb\": container with ID starting with ec5d02877d0912bb1d8b7ab44685b887e13f186b04e6e03925fe640bb38b80bb not found: ID does not exist" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.769473 4846 scope.go:117] "RemoveContainer" containerID="a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c" Feb 02 12:32:05 crc kubenswrapper[4846]: E0202 12:32:05.769867 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c\": container with ID starting with a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c not found: ID does not exist" containerID="a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.769922 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c"} err="failed to get container status \"a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c\": rpc error: code = NotFound desc = could not find container \"a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c\": container with ID starting with a943470a59cc98b99b0619915a61ee1845412f8f7ec8c3761aec115a7c02ce1c not found: ID does not exist" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.777089 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.777130 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.777206 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-scripts\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.777233 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-run-httpd\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.777247 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-log-httpd\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.777272 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-config-data\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.777346 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.777488 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnhmc\" (UniqueName: \"kubernetes.io/projected/92819684-ecb5-4c4d-950c-edacf38e269b-kube-api-access-xnhmc\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.879244 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.879310 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.879407 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-scripts\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.879444 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-run-httpd\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.879466 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-log-httpd\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.879498 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-config-data\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.879522 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.879570 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnhmc\" (UniqueName: \"kubernetes.io/projected/92819684-ecb5-4c4d-950c-edacf38e269b-kube-api-access-xnhmc\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.880934 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-run-httpd\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.881035 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-log-httpd\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.897812 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.897819 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.899079 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-scripts\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.899900 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-config-data\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.900121 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnhmc\" (UniqueName: \"kubernetes.io/projected/92819684-ecb5-4c4d-950c-edacf38e269b-kube-api-access-xnhmc\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:05 crc kubenswrapper[4846]: I0202 12:32:05.907373 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " pod="openstack/ceilometer-0" Feb 02 12:32:06 crc kubenswrapper[4846]: I0202 12:32:06.061769 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:32:06 crc kubenswrapper[4846]: I0202 12:32:06.509759 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:32:06 crc kubenswrapper[4846]: I0202 12:32:06.632537 4846 generic.go:334] "Generic (PLEG): container finished" podID="a7696a86-1c3e-4740-a2ce-ed1a45896127" containerID="e4a716613f696100470f5f89eb0a597105b79f131836ad346870b04f41dfc79d" exitCode=0 Feb 02 12:32:06 crc kubenswrapper[4846]: I0202 12:32:06.632608 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gmb9m" event={"ID":"a7696a86-1c3e-4740-a2ce-ed1a45896127","Type":"ContainerDied","Data":"e4a716613f696100470f5f89eb0a597105b79f131836ad346870b04f41dfc79d"} Feb 02 12:32:06 crc kubenswrapper[4846]: I0202 12:32:06.634007 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92819684-ecb5-4c4d-950c-edacf38e269b","Type":"ContainerStarted","Data":"7812dfc323a65a36cf25f3ba6d08f1f304a1dc41c215d04527f26995552b4c29"} Feb 02 12:32:07 crc kubenswrapper[4846]: I0202 12:32:07.434220 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95ab55a6-e416-4f23-831e-3fb81fc1aaf5" path="/var/lib/kubelet/pods/95ab55a6-e416-4f23-831e-3fb81fc1aaf5/volumes" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.031925 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.118727 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gphf\" (UniqueName: \"kubernetes.io/projected/a7696a86-1c3e-4740-a2ce-ed1a45896127-kube-api-access-8gphf\") pod \"a7696a86-1c3e-4740-a2ce-ed1a45896127\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.118853 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-config-data\") pod \"a7696a86-1c3e-4740-a2ce-ed1a45896127\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.118968 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-combined-ca-bundle\") pod \"a7696a86-1c3e-4740-a2ce-ed1a45896127\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.118998 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-scripts\") pod \"a7696a86-1c3e-4740-a2ce-ed1a45896127\" (UID: \"a7696a86-1c3e-4740-a2ce-ed1a45896127\") " Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.126811 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-scripts" (OuterVolumeSpecName: "scripts") pod "a7696a86-1c3e-4740-a2ce-ed1a45896127" (UID: "a7696a86-1c3e-4740-a2ce-ed1a45896127"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.126811 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7696a86-1c3e-4740-a2ce-ed1a45896127-kube-api-access-8gphf" (OuterVolumeSpecName: "kube-api-access-8gphf") pod "a7696a86-1c3e-4740-a2ce-ed1a45896127" (UID: "a7696a86-1c3e-4740-a2ce-ed1a45896127"). InnerVolumeSpecName "kube-api-access-8gphf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.159076 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7696a86-1c3e-4740-a2ce-ed1a45896127" (UID: "a7696a86-1c3e-4740-a2ce-ed1a45896127"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.159666 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-config-data" (OuterVolumeSpecName: "config-data") pod "a7696a86-1c3e-4740-a2ce-ed1a45896127" (UID: "a7696a86-1c3e-4740-a2ce-ed1a45896127"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.221124 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.221172 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.221184 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gphf\" (UniqueName: \"kubernetes.io/projected/a7696a86-1c3e-4740-a2ce-ed1a45896127-kube-api-access-8gphf\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.221197 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7696a86-1c3e-4740-a2ce-ed1a45896127-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.655437 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gmb9m" event={"ID":"a7696a86-1c3e-4740-a2ce-ed1a45896127","Type":"ContainerDied","Data":"2424c630d43ce691d372bfb0e516b61cc2c911d5246b4b8ad3856468f4b3bd57"} Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.655488 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2424c630d43ce691d372bfb0e516b61cc2c911d5246b4b8ad3856468f4b3bd57" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.656937 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92819684-ecb5-4c4d-950c-edacf38e269b","Type":"ContainerStarted","Data":"e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea"} Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.657014 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gmb9m" Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.852715 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.853126 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" containerName="nova-api-log" containerID="cri-o://514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b" gracePeriod=30 Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.853487 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" containerName="nova-api-api" containerID="cri-o://fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332" gracePeriod=30 Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.896782 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.897402 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c3207510-5457-46af-b2ea-05b33efc61fb" containerName="nova-scheduler-scheduler" containerID="cri-o://42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c" gracePeriod=30 Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.916059 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.916336 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-log" containerID="cri-o://ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b" gracePeriod=30 Feb 02 12:32:08 crc kubenswrapper[4846]: I0202 12:32:08.916500 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-metadata" containerID="cri-o://7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021" gracePeriod=30 Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.385006 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.479578 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f604b0-a17e-4da4-b12d-32a71bd3c687-logs\") pod \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.480366 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6xtk\" (UniqueName: \"kubernetes.io/projected/a0f604b0-a17e-4da4-b12d-32a71bd3c687-kube-api-access-n6xtk\") pod \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.480169 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f604b0-a17e-4da4-b12d-32a71bd3c687-logs" (OuterVolumeSpecName: "logs") pod "a0f604b0-a17e-4da4-b12d-32a71bd3c687" (UID: "a0f604b0-a17e-4da4-b12d-32a71bd3c687"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.481044 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-config-data\") pod \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.481192 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-combined-ca-bundle\") pod \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.481277 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-internal-tls-certs\") pod \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.481306 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-public-tls-certs\") pod \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\" (UID: \"a0f604b0-a17e-4da4-b12d-32a71bd3c687\") " Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.482179 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0f604b0-a17e-4da4-b12d-32a71bd3c687-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.485688 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f604b0-a17e-4da4-b12d-32a71bd3c687-kube-api-access-n6xtk" (OuterVolumeSpecName: "kube-api-access-n6xtk") pod "a0f604b0-a17e-4da4-b12d-32a71bd3c687" (UID: "a0f604b0-a17e-4da4-b12d-32a71bd3c687"). InnerVolumeSpecName "kube-api-access-n6xtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.505776 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0f604b0-a17e-4da4-b12d-32a71bd3c687" (UID: "a0f604b0-a17e-4da4-b12d-32a71bd3c687"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.506206 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-config-data" (OuterVolumeSpecName: "config-data") pod "a0f604b0-a17e-4da4-b12d-32a71bd3c687" (UID: "a0f604b0-a17e-4da4-b12d-32a71bd3c687"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.535766 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a0f604b0-a17e-4da4-b12d-32a71bd3c687" (UID: "a0f604b0-a17e-4da4-b12d-32a71bd3c687"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.536594 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a0f604b0-a17e-4da4-b12d-32a71bd3c687" (UID: "a0f604b0-a17e-4da4-b12d-32a71bd3c687"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.584442 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6xtk\" (UniqueName: \"kubernetes.io/projected/a0f604b0-a17e-4da4-b12d-32a71bd3c687-kube-api-access-n6xtk\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.584476 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.584485 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.584493 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.584502 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0f604b0-a17e-4da4-b12d-32a71bd3c687-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.665265 4846 generic.go:334] "Generic (PLEG): container finished" podID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerID="ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b" exitCode=143 Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.665349 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"053e0de3-f1a2-451d-9f01-8f26cc67fd92","Type":"ContainerDied","Data":"ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b"} Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.666852 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92819684-ecb5-4c4d-950c-edacf38e269b","Type":"ContainerStarted","Data":"b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f"} Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.668542 4846 generic.go:334] "Generic (PLEG): container finished" podID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" containerID="fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332" exitCode=0 Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.668703 4846 generic.go:334] "Generic (PLEG): container finished" podID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" containerID="514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b" exitCode=143 Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.668644 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0f604b0-a17e-4da4-b12d-32a71bd3c687","Type":"ContainerDied","Data":"fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332"} Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.668932 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0f604b0-a17e-4da4-b12d-32a71bd3c687","Type":"ContainerDied","Data":"514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b"} Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.669001 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a0f604b0-a17e-4da4-b12d-32a71bd3c687","Type":"ContainerDied","Data":"bf65f70f5c576003b41f29420eee613805ec9f0ed6aeea5308c682092b1ba70b"} Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.669079 4846 scope.go:117] "RemoveContainer" containerID="fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.668604 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.697243 4846 scope.go:117] "RemoveContainer" containerID="514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.705670 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.716690 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.734737 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 12:32:09 crc kubenswrapper[4846]: E0202 12:32:09.735189 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" containerName="nova-api-api" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.735211 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" containerName="nova-api-api" Feb 02 12:32:09 crc kubenswrapper[4846]: E0202 12:32:09.735223 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" containerName="nova-api-log" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.735231 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" containerName="nova-api-log" Feb 02 12:32:09 crc kubenswrapper[4846]: E0202 12:32:09.735259 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7696a86-1c3e-4740-a2ce-ed1a45896127" containerName="nova-manage" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.735265 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7696a86-1c3e-4740-a2ce-ed1a45896127" containerName="nova-manage" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.735448 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7696a86-1c3e-4740-a2ce-ed1a45896127" containerName="nova-manage" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.735467 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" containerName="nova-api-api" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.735488 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" containerName="nova-api-log" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.736528 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.740794 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.740894 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.741429 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.750471 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.770068 4846 scope.go:117] "RemoveContainer" containerID="fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332" Feb 02 12:32:09 crc kubenswrapper[4846]: E0202 12:32:09.773993 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332\": container with ID starting with fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332 not found: ID does not exist" containerID="fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.774043 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332"} err="failed to get container status \"fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332\": rpc error: code = NotFound desc = could not find container \"fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332\": container with ID starting with fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332 not found: ID does not exist" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.774077 4846 scope.go:117] "RemoveContainer" containerID="514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b" Feb 02 12:32:09 crc kubenswrapper[4846]: E0202 12:32:09.774522 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b\": container with ID starting with 514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b not found: ID does not exist" containerID="514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.774559 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b"} err="failed to get container status \"514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b\": rpc error: code = NotFound desc = could not find container \"514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b\": container with ID starting with 514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b not found: ID does not exist" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.774584 4846 scope.go:117] "RemoveContainer" containerID="fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.774987 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332"} err="failed to get container status \"fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332\": rpc error: code = NotFound desc = could not find container \"fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332\": container with ID starting with fac234f11bf7491532d4701b44030d4db6c58a5cedd2f48679948a6df535b332 not found: ID does not exist" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.775006 4846 scope.go:117] "RemoveContainer" containerID="514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.775933 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b"} err="failed to get container status \"514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b\": rpc error: code = NotFound desc = could not find container \"514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b\": container with ID starting with 514f44106cce00e83d9a8ec8b1a6937efe37fd2bdf508e8ccb45a370bab3636b not found: ID does not exist" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.892352 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-config-data\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.892432 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.892453 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-public-tls-certs\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.892513 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c452f22-eef8-4e05-a1dd-f6c841833ab4-logs\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.892559 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.892577 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkjz7\" (UniqueName: \"kubernetes.io/projected/3c452f22-eef8-4e05-a1dd-f6c841833ab4-kube-api-access-xkjz7\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.993776 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.993817 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkjz7\" (UniqueName: \"kubernetes.io/projected/3c452f22-eef8-4e05-a1dd-f6c841833ab4-kube-api-access-xkjz7\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.993893 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-config-data\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.993964 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.993985 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-public-tls-certs\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.994071 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c452f22-eef8-4e05-a1dd-f6c841833ab4-logs\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.994436 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c452f22-eef8-4e05-a1dd-f6c841833ab4-logs\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.998460 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.999333 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:09 crc kubenswrapper[4846]: I0202 12:32:09.999341 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-config-data\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:10 crc kubenswrapper[4846]: I0202 12:32:10.010545 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-public-tls-certs\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:10 crc kubenswrapper[4846]: I0202 12:32:10.015321 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkjz7\" (UniqueName: \"kubernetes.io/projected/3c452f22-eef8-4e05-a1dd-f6c841833ab4-kube-api-access-xkjz7\") pod \"nova-api-0\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " pod="openstack/nova-api-0" Feb 02 12:32:10 crc kubenswrapper[4846]: I0202 12:32:10.076416 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:32:10 crc kubenswrapper[4846]: E0202 12:32:10.499744 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 12:32:10 crc kubenswrapper[4846]: E0202 12:32:10.503846 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 12:32:10 crc kubenswrapper[4846]: E0202 12:32:10.505471 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 12:32:10 crc kubenswrapper[4846]: E0202 12:32:10.505511 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c3207510-5457-46af-b2ea-05b33efc61fb" containerName="nova-scheduler-scheduler" Feb 02 12:32:10 crc kubenswrapper[4846]: I0202 12:32:10.544103 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:32:10 crc kubenswrapper[4846]: W0202 12:32:10.544771 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c452f22_eef8_4e05_a1dd_f6c841833ab4.slice/crio-a0d977765916383faa5992a6fbbed67a8ead03c2da665089b66deda768013551 WatchSource:0}: Error finding container a0d977765916383faa5992a6fbbed67a8ead03c2da665089b66deda768013551: Status 404 returned error can't find the container with id a0d977765916383faa5992a6fbbed67a8ead03c2da665089b66deda768013551 Feb 02 12:32:10 crc kubenswrapper[4846]: I0202 12:32:10.679252 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92819684-ecb5-4c4d-950c-edacf38e269b","Type":"ContainerStarted","Data":"821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91"} Feb 02 12:32:10 crc kubenswrapper[4846]: I0202 12:32:10.680439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c452f22-eef8-4e05-a1dd-f6c841833ab4","Type":"ContainerStarted","Data":"a0d977765916383faa5992a6fbbed67a8ead03c2da665089b66deda768013551"} Feb 02 12:32:11 crc kubenswrapper[4846]: I0202 12:32:11.435459 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f604b0-a17e-4da4-b12d-32a71bd3c687" path="/var/lib/kubelet/pods/a0f604b0-a17e-4da4-b12d-32a71bd3c687/volumes" Feb 02 12:32:11 crc kubenswrapper[4846]: I0202 12:32:11.729852 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c452f22-eef8-4e05-a1dd-f6c841833ab4","Type":"ContainerStarted","Data":"99b0a8947603d649bb9d1f3decf69b9ee0004e7f9c373151604124884a5d8445"} Feb 02 12:32:11 crc kubenswrapper[4846]: I0202 12:32:11.729905 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c452f22-eef8-4e05-a1dd-f6c841833ab4","Type":"ContainerStarted","Data":"ceb10096d86d9be3fc00e39352a6fcbaf81ecf7bd346d3e24a99c69fe60cd9f2"} Feb 02 12:32:11 crc kubenswrapper[4846]: I0202 12:32:11.753554 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.753538029 podStartE2EDuration="2.753538029s" podCreationTimestamp="2026-02-02 12:32:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:32:11.749088639 +0000 UTC m=+1362.977675512" watchObservedRunningTime="2026-02-02 12:32:11.753538029 +0000 UTC m=+1362.982124892" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.078900 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": read tcp 10.217.0.2:58862->10.217.0.196:8775: read: connection reset by peer" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.078964 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": read tcp 10.217.0.2:58852->10.217.0.196:8775: read: connection reset by peer" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.561583 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.652755 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/053e0de3-f1a2-451d-9f01-8f26cc67fd92-logs\") pod \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.653249 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-config-data\") pod \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.653328 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-nova-metadata-tls-certs\") pod \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.653444 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-combined-ca-bundle\") pod \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.653508 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hwzn\" (UniqueName: \"kubernetes.io/projected/053e0de3-f1a2-451d-9f01-8f26cc67fd92-kube-api-access-8hwzn\") pod \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\" (UID: \"053e0de3-f1a2-451d-9f01-8f26cc67fd92\") " Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.654247 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/053e0de3-f1a2-451d-9f01-8f26cc67fd92-logs" (OuterVolumeSpecName: "logs") pod "053e0de3-f1a2-451d-9f01-8f26cc67fd92" (UID: "053e0de3-f1a2-451d-9f01-8f26cc67fd92"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.661056 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/053e0de3-f1a2-451d-9f01-8f26cc67fd92-kube-api-access-8hwzn" (OuterVolumeSpecName: "kube-api-access-8hwzn") pod "053e0de3-f1a2-451d-9f01-8f26cc67fd92" (UID: "053e0de3-f1a2-451d-9f01-8f26cc67fd92"). InnerVolumeSpecName "kube-api-access-8hwzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.691000 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-config-data" (OuterVolumeSpecName: "config-data") pod "053e0de3-f1a2-451d-9f01-8f26cc67fd92" (UID: "053e0de3-f1a2-451d-9f01-8f26cc67fd92"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.694749 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "053e0de3-f1a2-451d-9f01-8f26cc67fd92" (UID: "053e0de3-f1a2-451d-9f01-8f26cc67fd92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.718469 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "053e0de3-f1a2-451d-9f01-8f26cc67fd92" (UID: "053e0de3-f1a2-451d-9f01-8f26cc67fd92"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.741552 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92819684-ecb5-4c4d-950c-edacf38e269b","Type":"ContainerStarted","Data":"c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e"} Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.741682 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.743934 4846 generic.go:334] "Generic (PLEG): container finished" podID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerID="7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021" exitCode=0 Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.744456 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"053e0de3-f1a2-451d-9f01-8f26cc67fd92","Type":"ContainerDied","Data":"7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021"} Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.744530 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"053e0de3-f1a2-451d-9f01-8f26cc67fd92","Type":"ContainerDied","Data":"9eed17d9c7dd77001d75ff7544c20c5736721d23ee18f3567c3f2b1307ffe755"} Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.744464 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.744549 4846 scope.go:117] "RemoveContainer" containerID="7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.755848 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.755912 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hwzn\" (UniqueName: \"kubernetes.io/projected/053e0de3-f1a2-451d-9f01-8f26cc67fd92-kube-api-access-8hwzn\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.755925 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/053e0de3-f1a2-451d-9f01-8f26cc67fd92-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.755935 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.755943 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/053e0de3-f1a2-451d-9f01-8f26cc67fd92-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.768261 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.345329461 podStartE2EDuration="7.768245478s" podCreationTimestamp="2026-02-02 12:32:05 +0000 UTC" firstStartedPulling="2026-02-02 12:32:06.527573825 +0000 UTC m=+1357.756160698" lastFinishedPulling="2026-02-02 12:32:11.950489842 +0000 UTC m=+1363.179076715" observedRunningTime="2026-02-02 12:32:12.766821203 +0000 UTC m=+1363.995408086" watchObservedRunningTime="2026-02-02 12:32:12.768245478 +0000 UTC m=+1363.996832341" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.776658 4846 scope.go:117] "RemoveContainer" containerID="ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.788393 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.796852 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.799272 4846 scope.go:117] "RemoveContainer" containerID="7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021" Feb 02 12:32:12 crc kubenswrapper[4846]: E0202 12:32:12.799788 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021\": container with ID starting with 7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021 not found: ID does not exist" containerID="7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.799823 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021"} err="failed to get container status \"7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021\": rpc error: code = NotFound desc = could not find container \"7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021\": container with ID starting with 7f91ae991d96d5855b3881b422bd31673aa58308653fdc45ca20abda60525021 not found: ID does not exist" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.799849 4846 scope.go:117] "RemoveContainer" containerID="ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b" Feb 02 12:32:12 crc kubenswrapper[4846]: E0202 12:32:12.800216 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b\": container with ID starting with ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b not found: ID does not exist" containerID="ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.800255 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b"} err="failed to get container status \"ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b\": rpc error: code = NotFound desc = could not find container \"ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b\": container with ID starting with ed2f8df3b5f8d10eb4bcc2de2e8a2c53980df46be135c278d4a6a63772622f9b not found: ID does not exist" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.814646 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:32:12 crc kubenswrapper[4846]: E0202 12:32:12.815255 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-log" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.815361 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-log" Feb 02 12:32:12 crc kubenswrapper[4846]: E0202 12:32:12.815428 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-metadata" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.815482 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-metadata" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.815717 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-metadata" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.815784 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" containerName="nova-metadata-log" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.816749 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.823782 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.824119 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.849809 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.960258 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.960331 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhbpb\" (UniqueName: \"kubernetes.io/projected/5e2d348d-dbbf-4046-b9fb-94f8aab64573-kube-api-access-xhbpb\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.960383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.960416 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e2d348d-dbbf-4046-b9fb-94f8aab64573-logs\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:12 crc kubenswrapper[4846]: I0202 12:32:12.960456 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-config-data\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.062395 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.062441 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e2d348d-dbbf-4046-b9fb-94f8aab64573-logs\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.062486 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-config-data\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.062569 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.062598 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhbpb\" (UniqueName: \"kubernetes.io/projected/5e2d348d-dbbf-4046-b9fb-94f8aab64573-kube-api-access-xhbpb\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.062971 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e2d348d-dbbf-4046-b9fb-94f8aab64573-logs\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.065937 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.066307 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.067291 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-config-data\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.078637 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhbpb\" (UniqueName: \"kubernetes.io/projected/5e2d348d-dbbf-4046-b9fb-94f8aab64573-kube-api-access-xhbpb\") pod \"nova-metadata-0\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.131197 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.441644 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="053e0de3-f1a2-451d-9f01-8f26cc67fd92" path="/var/lib/kubelet/pods/053e0de3-f1a2-451d-9f01-8f26cc67fd92/volumes" Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.606879 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:32:13 crc kubenswrapper[4846]: W0202 12:32:13.609181 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e2d348d_dbbf_4046_b9fb_94f8aab64573.slice/crio-9e37fc284b289789211d3d4abe18eeeb82d5df4b59f640e5b8743b4bd5f0b315 WatchSource:0}: Error finding container 9e37fc284b289789211d3d4abe18eeeb82d5df4b59f640e5b8743b4bd5f0b315: Status 404 returned error can't find the container with id 9e37fc284b289789211d3d4abe18eeeb82d5df4b59f640e5b8743b4bd5f0b315 Feb 02 12:32:13 crc kubenswrapper[4846]: I0202 12:32:13.754291 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e2d348d-dbbf-4046-b9fb-94f8aab64573","Type":"ContainerStarted","Data":"9e37fc284b289789211d3d4abe18eeeb82d5df4b59f640e5b8743b4bd5f0b315"} Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.576066 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.703373 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-combined-ca-bundle\") pod \"c3207510-5457-46af-b2ea-05b33efc61fb\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.703489 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-config-data\") pod \"c3207510-5457-46af-b2ea-05b33efc61fb\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.703683 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkd8q\" (UniqueName: \"kubernetes.io/projected/c3207510-5457-46af-b2ea-05b33efc61fb-kube-api-access-fkd8q\") pod \"c3207510-5457-46af-b2ea-05b33efc61fb\" (UID: \"c3207510-5457-46af-b2ea-05b33efc61fb\") " Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.706971 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3207510-5457-46af-b2ea-05b33efc61fb-kube-api-access-fkd8q" (OuterVolumeSpecName: "kube-api-access-fkd8q") pod "c3207510-5457-46af-b2ea-05b33efc61fb" (UID: "c3207510-5457-46af-b2ea-05b33efc61fb"). InnerVolumeSpecName "kube-api-access-fkd8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.749155 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-config-data" (OuterVolumeSpecName: "config-data") pod "c3207510-5457-46af-b2ea-05b33efc61fb" (UID: "c3207510-5457-46af-b2ea-05b33efc61fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.755778 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3207510-5457-46af-b2ea-05b33efc61fb" (UID: "c3207510-5457-46af-b2ea-05b33efc61fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.769223 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e2d348d-dbbf-4046-b9fb-94f8aab64573","Type":"ContainerStarted","Data":"ba2153f30138c7492692991e0000a0e9232db386da8169c4c360432f8ac0b9a2"} Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.769262 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e2d348d-dbbf-4046-b9fb-94f8aab64573","Type":"ContainerStarted","Data":"66365bb3d3091805961a6da62564f6a7bf8e51dc20054684c80e03009a5d18dd"} Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.781945 4846 generic.go:334] "Generic (PLEG): container finished" podID="c3207510-5457-46af-b2ea-05b33efc61fb" containerID="42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c" exitCode=0 Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.782016 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c3207510-5457-46af-b2ea-05b33efc61fb","Type":"ContainerDied","Data":"42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c"} Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.782040 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c3207510-5457-46af-b2ea-05b33efc61fb","Type":"ContainerDied","Data":"ad144c609948fc4b457c6a8013010d9fe6baa30f993db447fcefd81150db17fb"} Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.782056 4846 scope.go:117] "RemoveContainer" containerID="42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.782165 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.809226 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkd8q\" (UniqueName: \"kubernetes.io/projected/c3207510-5457-46af-b2ea-05b33efc61fb-kube-api-access-fkd8q\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.809253 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.809264 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3207510-5457-46af-b2ea-05b33efc61fb-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.815430 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.8154096859999997 podStartE2EDuration="2.815409686s" podCreationTimestamp="2026-02-02 12:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:32:14.793025687 +0000 UTC m=+1366.021612570" watchObservedRunningTime="2026-02-02 12:32:14.815409686 +0000 UTC m=+1366.043996549" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.835350 4846 scope.go:117] "RemoveContainer" containerID="42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c" Feb 02 12:32:14 crc kubenswrapper[4846]: E0202 12:32:14.849199 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c\": container with ID starting with 42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c not found: ID does not exist" containerID="42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.849333 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c"} err="failed to get container status \"42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c\": rpc error: code = NotFound desc = could not find container \"42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c\": container with ID starting with 42800273a9456f6de46c47c01f00183a28d1e1ac3f64871c32cc1217165a898c not found: ID does not exist" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.878525 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.888137 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.896944 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:32:14 crc kubenswrapper[4846]: E0202 12:32:14.897399 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3207510-5457-46af-b2ea-05b33efc61fb" containerName="nova-scheduler-scheduler" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.897420 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3207510-5457-46af-b2ea-05b33efc61fb" containerName="nova-scheduler-scheduler" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.897887 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3207510-5457-46af-b2ea-05b33efc61fb" containerName="nova-scheduler-scheduler" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.898738 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.903927 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 02 12:32:14 crc kubenswrapper[4846]: I0202 12:32:14.905424 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.012661 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " pod="openstack/nova-scheduler-0" Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.013020 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-config-data\") pod \"nova-scheduler-0\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " pod="openstack/nova-scheduler-0" Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.013211 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7prsb\" (UniqueName: \"kubernetes.io/projected/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-kube-api-access-7prsb\") pod \"nova-scheduler-0\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " pod="openstack/nova-scheduler-0" Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.115093 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-config-data\") pod \"nova-scheduler-0\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " pod="openstack/nova-scheduler-0" Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.115239 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7prsb\" (UniqueName: \"kubernetes.io/projected/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-kube-api-access-7prsb\") pod \"nova-scheduler-0\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " pod="openstack/nova-scheduler-0" Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.115265 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " pod="openstack/nova-scheduler-0" Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.123237 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-config-data\") pod \"nova-scheduler-0\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " pod="openstack/nova-scheduler-0" Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.123297 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " pod="openstack/nova-scheduler-0" Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.138018 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7prsb\" (UniqueName: \"kubernetes.io/projected/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-kube-api-access-7prsb\") pod \"nova-scheduler-0\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " pod="openstack/nova-scheduler-0" Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.216406 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.435424 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3207510-5457-46af-b2ea-05b33efc61fb" path="/var/lib/kubelet/pods/c3207510-5457-46af-b2ea-05b33efc61fb/volumes" Feb 02 12:32:15 crc kubenswrapper[4846]: W0202 12:32:15.642824 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f25197c_8c4e_41ef_b7ef_1ed0f141599b.slice/crio-585dc97a1f45b5b4dabb82cd33a5932a469364646b448d0b8364bf69db252fa3 WatchSource:0}: Error finding container 585dc97a1f45b5b4dabb82cd33a5932a469364646b448d0b8364bf69db252fa3: Status 404 returned error can't find the container with id 585dc97a1f45b5b4dabb82cd33a5932a469364646b448d0b8364bf69db252fa3 Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.645854 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:32:15 crc kubenswrapper[4846]: I0202 12:32:15.795771 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7f25197c-8c4e-41ef-b7ef-1ed0f141599b","Type":"ContainerStarted","Data":"585dc97a1f45b5b4dabb82cd33a5932a469364646b448d0b8364bf69db252fa3"} Feb 02 12:32:16 crc kubenswrapper[4846]: I0202 12:32:16.805221 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7f25197c-8c4e-41ef-b7ef-1ed0f141599b","Type":"ContainerStarted","Data":"24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8"} Feb 02 12:32:16 crc kubenswrapper[4846]: I0202 12:32:16.827096 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.827073584 podStartE2EDuration="2.827073584s" podCreationTimestamp="2026-02-02 12:32:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 12:32:16.820955004 +0000 UTC m=+1368.049541887" watchObservedRunningTime="2026-02-02 12:32:16.827073584 +0000 UTC m=+1368.055660447" Feb 02 12:32:18 crc kubenswrapper[4846]: I0202 12:32:18.131832 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 12:32:18 crc kubenswrapper[4846]: I0202 12:32:18.131890 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 12:32:20 crc kubenswrapper[4846]: I0202 12:32:20.077335 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 12:32:20 crc kubenswrapper[4846]: I0202 12:32:20.077706 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 12:32:20 crc kubenswrapper[4846]: I0202 12:32:20.217515 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 02 12:32:21 crc kubenswrapper[4846]: I0202 12:32:21.092868 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 12:32:21 crc kubenswrapper[4846]: I0202 12:32:21.092887 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 12:32:23 crc kubenswrapper[4846]: I0202 12:32:23.131733 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 12:32:23 crc kubenswrapper[4846]: I0202 12:32:23.132118 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 12:32:24 crc kubenswrapper[4846]: I0202 12:32:24.151818 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 12:32:24 crc kubenswrapper[4846]: I0202 12:32:24.151818 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 12:32:25 crc kubenswrapper[4846]: I0202 12:32:25.217100 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 02 12:32:25 crc kubenswrapper[4846]: I0202 12:32:25.245680 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 02 12:32:25 crc kubenswrapper[4846]: I0202 12:32:25.956125 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 02 12:32:30 crc kubenswrapper[4846]: I0202 12:32:30.089397 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 12:32:30 crc kubenswrapper[4846]: I0202 12:32:30.091751 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 12:32:30 crc kubenswrapper[4846]: I0202 12:32:30.099652 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 12:32:30 crc kubenswrapper[4846]: I0202 12:32:30.103688 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 12:32:30 crc kubenswrapper[4846]: I0202 12:32:30.940347 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 12:32:30 crc kubenswrapper[4846]: I0202 12:32:30.946215 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 12:32:33 crc kubenswrapper[4846]: I0202 12:32:33.135996 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 12:32:33 crc kubenswrapper[4846]: I0202 12:32:33.137297 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 12:32:33 crc kubenswrapper[4846]: I0202 12:32:33.142310 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 12:32:33 crc kubenswrapper[4846]: I0202 12:32:33.144895 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 12:32:36 crc kubenswrapper[4846]: I0202 12:32:36.072892 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.345015 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c2bc-account-create-update-2r5pb"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.358564 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c2bc-account-create-update-2r5pb"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.366683 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c2bc-account-create-update-qz9kb"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.368191 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c2bc-account-create-update-qz9kb" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.375056 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.381718 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c2bc-account-create-update-qz9kb"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.447841 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-f94a-account-create-update-zg8j8"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.449096 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f94a-account-create-update-zg8j8" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.458005 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdgzg\" (UniqueName: \"kubernetes.io/projected/3915f1bc-20d3-4e7d-b627-46406d0d0461-kube-api-access-bdgzg\") pod \"barbican-c2bc-account-create-update-qz9kb\" (UID: \"3915f1bc-20d3-4e7d-b627-46406d0d0461\") " pod="openstack/barbican-c2bc-account-create-update-qz9kb" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.458037 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3915f1bc-20d3-4e7d-b627-46406d0d0461-operator-scripts\") pod \"barbican-c2bc-account-create-update-qz9kb\" (UID: \"3915f1bc-20d3-4e7d-b627-46406d0d0461\") " pod="openstack/barbican-c2bc-account-create-update-qz9kb" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.464179 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.470361 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f94a-account-create-update-zg8j8"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.487371 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-rr425"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.488833 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rr425" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.497550 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.518929 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-33ff-account-create-update-vz5ls"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.520085 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-33ff-account-create-update-vz5ls" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.533436 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.558957 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rr425"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.561056 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm67h\" (UniqueName: \"kubernetes.io/projected/71584556-0f8c-4781-bbd6-2e3f6913f0e1-kube-api-access-bm67h\") pod \"root-account-create-update-rr425\" (UID: \"71584556-0f8c-4781-bbd6-2e3f6913f0e1\") " pod="openstack/root-account-create-update-rr425" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.561130 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdgzg\" (UniqueName: \"kubernetes.io/projected/3915f1bc-20d3-4e7d-b627-46406d0d0461-kube-api-access-bdgzg\") pod \"barbican-c2bc-account-create-update-qz9kb\" (UID: \"3915f1bc-20d3-4e7d-b627-46406d0d0461\") " pod="openstack/barbican-c2bc-account-create-update-qz9kb" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.561154 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3915f1bc-20d3-4e7d-b627-46406d0d0461-operator-scripts\") pod \"barbican-c2bc-account-create-update-qz9kb\" (UID: \"3915f1bc-20d3-4e7d-b627-46406d0d0461\") " pod="openstack/barbican-c2bc-account-create-update-qz9kb" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.561210 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71584556-0f8c-4781-bbd6-2e3f6913f0e1-operator-scripts\") pod \"root-account-create-update-rr425\" (UID: \"71584556-0f8c-4781-bbd6-2e3f6913f0e1\") " pod="openstack/root-account-create-update-rr425" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.561282 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2sw2\" (UniqueName: \"kubernetes.io/projected/b6f74db8-a48c-4358-81d0-41391a207008-kube-api-access-q2sw2\") pod \"glance-f94a-account-create-update-zg8j8\" (UID: \"b6f74db8-a48c-4358-81d0-41391a207008\") " pod="openstack/glance-f94a-account-create-update-zg8j8" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.561317 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f74db8-a48c-4358-81d0-41391a207008-operator-scripts\") pod \"glance-f94a-account-create-update-zg8j8\" (UID: \"b6f74db8-a48c-4358-81d0-41391a207008\") " pod="openstack/glance-f94a-account-create-update-zg8j8" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.563466 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3915f1bc-20d3-4e7d-b627-46406d0d0461-operator-scripts\") pod \"barbican-c2bc-account-create-update-qz9kb\" (UID: \"3915f1bc-20d3-4e7d-b627-46406d0d0461\") " pod="openstack/barbican-c2bc-account-create-update-qz9kb" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.571696 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f94a-account-create-update-vwrmd"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.601871 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-f94a-account-create-update-vwrmd"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.635749 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-33ff-account-create-update-vz5ls"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.645575 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdgzg\" (UniqueName: \"kubernetes.io/projected/3915f1bc-20d3-4e7d-b627-46406d0d0461-kube-api-access-bdgzg\") pod \"barbican-c2bc-account-create-update-qz9kb\" (UID: \"3915f1bc-20d3-4e7d-b627-46406d0d0461\") " pod="openstack/barbican-c2bc-account-create-update-qz9kb" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.667815 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2sw2\" (UniqueName: \"kubernetes.io/projected/b6f74db8-a48c-4358-81d0-41391a207008-kube-api-access-q2sw2\") pod \"glance-f94a-account-create-update-zg8j8\" (UID: \"b6f74db8-a48c-4358-81d0-41391a207008\") " pod="openstack/glance-f94a-account-create-update-zg8j8" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.667873 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/318e1583-4759-4158-9683-14524ad771c2-operator-scripts\") pod \"neutron-33ff-account-create-update-vz5ls\" (UID: \"318e1583-4759-4158-9683-14524ad771c2\") " pod="openstack/neutron-33ff-account-create-update-vz5ls" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.667895 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f74db8-a48c-4358-81d0-41391a207008-operator-scripts\") pod \"glance-f94a-account-create-update-zg8j8\" (UID: \"b6f74db8-a48c-4358-81d0-41391a207008\") " pod="openstack/glance-f94a-account-create-update-zg8j8" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.667941 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm67h\" (UniqueName: \"kubernetes.io/projected/71584556-0f8c-4781-bbd6-2e3f6913f0e1-kube-api-access-bm67h\") pod \"root-account-create-update-rr425\" (UID: \"71584556-0f8c-4781-bbd6-2e3f6913f0e1\") " pod="openstack/root-account-create-update-rr425" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.667983 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44x6w\" (UniqueName: \"kubernetes.io/projected/318e1583-4759-4158-9683-14524ad771c2-kube-api-access-44x6w\") pod \"neutron-33ff-account-create-update-vz5ls\" (UID: \"318e1583-4759-4158-9683-14524ad771c2\") " pod="openstack/neutron-33ff-account-create-update-vz5ls" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.668030 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71584556-0f8c-4781-bbd6-2e3f6913f0e1-operator-scripts\") pod \"root-account-create-update-rr425\" (UID: \"71584556-0f8c-4781-bbd6-2e3f6913f0e1\") " pod="openstack/root-account-create-update-rr425" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.668733 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71584556-0f8c-4781-bbd6-2e3f6913f0e1-operator-scripts\") pod \"root-account-create-update-rr425\" (UID: \"71584556-0f8c-4781-bbd6-2e3f6913f0e1\") " pod="openstack/root-account-create-update-rr425" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.669465 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f74db8-a48c-4358-81d0-41391a207008-operator-scripts\") pod \"glance-f94a-account-create-update-zg8j8\" (UID: \"b6f74db8-a48c-4358-81d0-41391a207008\") " pod="openstack/glance-f94a-account-create-update-zg8j8" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.682686 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-cd5d-account-create-update-v79s9"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.683853 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cd5d-account-create-update-v79s9" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.686984 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c2bc-account-create-update-qz9kb" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.687817 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.718174 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-33ff-account-create-update-mpf9l"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.768695 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2sw2\" (UniqueName: \"kubernetes.io/projected/b6f74db8-a48c-4358-81d0-41391a207008-kube-api-access-q2sw2\") pod \"glance-f94a-account-create-update-zg8j8\" (UID: \"b6f74db8-a48c-4358-81d0-41391a207008\") " pod="openstack/glance-f94a-account-create-update-zg8j8" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.781604 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/318e1583-4759-4158-9683-14524ad771c2-operator-scripts\") pod \"neutron-33ff-account-create-update-vz5ls\" (UID: \"318e1583-4759-4158-9683-14524ad771c2\") " pod="openstack/neutron-33ff-account-create-update-vz5ls" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.781696 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gswh5\" (UniqueName: \"kubernetes.io/projected/db660713-b11b-416f-9fb6-075facb7ecf0-kube-api-access-gswh5\") pod \"cinder-cd5d-account-create-update-v79s9\" (UID: \"db660713-b11b-416f-9fb6-075facb7ecf0\") " pod="openstack/cinder-cd5d-account-create-update-v79s9" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.781837 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44x6w\" (UniqueName: \"kubernetes.io/projected/318e1583-4759-4158-9683-14524ad771c2-kube-api-access-44x6w\") pod \"neutron-33ff-account-create-update-vz5ls\" (UID: \"318e1583-4759-4158-9683-14524ad771c2\") " pod="openstack/neutron-33ff-account-create-update-vz5ls" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.781988 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db660713-b11b-416f-9fb6-075facb7ecf0-operator-scripts\") pod \"cinder-cd5d-account-create-update-v79s9\" (UID: \"db660713-b11b-416f-9fb6-075facb7ecf0\") " pod="openstack/cinder-cd5d-account-create-update-v79s9" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.787127 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm67h\" (UniqueName: \"kubernetes.io/projected/71584556-0f8c-4781-bbd6-2e3f6913f0e1-kube-api-access-bm67h\") pod \"root-account-create-update-rr425\" (UID: \"71584556-0f8c-4781-bbd6-2e3f6913f0e1\") " pod="openstack/root-account-create-update-rr425" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.880199 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-33ff-account-create-update-mpf9l"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.886170 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f94a-account-create-update-zg8j8" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.886955 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/318e1583-4759-4158-9683-14524ad771c2-operator-scripts\") pod \"neutron-33ff-account-create-update-vz5ls\" (UID: \"318e1583-4759-4158-9683-14524ad771c2\") " pod="openstack/neutron-33ff-account-create-update-vz5ls" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.893228 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rr425" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.895890 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gswh5\" (UniqueName: \"kubernetes.io/projected/db660713-b11b-416f-9fb6-075facb7ecf0-kube-api-access-gswh5\") pod \"cinder-cd5d-account-create-update-v79s9\" (UID: \"db660713-b11b-416f-9fb6-075facb7ecf0\") " pod="openstack/cinder-cd5d-account-create-update-v79s9" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.937197 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a0f2-account-create-update-g95vn"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.938735 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0f2-account-create-update-g95vn" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.969114 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.969649 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cd5d-account-create-update-v79s9"] Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.998034 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24671593-6e2f-44b9-909f-66b7a794c372-operator-scripts\") pod \"placement-a0f2-account-create-update-g95vn\" (UID: \"24671593-6e2f-44b9-909f-66b7a794c372\") " pod="openstack/placement-a0f2-account-create-update-g95vn" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.998099 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m88rs\" (UniqueName: \"kubernetes.io/projected/24671593-6e2f-44b9-909f-66b7a794c372-kube-api-access-m88rs\") pod \"placement-a0f2-account-create-update-g95vn\" (UID: \"24671593-6e2f-44b9-909f-66b7a794c372\") " pod="openstack/placement-a0f2-account-create-update-g95vn" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.998104 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gswh5\" (UniqueName: \"kubernetes.io/projected/db660713-b11b-416f-9fb6-075facb7ecf0-kube-api-access-gswh5\") pod \"cinder-cd5d-account-create-update-v79s9\" (UID: \"db660713-b11b-416f-9fb6-075facb7ecf0\") " pod="openstack/cinder-cd5d-account-create-update-v79s9" Feb 02 12:32:54 crc kubenswrapper[4846]: I0202 12:32:54.998245 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db660713-b11b-416f-9fb6-075facb7ecf0-operator-scripts\") pod \"cinder-cd5d-account-create-update-v79s9\" (UID: \"db660713-b11b-416f-9fb6-075facb7ecf0\") " pod="openstack/cinder-cd5d-account-create-update-v79s9" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:54.999093 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db660713-b11b-416f-9fb6-075facb7ecf0-operator-scripts\") pod \"cinder-cd5d-account-create-update-v79s9\" (UID: \"db660713-b11b-416f-9fb6-075facb7ecf0\") " pod="openstack/cinder-cd5d-account-create-update-v79s9" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.018245 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44x6w\" (UniqueName: \"kubernetes.io/projected/318e1583-4759-4158-9683-14524ad771c2-kube-api-access-44x6w\") pod \"neutron-33ff-account-create-update-vz5ls\" (UID: \"318e1583-4759-4158-9683-14524ad771c2\") " pod="openstack/neutron-33ff-account-create-update-vz5ls" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.018764 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a0f2-account-create-update-g95vn"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.063692 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.102828 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24671593-6e2f-44b9-909f-66b7a794c372-operator-scripts\") pod \"placement-a0f2-account-create-update-g95vn\" (UID: \"24671593-6e2f-44b9-909f-66b7a794c372\") " pod="openstack/placement-a0f2-account-create-update-g95vn" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.102900 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m88rs\" (UniqueName: \"kubernetes.io/projected/24671593-6e2f-44b9-909f-66b7a794c372-kube-api-access-m88rs\") pod \"placement-a0f2-account-create-update-g95vn\" (UID: \"24671593-6e2f-44b9-909f-66b7a794c372\") " pod="openstack/placement-a0f2-account-create-update-g95vn" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.103958 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24671593-6e2f-44b9-909f-66b7a794c372-operator-scripts\") pod \"placement-a0f2-account-create-update-g95vn\" (UID: \"24671593-6e2f-44b9-909f-66b7a794c372\") " pod="openstack/placement-a0f2-account-create-update-g95vn" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.147821 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.148392 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="ba822b20-88d5-404b-850c-7f9441933baf" containerName="openstackclient" containerID="cri-o://b4630774f6b8174e6eaa2407dc38ee14e1e53c6776c23ce799141cd61d5f6c49" gracePeriod=2 Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.173424 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m88rs\" (UniqueName: \"kubernetes.io/projected/24671593-6e2f-44b9-909f-66b7a794c372-kube-api-access-m88rs\") pod \"placement-a0f2-account-create-update-g95vn\" (UID: \"24671593-6e2f-44b9-909f-66b7a794c372\") " pod="openstack/placement-a0f2-account-create-update-g95vn" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.184181 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.184345 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-33ff-account-create-update-vz5ls" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.225079 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cd5d-account-create-update-v79s9" Feb 02 12:32:55 crc kubenswrapper[4846]: E0202 12:32:55.225672 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Feb 02 12:32:55 crc kubenswrapper[4846]: E0202 12:32:55.225717 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data podName:327c8143-b3ed-4431-a206-d2019ddcb606 nodeName:}" failed. No retries permitted until 2026-02-02 12:32:55.725698933 +0000 UTC m=+1406.954285796 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data") pod "rabbitmq-cell1-server-0" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606") : configmap "rabbitmq-cell1-config-data" not found Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.230224 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-c26kj"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.286039 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-c26kj"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.354529 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-b2hcc"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.388755 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-nlrm4"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.389033 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-nlrm4" podUID="fd01ba51-8a4c-4b74-a7ab-382600f93e68" containerName="openstack-network-exporter" containerID="cri-o://0448b20de6b4b9dfd9b17913d840a0f18245b4ede9e0f0f9853c064551787006" gracePeriod=30 Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.391193 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0f2-account-create-update-g95vn" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.415690 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-cd5d-account-create-update-7xmpx"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.504170 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93eff246-ccb3-472c-bf51-f8ca43a5c4a2" path="/var/lib/kubelet/pods/93eff246-ccb3-472c-bf51-f8ca43a5c4a2/volumes" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.504916 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a488c453-b188-4d57-a136-0a002cd342bb" path="/var/lib/kubelet/pods/a488c453-b188-4d57-a136-0a002cd342bb/volumes" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.505430 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f" path="/var/lib/kubelet/pods/ae8af97a-a4b6-4f5a-b2a4-f1863fabb04f/volumes" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.506226 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1" path="/var/lib/kubelet/pods/c39fb0e2-eb96-4f4e-a8d2-a268b0d8b4f1/volumes" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.507220 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-cd5d-account-create-update-7xmpx"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.507252 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-798b-account-create-update-7g5s7"] Feb 02 12:32:55 crc kubenswrapper[4846]: E0202 12:32:55.507556 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba822b20-88d5-404b-850c-7f9441933baf" containerName="openstackclient" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.507572 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba822b20-88d5-404b-850c-7f9441933baf" containerName="openstackclient" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.516832 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba822b20-88d5-404b-850c-7f9441933baf" containerName="openstackclient" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.517595 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-798b-account-create-update-7g5s7" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.523756 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-798b-account-create-update-7g5s7"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.540598 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.544428 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-m79vg"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.563387 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a0f2-account-create-update-pmqhw"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.583689 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a0f2-account-create-update-pmqhw"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.609421 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-lwh8p"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.622664 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-operator-scripts\") pod \"nova-api-798b-account-create-update-7g5s7\" (UID: \"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb\") " pod="openstack/nova-api-798b-account-create-update-7g5s7" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.622792 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d4jf\" (UniqueName: \"kubernetes.io/projected/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-kube-api-access-7d4jf\") pod \"nova-api-798b-account-create-update-7g5s7\" (UID: \"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb\") " pod="openstack/nova-api-798b-account-create-update-7g5s7" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.644432 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.644829 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerName="ovn-northd" containerID="cri-o://a60702e8aab62dc62e3d8fd855c2cb15d42dc1d569a8a5f56df75bb229c8672e" gracePeriod=30 Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.645465 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerName="openstack-network-exporter" containerID="cri-o://a6d2a885c9e46febf0a5624587138a27dc292d388c2adfac2d4f55ce324f5505" gracePeriod=30 Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.671930 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-lwh8p"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.713534 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-5vmgj"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.724092 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d4jf\" (UniqueName: \"kubernetes.io/projected/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-kube-api-access-7d4jf\") pod \"nova-api-798b-account-create-update-7g5s7\" (UID: \"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb\") " pod="openstack/nova-api-798b-account-create-update-7g5s7" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.724241 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-operator-scripts\") pod \"nova-api-798b-account-create-update-7g5s7\" (UID: \"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb\") " pod="openstack/nova-api-798b-account-create-update-7g5s7" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.725229 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-operator-scripts\") pod \"nova-api-798b-account-create-update-7g5s7\" (UID: \"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb\") " pod="openstack/nova-api-798b-account-create-update-7g5s7" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.731769 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-kd7sk"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.756715 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-5vmgj"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.766676 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-kd7sk"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.779972 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d4jf\" (UniqueName: \"kubernetes.io/projected/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-kube-api-access-7d4jf\") pod \"nova-api-798b-account-create-update-7g5s7\" (UID: \"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb\") " pod="openstack/nova-api-798b-account-create-update-7g5s7" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.782683 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-798b-account-create-update-gjnrv"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.792524 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-798b-account-create-update-gjnrv"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.814158 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-2z995"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.830867 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-2z995"] Feb 02 12:32:55 crc kubenswrapper[4846]: E0202 12:32:55.832968 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Feb 02 12:32:55 crc kubenswrapper[4846]: E0202 12:32:55.833028 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data podName:327c8143-b3ed-4431-a206-d2019ddcb606 nodeName:}" failed. No retries permitted until 2026-02-02 12:32:56.833013513 +0000 UTC m=+1408.061600376 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data") pod "rabbitmq-cell1-server-0" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606") : configmap "rabbitmq-cell1-config-data" not found Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.864414 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-798b-account-create-update-7g5s7" Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.883397 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-s2hlh"] Feb 02 12:32:55 crc kubenswrapper[4846]: I0202 12:32:55.970207 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-858jm"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.055588 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-c48q4"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.055872 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" podUID="b7187614-7a06-41a1-9835-d309ceb1ac9f" containerName="dnsmasq-dns" containerID="cri-o://e9afbbe94cd0b66d2701c6bfe73f4de7c7680ba854f612982eafd6a0db29d1a5" gracePeriod=10 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.083893 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-s2hlh"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.098574 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-858jm"] Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.132407 4846 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-b2hcc" message="Exiting ovn-controller (1) " Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.132454 4846 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-b2hcc" podUID="bda23288-2fee-45ad-81e2-a14654c16589" containerName="ovn-controller" containerID="cri-o://ebe246dd795f71a1ad40ab382816989fc7965fac63ac05fe50ffa37d6e3182a0" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.132487 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-b2hcc" podUID="bda23288-2fee-45ad-81e2-a14654c16589" containerName="ovn-controller" containerID="cri-o://ebe246dd795f71a1ad40ab382816989fc7965fac63ac05fe50ffa37d6e3182a0" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.142281 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.161994 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.162831 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerName="openstack-network-exporter" containerID="cri-o://8cd3e31565aad9a3a479cdc1ab4236b7a0085ec750392646a09a6087a0894db0" gracePeriod=300 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.188311 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.190256 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="36061d30-6265-47ca-8a87-5275f41d03ed" containerName="openstack-network-exporter" containerID="cri-o://0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341" gracePeriod=300 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.231088 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-f9gxt"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.232411 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.236100 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.251735 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.251796 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data podName:9cec353a-92f9-4ed4-8ec9-15271cbc6912 nodeName:}" failed. No retries permitted until 2026-02-02 12:32:56.751781524 +0000 UTC m=+1407.980368387 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data") pod "rabbitmq-server-0" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912") : configmap "rabbitmq-config-data" not found Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.269574 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-wj5sj"] Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.274275 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:32:56 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: if [ -n "barbican" ]; then Feb 02 12:32:56 crc kubenswrapper[4846]: GRANT_DATABASE="barbican" Feb 02 12:32:56 crc kubenswrapper[4846]: else Feb 02 12:32:56 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:32:56 crc kubenswrapper[4846]: fi Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:32:56 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:32:56 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:32:56 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:32:56 crc kubenswrapper[4846]: # support updates Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.275374 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"barbican-db-secret\\\" not found\"" pod="openstack/barbican-c2bc-account-create-update-qz9kb" podUID="3915f1bc-20d3-4e7d-b627-46406d0d0461" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.315382 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f3b0-account-create-update-5xhdn"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.355478 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-operator-scripts\") pod \"nova-cell0-ad5a-account-create-update-f9gxt\" (UID: \"1d163475-4fe6-4992-90ae-4cfc97a1e6cd\") " pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.355684 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x59n9\" (UniqueName: \"kubernetes.io/projected/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-kube-api-access-x59n9\") pod \"nova-cell0-ad5a-account-create-update-f9gxt\" (UID: \"1d163475-4fe6-4992-90ae-4cfc97a1e6cd\") " pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.372108 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f3b0-account-create-update-5xhdn"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.393516 4846 generic.go:334] "Generic (PLEG): container finished" podID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerID="a6d2a885c9e46febf0a5624587138a27dc292d388c2adfac2d4f55ce324f5505" exitCode=2 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.393604 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e","Type":"ContainerDied","Data":"a6d2a885c9e46febf0a5624587138a27dc292d388c2adfac2d4f55ce324f5505"} Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.409311 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-f9gxt"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.427296 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="36061d30-6265-47ca-8a87-5275f41d03ed" containerName="ovsdbserver-nb" containerID="cri-o://b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389" gracePeriod=300 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.441766 4846 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder-scheduler-0" secret="" err="secret \"cinder-cinder-dockercfg-plfcv\" not found" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.442179 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-wj5sj"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.468587 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-operator-scripts\") pod \"nova-cell0-ad5a-account-create-update-f9gxt\" (UID: \"1d163475-4fe6-4992-90ae-4cfc97a1e6cd\") " pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.468916 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x59n9\" (UniqueName: \"kubernetes.io/projected/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-kube-api-access-x59n9\") pod \"nova-cell0-ad5a-account-create-update-f9gxt\" (UID: \"1d163475-4fe6-4992-90ae-4cfc97a1e6cd\") " pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.470081 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-operator-scripts\") pod \"nova-cell0-ad5a-account-create-update-f9gxt\" (UID: \"1d163475-4fe6-4992-90ae-4cfc97a1e6cd\") " pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.502863 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.503161 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="271eeb3d-6d83-49c6-9d80-75b2276a304a" containerName="glance-log" containerID="cri-o://a3bf2b4a419820c2d36fe369e8dac05f3052d401b7c0bdcc51a1cdf750ef3a24" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.503590 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="271eeb3d-6d83-49c6-9d80-75b2276a304a" containerName="glance-httpd" containerID="cri-o://dbd55a0c1904df42f4b3eb5b57e2aa4fadf58f7e576128bd91743f3ff542702f" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.503643 4846 generic.go:334] "Generic (PLEG): container finished" podID="b7187614-7a06-41a1-9835-d309ceb1ac9f" containerID="e9afbbe94cd0b66d2701c6bfe73f4de7c7680ba854f612982eafd6a0db29d1a5" exitCode=0 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.503739 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" event={"ID":"b7187614-7a06-41a1-9835-d309ceb1ac9f","Type":"ContainerDied","Data":"e9afbbe94cd0b66d2701c6bfe73f4de7c7680ba854f612982eafd6a0db29d1a5"} Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.526542 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerName="ovsdbserver-sb" containerID="cri-o://ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835" gracePeriod=300 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.537546 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x59n9\" (UniqueName: \"kubernetes.io/projected/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-kube-api-access-x59n9\") pod \"nova-cell0-ad5a-account-create-update-f9gxt\" (UID: \"1d163475-4fe6-4992-90ae-4cfc97a1e6cd\") " pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.539317 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c2bc-account-create-update-qz9kb"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.556312 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c2bc-account-create-update-qz9kb" event={"ID":"3915f1bc-20d3-4e7d-b627-46406d0d0461","Type":"ContainerStarted","Data":"327a79df04532b0e25d4d566df398944e9876a11015e055cd7ba2ea7dc899bbb"} Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.571851 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-c64lq"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.579286 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nlrm4_fd01ba51-8a4c-4b74-a7ab-382600f93e68/openstack-network-exporter/0.log" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.579334 4846 generic.go:334] "Generic (PLEG): container finished" podID="fd01ba51-8a4c-4b74-a7ab-382600f93e68" containerID="0448b20de6b4b9dfd9b17913d840a0f18245b4ede9e0f0f9853c064551787006" exitCode=2 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.579405 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nlrm4" event={"ID":"fd01ba51-8a4c-4b74-a7ab-382600f93e68","Type":"ContainerDied","Data":"0448b20de6b4b9dfd9b17913d840a0f18245b4ede9e0f0f9853c064551787006"} Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.597831 4846 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.597898 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:32:57.097879373 +0000 UTC m=+1408.326466316 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scripts" not found Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.598336 4846 secret.go:188] Couldn't get secret openstack/cinder-scheduler-config-data: secret "cinder-scheduler-config-data" not found Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.598375 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:32:57.098365675 +0000 UTC m=+1408.326952538 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scheduler-config-data" not found Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.604785 4846 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.604861 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:32:57.104842334 +0000 UTC m=+1408.333429197 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-config-data" not found Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.608079 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:32:56 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: if [ -n "barbican" ]; then Feb 02 12:32:56 crc kubenswrapper[4846]: GRANT_DATABASE="barbican" Feb 02 12:32:56 crc kubenswrapper[4846]: else Feb 02 12:32:56 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:32:56 crc kubenswrapper[4846]: fi Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:32:56 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:32:56 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:32:56 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:32:56 crc kubenswrapper[4846]: # support updates Feb 02 12:32:56 crc kubenswrapper[4846]: Feb 02 12:32:56 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.609838 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.614048 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"barbican-db-secret\\\" not found\"" pod="openstack/barbican-c2bc-account-create-update-qz9kb" podUID="3915f1bc-20d3-4e7d-b627-46406d0d0461" Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.656145 4846 generic.go:334] "Generic (PLEG): container finished" podID="bda23288-2fee-45ad-81e2-a14654c16589" containerID="ebe246dd795f71a1ad40ab382816989fc7965fac63ac05fe50ffa37d6e3182a0" exitCode=0 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.656211 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc" event={"ID":"bda23288-2fee-45ad-81e2-a14654c16589","Type":"ContainerDied","Data":"ebe246dd795f71a1ad40ab382816989fc7965fac63ac05fe50ffa37d6e3182a0"} Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.687714 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.688946 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-server" containerID="cri-o://eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.689468 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="swift-recon-cron" containerID="cri-o://c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.689560 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="rsync" containerID="cri-o://e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.689616 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-expirer" containerID="cri-o://8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.689686 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-updater" containerID="cri-o://38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.690449 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-auditor" containerID="cri-o://96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.690522 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-replicator" containerID="cri-o://5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.691016 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-server" containerID="cri-o://29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.691089 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-updater" containerID="cri-o://80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.691156 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-auditor" containerID="cri-o://0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.691233 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-replicator" containerID="cri-o://0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.691303 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-server" containerID="cri-o://9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.691337 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-reaper" containerID="cri-o://387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.691389 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-auditor" containerID="cri-o://6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.691419 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-replicator" containerID="cri-o://28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.711650 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f94a-account-create-update-zg8j8"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.749805 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-c64lq"] Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.806257 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.806338 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data podName:9cec353a-92f9-4ed4-8ec9-15271cbc6912 nodeName:}" failed. No retries permitted until 2026-02-02 12:32:57.80631643 +0000 UTC m=+1409.034903293 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data") pod "rabbitmq-server-0" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912") : configmap "rabbitmq-config-data" not found Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.822724 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.823052 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" containerName="glance-log" containerID="cri-o://7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.823531 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" containerName="glance-httpd" containerID="cri-o://265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.917557 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Feb 02 12:32:56 crc kubenswrapper[4846]: E0202 12:32:56.917628 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data podName:327c8143-b3ed-4431-a206-d2019ddcb606 nodeName:}" failed. No retries permitted until 2026-02-02 12:32:58.917601766 +0000 UTC m=+1410.146188629 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data") pod "rabbitmq-cell1-server-0" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606") : configmap "rabbitmq-cell1-config-data" not found Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.953744 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6497c67f7-24wxw"] Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.954091 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6497c67f7-24wxw" podUID="5a1c1394-3713-41ee-948e-3800e5d51020" containerName="neutron-api" containerID="cri-o://38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd" gracePeriod=30 Feb 02 12:32:56 crc kubenswrapper[4846]: I0202 12:32:56.954590 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6497c67f7-24wxw" podUID="5a1c1394-3713-41ee-948e-3800e5d51020" containerName="neutron-httpd" containerID="cri-o://f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.006668 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c2bc-account-create-update-qz9kb"] Feb 02 12:32:57 crc kubenswrapper[4846]: W0202 12:32:57.012191 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6f74db8_a48c_4358_81d0_41391a207008.slice/crio-791a74da2fcdd8a47c73f68a10370c78c2e796b5930e01db19b535181176d76b WatchSource:0}: Error finding container 791a74da2fcdd8a47c73f68a10370c78c2e796b5930e01db19b535181176d76b: Status 404 returned error can't find the container with id 791a74da2fcdd8a47c73f68a10370c78c2e796b5930e01db19b535181176d76b Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.055533 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-2z6vf"] Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.093030 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:32:57 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: if [ -n "glance" ]; then Feb 02 12:32:57 crc kubenswrapper[4846]: GRANT_DATABASE="glance" Feb 02 12:32:57 crc kubenswrapper[4846]: else Feb 02 12:32:57 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:32:57 crc kubenswrapper[4846]: fi Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:32:57 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:32:57 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:32:57 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:32:57 crc kubenswrapper[4846]: # support updates Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.095074 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"glance-db-secret\\\" not found\"" pod="openstack/glance-f94a-account-create-update-zg8j8" podUID="b6f74db8-a48c-4358-81d0-41391a207008" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.095687 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.099096 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835 is running failed: container process not found" containerID="ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835" cmd=["/usr/bin/pidof","ovsdb-server"] Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.102098 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835 is running failed: container process not found" containerID="ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835" cmd=["/usr/bin/pidof","ovsdb-server"] Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.103331 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835 is running failed: container process not found" containerID="ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835" cmd=["/usr/bin/pidof","ovsdb-server"] Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.103381 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerName="ovsdbserver-sb" Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.121875 4846 secret.go:188] Couldn't get secret openstack/cinder-scheduler-config-data: secret "cinder-scheduler-config-data" not found Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.121945 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:32:58.121928623 +0000 UTC m=+1409.350515486 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scheduler-config-data" not found Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.122294 4846 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.122334 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:32:58.122324612 +0000 UTC m=+1409.350911475 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-config-data" not found Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.125346 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-2z6vf"] Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.125822 4846 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.125892 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:32:58.125869699 +0000 UTC m=+1409.354456562 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scripts" not found Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.149228 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-gmb9m"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.160792 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-gmb9m"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.184719 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nlrm4_fd01ba51-8a4c-4b74-a7ab-382600f93e68/openstack-network-exporter/0.log" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.184777 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.198657 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c75698f54-h7x5d"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.201439 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-c75698f54-h7x5d" podUID="525708d2-15d1-4bfb-8258-bf216ebc4bab" containerName="placement-log" containerID="cri-o://4f2c0bed69df185a3aca7bd0bcabf6f5afe33072cfbe8448804ba379ea4ff18e" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.201748 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-c75698f54-h7x5d" podUID="525708d2-15d1-4bfb-8258-bf216ebc4bab" containerName="placement-api" containerID="cri-o://f7527fb5dfdb0a8afb41a261ae73e292db1a4a3cb90a3972583cde5f54856e3b" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.232942 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.234183 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovs-vswitchd" containerID="cri-o://c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" gracePeriod=29 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.235097 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerName="cinder-api-log" containerID="cri-o://428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.235185 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerName="cinder-api" containerID="cri-o://afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.255713 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-pnzdq"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.287802 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-pnzdq"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.325776 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovs-rundir\") pod \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.325875 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-metrics-certs-tls-certs\") pod \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.325928 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-combined-ca-bundle\") pod \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.325984 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cbhc\" (UniqueName: \"kubernetes.io/projected/fd01ba51-8a4c-4b74-a7ab-382600f93e68-kube-api-access-8cbhc\") pod \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.326053 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd01ba51-8a4c-4b74-a7ab-382600f93e68-config\") pod \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.326148 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovn-rundir\") pod \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\" (UID: \"fd01ba51-8a4c-4b74-a7ab-382600f93e68\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.327133 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "fd01ba51-8a4c-4b74-a7ab-382600f93e68" (UID: "fd01ba51-8a4c-4b74-a7ab-382600f93e68"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.327188 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "fd01ba51-8a4c-4b74-a7ab-382600f93e68" (UID: "fd01ba51-8a4c-4b74-a7ab-382600f93e68"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.334022 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd01ba51-8a4c-4b74-a7ab-382600f93e68-config" (OuterVolumeSpecName: "config") pod "fd01ba51-8a4c-4b74-a7ab-382600f93e68" (UID: "fd01ba51-8a4c-4b74-a7ab-382600f93e68"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.344838 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd01ba51-8a4c-4b74-a7ab-382600f93e68-kube-api-access-8cbhc" (OuterVolumeSpecName: "kube-api-access-8cbhc") pod "fd01ba51-8a4c-4b74-a7ab-382600f93e68" (UID: "fd01ba51-8a4c-4b74-a7ab-382600f93e68"). InnerVolumeSpecName "kube-api-access-8cbhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.345236 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-33ff-account-create-update-vz5ls"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.371258 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-nm26b"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.384809 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd01ba51-8a4c-4b74-a7ab-382600f93e68" (UID: "fd01ba51-8a4c-4b74-a7ab-382600f93e68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.388923 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-nm26b"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.410672 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-tb9t6"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.420734 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-tb9t6"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.430799 4846 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovs-rundir\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.430834 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.430845 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cbhc\" (UniqueName: \"kubernetes.io/projected/fd01ba51-8a4c-4b74-a7ab-382600f93e68-kube-api-access-8cbhc\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.430856 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd01ba51-8a4c-4b74-a7ab-382600f93e68-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.430868 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fd01ba51-8a4c-4b74-a7ab-382600f93e68-ovn-rundir\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.447436 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="037eb6da-9f07-4b72-9efe-5000aaa83b94" path="/var/lib/kubelet/pods/037eb6da-9f07-4b72-9efe-5000aaa83b94/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.448478 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17190291-9e13-48b3-aff4-fc7e84e966bd" path="/var/lib/kubelet/pods/17190291-9e13-48b3-aff4-fc7e84e966bd/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.449318 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18307d58-f249-4ec0-8556-fd98e4febb66" path="/var/lib/kubelet/pods/18307d58-f249-4ec0-8556-fd98e4febb66/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.450266 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28d3c741-e2ee-45b6-bb97-61485cf748be" path="/var/lib/kubelet/pods/28d3c741-e2ee-45b6-bb97-61485cf748be/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.451561 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dd98e97-bd56-4dde-8592-8836bbc2a0ee" path="/var/lib/kubelet/pods/3dd98e97-bd56-4dde-8592-8836bbc2a0ee/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.452362 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77c4e938-e657-40a2-9d0d-7261afc00a38" path="/var/lib/kubelet/pods/77c4e938-e657-40a2-9d0d-7261afc00a38/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.453730 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad01c88-75f3-49ef-be2a-19ca711923bd" path="/var/lib/kubelet/pods/8ad01c88-75f3-49ef-be2a-19ca711923bd/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.454457 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b58cb4f-6d15-4d97-871d-724c160ec765" path="/var/lib/kubelet/pods/8b58cb4f-6d15-4d97-871d-724c160ec765/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.455411 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a78e75-b5be-43f1-ac04-911885a75395" path="/var/lib/kubelet/pods/94a78e75-b5be-43f1-ac04-911885a75395/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.456902 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4d8a08e-0bf2-4830-b1f0-c5d599fc5757" path="/var/lib/kubelet/pods/a4d8a08e-0bf2-4830-b1f0-c5d599fc5757/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.458989 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7696a86-1c3e-4740-a2ce-ed1a45896127" path="/var/lib/kubelet/pods/a7696a86-1c3e-4740-a2ce-ed1a45896127/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.459800 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbe30827-4587-4d5a-a651-df65604399ee" path="/var/lib/kubelet/pods/bbe30827-4587-4d5a-a651-df65604399ee/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.462307 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdb0f9a1-5ddc-45ba-af84-87dc5e2af531" path="/var/lib/kubelet/pods/bdb0f9a1-5ddc-45ba-af84-87dc5e2af531/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.464026 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d670b5c2-8eea-4413-90df-0f4b72d8acd7" path="/var/lib/kubelet/pods/d670b5c2-8eea-4413-90df-0f4b72d8acd7/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.467197 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e02dd0b8-6bd8-40b2-8a18-64d298c72a67" path="/var/lib/kubelet/pods/e02dd0b8-6bd8-40b2-8a18-64d298c72a67/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.467564 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.469219 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e924adc7-fffe-4682-a601-cac1917e8f54" path="/var/lib/kubelet/pods/e924adc7-fffe-4682-a601-cac1917e8f54/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.472643 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ece3e0cc-5193-4c71-982b-84fe26ad56e1" path="/var/lib/kubelet/pods/ece3e0cc-5193-4c71-982b-84fe26ad56e1/volumes" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.479458 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-cd5d-account-create-update-v79s9"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.479499 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-df8f98698-ddglj"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.479871 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" podUID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" containerName="barbican-keystone-listener-log" containerID="cri-o://199fdc9812e4067e4c230bdb3772de61e8c1276ecc5f01f740fc1af2fa9d74c4" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.479955 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" podUID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" containerName="barbican-keystone-listener" containerID="cri-o://319bed97e3756574fc101f5461b63927d20d9217f036ffcd758a126ac4c88d35" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.506388 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_36061d30-6265-47ca-8a87-5275f41d03ed/ovsdbserver-nb/0.log" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.506470 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.508699 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-f6447845b-zlnf5"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.508933 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-f6447845b-zlnf5" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerName="barbican-api-log" containerID="cri-o://f357c7a0e61fbd03ef490e5638f0b35530ac99e278d44d9f752d0369a28df54b" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.509068 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-f6447845b-zlnf5" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerName="barbican-api" containerID="cri-o://b358fa2e3b3d3040b293911322aa36237a5340d9d50ce3922695a3ff266a9b44" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.519224 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-5d9447b957-4tn8l"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.519480 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-5d9447b957-4tn8l" podUID="da322cd4-f770-42d7-af2f-e4ed438da7dd" containerName="barbican-worker-log" containerID="cri-o://c6820cbfbd66e9d71a05716d17fadd1549cc9619a0b848aee68552415ffa96a1" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.519609 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-5d9447b957-4tn8l" podUID="da322cd4-f770-42d7-af2f-e4ed438da7dd" containerName="barbican-worker" containerID="cri-o://77c5ca7cfef457fcaadda875671e3c5d3ab8c7606c3409257026722b9bd9fa0f" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.533613 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md7cs\" (UniqueName: \"kubernetes.io/projected/bda23288-2fee-45ad-81e2-a14654c16589-kube-api-access-md7cs\") pod \"bda23288-2fee-45ad-81e2-a14654c16589\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.533812 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bda23288-2fee-45ad-81e2-a14654c16589-scripts\") pod \"bda23288-2fee-45ad-81e2-a14654c16589\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.533899 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-ovn-controller-tls-certs\") pod \"bda23288-2fee-45ad-81e2-a14654c16589\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.533925 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run-ovn\") pod \"bda23288-2fee-45ad-81e2-a14654c16589\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.534065 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run\") pod \"bda23288-2fee-45ad-81e2-a14654c16589\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.534141 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-log-ovn\") pod \"bda23288-2fee-45ad-81e2-a14654c16589\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.534265 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-combined-ca-bundle\") pod \"bda23288-2fee-45ad-81e2-a14654c16589\" (UID: \"bda23288-2fee-45ad-81e2-a14654c16589\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.536968 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "bda23288-2fee-45ad-81e2-a14654c16589" (UID: "bda23288-2fee-45ad-81e2-a14654c16589"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.537751 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bda23288-2fee-45ad-81e2-a14654c16589-scripts" (OuterVolumeSpecName: "scripts") pod "bda23288-2fee-45ad-81e2-a14654c16589" (UID: "bda23288-2fee-45ad-81e2-a14654c16589"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.538794 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run" (OuterVolumeSpecName: "var-run") pod "bda23288-2fee-45ad-81e2-a14654c16589" (UID: "bda23288-2fee-45ad-81e2-a14654c16589"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.538848 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "bda23288-2fee-45ad-81e2-a14654c16589" (UID: "bda23288-2fee-45ad-81e2-a14654c16589"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.544336 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.546852 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bda23288-2fee-45ad-81e2-a14654c16589-kube-api-access-md7cs" (OuterVolumeSpecName: "kube-api-access-md7cs") pod "bda23288-2fee-45ad-81e2-a14654c16589" (UID: "bda23288-2fee-45ad-81e2-a14654c16589"). InnerVolumeSpecName "kube-api-access-md7cs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.546903 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-crpfz"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.569732 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "fd01ba51-8a4c-4b74-a7ab-382600f93e68" (UID: "fd01ba51-8a4c-4b74-a7ab-382600f93e68"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.586759 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bda23288-2fee-45ad-81e2-a14654c16589" (UID: "bda23288-2fee-45ad-81e2-a14654c16589"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.589540 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a0f2-account-create-update-g95vn"] Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.611973 4846 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Feb 02 12:32:57 crc kubenswrapper[4846]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Feb 02 12:32:57 crc kubenswrapper[4846]: + source /usr/local/bin/container-scripts/functions Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNBridge=br-int Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNRemote=tcp:localhost:6642 Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNEncapType=geneve Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNAvailabilityZones= Feb 02 12:32:57 crc kubenswrapper[4846]: ++ EnableChassisAsGateway=true Feb 02 12:32:57 crc kubenswrapper[4846]: ++ PhysicalNetworks= Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNHostName= Feb 02 12:32:57 crc kubenswrapper[4846]: ++ DB_FILE=/etc/openvswitch/conf.db Feb 02 12:32:57 crc kubenswrapper[4846]: ++ ovs_dir=/var/lib/openvswitch Feb 02 12:32:57 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Feb 02 12:32:57 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Feb 02 12:32:57 crc kubenswrapper[4846]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + sleep 0.5 Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + sleep 0.5 Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + sleep 0.5 Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + cleanup_ovsdb_server_semaphore Feb 02 12:32:57 crc kubenswrapper[4846]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 02 12:32:57 crc kubenswrapper[4846]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Feb 02 12:32:57 crc kubenswrapper[4846]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-m79vg" message=< Feb 02 12:32:57 crc kubenswrapper[4846]: Exiting ovsdb-server (5) [ OK ] Feb 02 12:32:57 crc kubenswrapper[4846]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Feb 02 12:32:57 crc kubenswrapper[4846]: + source /usr/local/bin/container-scripts/functions Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNBridge=br-int Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNRemote=tcp:localhost:6642 Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNEncapType=geneve Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNAvailabilityZones= Feb 02 12:32:57 crc kubenswrapper[4846]: ++ EnableChassisAsGateway=true Feb 02 12:32:57 crc kubenswrapper[4846]: ++ PhysicalNetworks= Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNHostName= Feb 02 12:32:57 crc kubenswrapper[4846]: ++ DB_FILE=/etc/openvswitch/conf.db Feb 02 12:32:57 crc kubenswrapper[4846]: ++ ovs_dir=/var/lib/openvswitch Feb 02 12:32:57 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Feb 02 12:32:57 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Feb 02 12:32:57 crc kubenswrapper[4846]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + sleep 0.5 Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + sleep 0.5 Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + sleep 0.5 Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + cleanup_ovsdb_server_semaphore Feb 02 12:32:57 crc kubenswrapper[4846]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 02 12:32:57 crc kubenswrapper[4846]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Feb 02 12:32:57 crc kubenswrapper[4846]: > Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.612022 4846 kuberuntime_container.go:691] "PreStop hook failed" err=< Feb 02 12:32:57 crc kubenswrapper[4846]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Feb 02 12:32:57 crc kubenswrapper[4846]: + source /usr/local/bin/container-scripts/functions Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNBridge=br-int Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNRemote=tcp:localhost:6642 Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNEncapType=geneve Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNAvailabilityZones= Feb 02 12:32:57 crc kubenswrapper[4846]: ++ EnableChassisAsGateway=true Feb 02 12:32:57 crc kubenswrapper[4846]: ++ PhysicalNetworks= Feb 02 12:32:57 crc kubenswrapper[4846]: ++ OVNHostName= Feb 02 12:32:57 crc kubenswrapper[4846]: ++ DB_FILE=/etc/openvswitch/conf.db Feb 02 12:32:57 crc kubenswrapper[4846]: ++ ovs_dir=/var/lib/openvswitch Feb 02 12:32:57 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Feb 02 12:32:57 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Feb 02 12:32:57 crc kubenswrapper[4846]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + sleep 0.5 Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + sleep 0.5 Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + sleep 0.5 Feb 02 12:32:57 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 02 12:32:57 crc kubenswrapper[4846]: + cleanup_ovsdb_server_semaphore Feb 02 12:32:57 crc kubenswrapper[4846]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 02 12:32:57 crc kubenswrapper[4846]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Feb 02 12:32:57 crc kubenswrapper[4846]: > pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" containerID="cri-o://7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.612066 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" containerID="cri-o://7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" gracePeriod=28 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.613391 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-crpfz"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.635307 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.635465 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbmtk\" (UniqueName: \"kubernetes.io/projected/b7187614-7a06-41a1-9835-d309ceb1ac9f-kube-api-access-lbmtk\") pod \"b7187614-7a06-41a1-9835-d309ceb1ac9f\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.635504 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"36061d30-6265-47ca-8a87-5275f41d03ed\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.635569 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-svc\") pod \"b7187614-7a06-41a1-9835-d309ceb1ac9f\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.635603 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqw5k\" (UniqueName: \"kubernetes.io/projected/36061d30-6265-47ca-8a87-5275f41d03ed-kube-api-access-dqw5k\") pod \"36061d30-6265-47ca-8a87-5275f41d03ed\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.635694 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-combined-ca-bundle\") pod \"36061d30-6265-47ca-8a87-5275f41d03ed\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.635755 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-scripts\") pod \"36061d30-6265-47ca-8a87-5275f41d03ed\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.638978 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-scripts" (OuterVolumeSpecName: "scripts") pod "36061d30-6265-47ca-8a87-5275f41d03ed" (UID: "36061d30-6265-47ca-8a87-5275f41d03ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.648602 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36061d30-6265-47ca-8a87-5275f41d03ed-kube-api-access-dqw5k" (OuterVolumeSpecName: "kube-api-access-dqw5k") pod "36061d30-6265-47ca-8a87-5275f41d03ed" (UID: "36061d30-6265-47ca-8a87-5275f41d03ed"). InnerVolumeSpecName "kube-api-access-dqw5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.648767 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "36061d30-6265-47ca-8a87-5275f41d03ed" (UID: "36061d30-6265-47ca-8a87-5275f41d03ed"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.648691 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.649390 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerName="nova-api-log" containerID="cri-o://ceb10096d86d9be3fc00e39352a6fcbaf81ecf7bd346d3e24a99c69fe60cd9f2" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.650030 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerName="nova-api-api" containerID="cri-o://99b0a8947603d649bb9d1f3decf69b9ee0004e7f9c373151604124884a5d8445" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.661756 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-metrics-certs-tls-certs\") pod \"36061d30-6265-47ca-8a87-5275f41d03ed\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.662406 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-nb\") pod \"b7187614-7a06-41a1-9835-d309ceb1ac9f\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.662822 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-config\") pod \"36061d30-6265-47ca-8a87-5275f41d03ed\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.662891 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-swift-storage-0\") pod \"b7187614-7a06-41a1-9835-d309ceb1ac9f\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.662930 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-config\") pod \"b7187614-7a06-41a1-9835-d309ceb1ac9f\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.663019 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-sb\") pod \"b7187614-7a06-41a1-9835-d309ceb1ac9f\" (UID: \"b7187614-7a06-41a1-9835-d309ceb1ac9f\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.663078 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdb-rundir\") pod \"36061d30-6265-47ca-8a87-5275f41d03ed\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.663116 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdbserver-nb-tls-certs\") pod \"36061d30-6265-47ca-8a87-5275f41d03ed\" (UID: \"36061d30-6265-47ca-8a87-5275f41d03ed\") " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.664862 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.666469 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7187614-7a06-41a1-9835-d309ceb1ac9f-kube-api-access-lbmtk" (OuterVolumeSpecName: "kube-api-access-lbmtk") pod "b7187614-7a06-41a1-9835-d309ceb1ac9f" (UID: "b7187614-7a06-41a1-9835-d309ceb1ac9f"). InnerVolumeSpecName "kube-api-access-lbmtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.667264 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-config" (OuterVolumeSpecName: "config") pod "36061d30-6265-47ca-8a87-5275f41d03ed" (UID: "36061d30-6265-47ca-8a87-5275f41d03ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.667679 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "36061d30-6265-47ca-8a87-5275f41d03ed" (UID: "36061d30-6265-47ca-8a87-5275f41d03ed"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684401 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684428 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bda23288-2fee-45ad-81e2-a14654c16589-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684438 4846 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684446 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684456 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-run\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684464 4846 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bda23288-2fee-45ad-81e2-a14654c16589-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684473 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbmtk\" (UniqueName: \"kubernetes.io/projected/b7187614-7a06-41a1-9835-d309ceb1ac9f-kube-api-access-lbmtk\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684498 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684510 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd01ba51-8a4c-4b74-a7ab-382600f93e68-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684522 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqw5k\" (UniqueName: \"kubernetes.io/projected/36061d30-6265-47ca-8a87-5275f41d03ed-kube-api-access-dqw5k\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684534 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684546 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36061d30-6265-47ca-8a87-5275f41d03ed-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.684557 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md7cs\" (UniqueName: \"kubernetes.io/projected/bda23288-2fee-45ad-81e2-a14654c16589-kube-api-access-md7cs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.713573 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.731345 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_36061d30-6265-47ca-8a87-5275f41d03ed/ovsdbserver-nb/0.log" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.731384 4846 generic.go:334] "Generic (PLEG): container finished" podID="36061d30-6265-47ca-8a87-5275f41d03ed" containerID="0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341" exitCode=2 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.731400 4846 generic.go:334] "Generic (PLEG): container finished" podID="36061d30-6265-47ca-8a87-5275f41d03ed" containerID="b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389" exitCode=143 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.731501 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"36061d30-6265-47ca-8a87-5275f41d03ed","Type":"ContainerDied","Data":"0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.731528 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"36061d30-6265-47ca-8a87-5275f41d03ed","Type":"ContainerDied","Data":"b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.731551 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"36061d30-6265-47ca-8a87-5275f41d03ed","Type":"ContainerDied","Data":"e56b4c06f56bec2f17b4f1e52b94306479c4be44b0443cd4e2e480cedf3904d7"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.731565 4846 scope.go:117] "RemoveContainer" containerID="0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.731706 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.745871 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-798b-account-create-update-7g5s7"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.752084 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-98cnq"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.753787 4846 generic.go:334] "Generic (PLEG): container finished" podID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" containerID="199fdc9812e4067e4c230bdb3772de61e8c1276ecc5f01f740fc1af2fa9d74c4" exitCode=143 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.753851 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" event={"ID":"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3","Type":"ContainerDied","Data":"199fdc9812e4067e4c230bdb3772de61e8c1276ecc5f01f740fc1af2fa9d74c4"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.767560 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="327c8143-b3ed-4431-a206-d2019ddcb606" containerName="rabbitmq" containerID="cri-o://ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56" gracePeriod=604800 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.777183 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-98cnq"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.803379 4846 generic.go:334] "Generic (PLEG): container finished" podID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" containerID="7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686" exitCode=143 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.803669 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9","Type":"ContainerDied","Data":"7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.805220 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.805393 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-4xltk"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.812682 4846 generic.go:334] "Generic (PLEG): container finished" podID="525708d2-15d1-4bfb-8258-bf216ebc4bab" containerID="4f2c0bed69df185a3aca7bd0bcabf6f5afe33072cfbe8448804ba379ea4ff18e" exitCode=143 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.812757 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c75698f54-h7x5d" event={"ID":"525708d2-15d1-4bfb-8258-bf216ebc4bab","Type":"ContainerDied","Data":"4f2c0bed69df185a3aca7bd0bcabf6f5afe33072cfbe8448804ba379ea4ff18e"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.814950 4846 generic.go:334] "Generic (PLEG): container finished" podID="5a1c1394-3713-41ee-948e-3800e5d51020" containerID="f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.815003 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6497c67f7-24wxw" event={"ID":"5a1c1394-3713-41ee-948e-3800e5d51020","Type":"ContainerDied","Data":"f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.820069 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36061d30-6265-47ca-8a87-5275f41d03ed" (UID: "36061d30-6265-47ca-8a87-5275f41d03ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.824539 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.824608 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-f9gxt"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.824656 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b99f9475-c48q4" event={"ID":"b7187614-7a06-41a1-9835-d309ceb1ac9f","Type":"ContainerDied","Data":"acb55e106bbee719c2bd0ceb2492ed79d97054ab454e95d754bd7535ef260c97"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.831767 4846 generic.go:334] "Generic (PLEG): container finished" podID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerID="f357c7a0e61fbd03ef490e5638f0b35530ac99e278d44d9f752d0369a28df54b" exitCode=143 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.832075 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6447845b-zlnf5" event={"ID":"69522beb-f7d5-4694-ada8-0d17d7d3f16b","Type":"ContainerDied","Data":"f357c7a0e61fbd03ef490e5638f0b35530ac99e278d44d9f752d0369a28df54b"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.842210 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b2hcc" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.842607 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b2hcc" event={"ID":"bda23288-2fee-45ad-81e2-a14654c16589","Type":"ContainerDied","Data":"de6f75ec04c419bcd53e3d033b604026fce641668fdb4d0cd503c6ac74c36ec4"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.846535 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.847187 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-log" containerID="cri-o://66365bb3d3091805961a6da62564f6a7bf8e51dc20054684c80e03009a5d18dd" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.847420 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-metadata" containerID="cri-o://ba2153f30138c7492692991e0000a0e9232db386da8169c4c360432f8ac0b9a2" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.855569 4846 generic.go:334] "Generic (PLEG): container finished" podID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.855814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m79vg" event={"ID":"53dd5f18-d111-4359-8170-102a8b30d0d9","Type":"ContainerDied","Data":"7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.859380 4846 generic.go:334] "Generic (PLEG): container finished" podID="ba822b20-88d5-404b-850c-7f9441933baf" containerID="b4630774f6b8174e6eaa2407dc38ee14e1e53c6776c23ce799141cd61d5f6c49" exitCode=137 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.860900 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nlrm4_fd01ba51-8a4c-4b74-a7ab-382600f93e68/openstack-network-exporter/0.log" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.860954 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nlrm4" event={"ID":"fd01ba51-8a4c-4b74-a7ab-382600f93e68","Type":"ContainerDied","Data":"14b0c39bd579b305c315f664d839d215ca357c7387892d18fcf466087cdd5b44"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.861027 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nlrm4" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.870102 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-4xltk"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.887801 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_797602a8-7d50-4ab4-8ba5-4afff6a34409/ovsdbserver-sb/0.log" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.889120 4846 generic.go:334] "Generic (PLEG): container finished" podID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerID="8cd3e31565aad9a3a479cdc1ab4236b7a0085ec750392646a09a6087a0894db0" exitCode=2 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.889145 4846 generic.go:334] "Generic (PLEG): container finished" podID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerID="ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835" exitCode=143 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.893105 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b7187614-7a06-41a1-9835-d309ceb1ac9f" (UID: "b7187614-7a06-41a1-9835-d309ceb1ac9f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.893202 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-gkxqm"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.893231 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"797602a8-7d50-4ab4-8ba5-4afff6a34409","Type":"ContainerDied","Data":"8cd3e31565aad9a3a479cdc1ab4236b7a0085ec750392646a09a6087a0894db0"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.893253 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"797602a8-7d50-4ab4-8ba5-4afff6a34409","Type":"ContainerDied","Data":"ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835"} Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.914140 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.915017 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data podName:9cec353a-92f9-4ed4-8ec9-15271cbc6912 nodeName:}" failed. No retries permitted until 2026-02-02 12:32:59.914995684 +0000 UTC m=+1411.143582597 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data") pod "rabbitmq-server-0" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912") : configmap "rabbitmq-config-data" not found Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.914251 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.918966 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.918311 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9447b957-4tn8l" event={"ID":"da322cd4-f770-42d7-af2f-e4ed438da7dd","Type":"ContainerDied","Data":"c6820cbfbd66e9d71a05716d17fadd1549cc9619a0b848aee68552415ffa96a1"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.918288 4846 generic.go:334] "Generic (PLEG): container finished" podID="da322cd4-f770-42d7-af2f-e4ed438da7dd" containerID="c6820cbfbd66e9d71a05716d17fadd1549cc9619a0b848aee68552415ffa96a1" exitCode=143 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.921986 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f94a-account-create-update-zg8j8"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.927038 4846 generic.go:334] "Generic (PLEG): container finished" podID="271eeb3d-6d83-49c6-9d80-75b2276a304a" containerID="a3bf2b4a419820c2d36fe369e8dac05f3052d401b7c0bdcc51a1cdf750ef3a24" exitCode=143 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.927092 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"271eeb3d-6d83-49c6-9d80-75b2276a304a","Type":"ContainerDied","Data":"a3bf2b4a419820c2d36fe369e8dac05f3052d401b7c0bdcc51a1cdf750ef3a24"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.927799 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-config" (OuterVolumeSpecName: "config") pod "b7187614-7a06-41a1-9835-d309ceb1ac9f" (UID: "b7187614-7a06-41a1-9835-d309ceb1ac9f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.930559 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-gkxqm"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.941773 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.941981 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="fa1b6480-161c-433d-975b-83079f2c979a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://c50d7d730873e682af6be480671b01b55a5ad561b1a9bbf213963fb50c01e8d1" gracePeriod=30 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.961336 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rr425"] Feb 02 12:32:57 crc kubenswrapper[4846]: W0202 12:32:57.966904 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71584556_0f8c_4781_bbd6_2e3f6913f0e1.slice/crio-babc16a10569ff1d331b10fa3f112c4f0c1c3761793bc6bca6f0a2b9212490c2 WatchSource:0}: Error finding container babc16a10569ff1d331b10fa3f112c4f0c1c3761793bc6bca6f0a2b9212490c2: Status 404 returned error can't find the container with id babc16a10569ff1d331b10fa3f112c4f0c1c3761793bc6bca6f0a2b9212490c2 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978517 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978543 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978572 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978580 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978587 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978595 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978601 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978606 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978612 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978629 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978635 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978641 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978647 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978652 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d" exitCode=0 Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978695 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978721 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978748 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978757 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978767 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978776 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978783 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978799 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978806 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978815 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978823 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978831 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.978840 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d"} Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.979357 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:32:57 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: if [ -n "" ]; then Feb 02 12:32:57 crc kubenswrapper[4846]: GRANT_DATABASE="" Feb 02 12:32:57 crc kubenswrapper[4846]: else Feb 02 12:32:57 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:32:57 crc kubenswrapper[4846]: fi Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:32:57 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:32:57 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:32:57 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:32:57 crc kubenswrapper[4846]: # support updates Feb 02 12:32:57 crc kubenswrapper[4846]: Feb 02 12:32:57 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.983035 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b7187614-7a06-41a1-9835-d309ceb1ac9f" (UID: "b7187614-7a06-41a1-9835-d309ceb1ac9f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:57 crc kubenswrapper[4846]: E0202 12:32:57.983105 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"openstack-cell1-mariadb-root-db-secret\\\" not found\"" pod="openstack/root-account-create-update-rr425" podUID="71584556-0f8c-4781-bbd6-2e3f6913f0e1" Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.983647 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f94a-account-create-update-zg8j8" event={"ID":"b6f74db8-a48c-4358-81d0-41391a207008","Type":"ContainerStarted","Data":"791a74da2fcdd8a47c73f68a10370c78c2e796b5930e01db19b535181176d76b"} Feb 02 12:32:57 crc kubenswrapper[4846]: I0202 12:32:57.991250 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.015106 4846 generic.go:334] "Generic (PLEG): container finished" podID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerID="428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646" exitCode=143 Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.015198 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"532236a1-a04d-4b99-9131-b1d16a3d5ed4","Type":"ContainerDied","Data":"428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646"} Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.015306 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" containerName="cinder-scheduler" containerID="cri-o://7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806" gracePeriod=30 Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.016350 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" containerName="probe" containerID="cri-o://642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210" gracePeriod=30 Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.025138 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.025168 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.031109 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:32:58 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: if [ -n "barbican" ]; then Feb 02 12:32:58 crc kubenswrapper[4846]: GRANT_DATABASE="barbican" Feb 02 12:32:58 crc kubenswrapper[4846]: else Feb 02 12:32:58 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:32:58 crc kubenswrapper[4846]: fi Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:32:58 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:32:58 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:32:58 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:32:58 crc kubenswrapper[4846]: # support updates Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.032886 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"barbican-db-secret\\\" not found\"" pod="openstack/barbican-c2bc-account-create-update-qz9kb" podUID="3915f1bc-20d3-4e7d-b627-46406d0d0461" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.039192 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b7187614-7a06-41a1-9835-d309ceb1ac9f" (UID: "b7187614-7a06-41a1-9835-d309ceb1ac9f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.041182 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "bda23288-2fee-45ad-81e2-a14654c16589" (UID: "bda23288-2fee-45ad-81e2-a14654c16589"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.066461 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b7187614-7a06-41a1-9835-d309ceb1ac9f" (UID: "b7187614-7a06-41a1-9835-d309ceb1ac9f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.067594 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.067844 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7f25197c-8c4e-41ef-b7ef-1ed0f141599b" containerName="nova-scheduler-scheduler" containerID="cri-o://24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8" gracePeriod=30 Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.072070 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="9cec353a-92f9-4ed4-8ec9-15271cbc6912" containerName="rabbitmq" containerID="cri-o://488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2" gracePeriod=604800 Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.104487 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rr425"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.188954 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "36061d30-6265-47ca-8a87-5275f41d03ed" (UID: "36061d30-6265-47ca-8a87-5275f41d03ed"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.216130 4846 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.216225 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:00.216196594 +0000 UTC m=+1411.444783457 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scripts" not found Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.217883 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bda23288-2fee-45ad-81e2-a14654c16589-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.231598 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.231731 4846 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.231797 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:00.231777575 +0000 UTC m=+1411.460364438 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-config-data" not found Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.232187 4846 secret.go:188] Couldn't get secret openstack/cinder-scheduler-config-data: secret "cinder-scheduler-config-data" not found Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.232243 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:00.232232836 +0000 UTC m=+1411.460819699 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scheduler-config-data" not found Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.232266 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.232283 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7187614-7a06-41a1-9835-d309ceb1ac9f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.234472 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_797602a8-7d50-4ab4-8ba5-4afff6a34409/ovsdbserver-sb/0.log" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.234566 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.249457 4846 scope.go:117] "RemoveContainer" containerID="b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.253825 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "36061d30-6265-47ca-8a87-5275f41d03ed" (UID: "36061d30-6265-47ca-8a87-5275f41d03ed"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.288067 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7fjsq"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.306301 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.306510 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="d5f78273-c1e2-452f-8bcb-a51c9bc238e4" containerName="nova-cell1-conductor-conductor" containerID="cri-o://b89c77647ebfc096292368b4c128270a751702e9c4f3ecf26b326d816dff5fbc" gracePeriod=30 Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.322003 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7fjsq"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.335675 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g5kjm"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.350591 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdb-rundir\") pod \"797602a8-7d50-4ab4-8ba5-4afff6a34409\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.350687 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-config\") pod \"797602a8-7d50-4ab4-8ba5-4afff6a34409\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.350723 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-combined-ca-bundle\") pod \"797602a8-7d50-4ab4-8ba5-4afff6a34409\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.350786 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-metrics-certs-tls-certs\") pod \"797602a8-7d50-4ab4-8ba5-4afff6a34409\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.350831 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-scripts\") pod \"797602a8-7d50-4ab4-8ba5-4afff6a34409\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.350900 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdbserver-sb-tls-certs\") pod \"797602a8-7d50-4ab4-8ba5-4afff6a34409\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.350936 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n9nc\" (UniqueName: \"kubernetes.io/projected/797602a8-7d50-4ab4-8ba5-4afff6a34409-kube-api-access-5n9nc\") pod \"797602a8-7d50-4ab4-8ba5-4afff6a34409\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.350999 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"797602a8-7d50-4ab4-8ba5-4afff6a34409\" (UID: \"797602a8-7d50-4ab4-8ba5-4afff6a34409\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.351578 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36061d30-6265-47ca-8a87-5275f41d03ed-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.352809 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-scripts" (OuterVolumeSpecName: "scripts") pod "797602a8-7d50-4ab4-8ba5-4afff6a34409" (UID: "797602a8-7d50-4ab4-8ba5-4afff6a34409"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.353752 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-config" (OuterVolumeSpecName: "config") pod "797602a8-7d50-4ab4-8ba5-4afff6a34409" (UID: "797602a8-7d50-4ab4-8ba5-4afff6a34409"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.359416 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "797602a8-7d50-4ab4-8ba5-4afff6a34409" (UID: "797602a8-7d50-4ab4-8ba5-4afff6a34409"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.364689 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-g5kjm"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.395801 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.396102 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="c7dd2f87-cea2-4575-ad0c-483944433a55" containerName="nova-cell0-conductor-conductor" containerID="cri-o://40dea2331c92832ae0ee4b74bc35edbdedf985815110294b594bde6cb57fb223" gracePeriod=30 Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.405911 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-nlrm4"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.415091 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-nlrm4"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.417931 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "797602a8-7d50-4ab4-8ba5-4afff6a34409" (UID: "797602a8-7d50-4ab4-8ba5-4afff6a34409"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.418018 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797602a8-7d50-4ab4-8ba5-4afff6a34409-kube-api-access-5n9nc" (OuterVolumeSpecName: "kube-api-access-5n9nc") pod "797602a8-7d50-4ab4-8ba5-4afff6a34409" (UID: "797602a8-7d50-4ab4-8ba5-4afff6a34409"). InnerVolumeSpecName "kube-api-access-5n9nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.418367 4846 scope.go:117] "RemoveContainer" containerID="0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.421203 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "797602a8-7d50-4ab4-8ba5-4afff6a34409" (UID: "797602a8-7d50-4ab4-8ba5-4afff6a34409"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.421300 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341\": container with ID starting with 0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341 not found: ID does not exist" containerID="0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.421361 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341"} err="failed to get container status \"0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341\": rpc error: code = NotFound desc = could not find container \"0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341\": container with ID starting with 0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341 not found: ID does not exist" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.421385 4846 scope.go:117] "RemoveContainer" containerID="b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389" Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.421719 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389\": container with ID starting with b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389 not found: ID does not exist" containerID="b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.421751 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389"} err="failed to get container status \"b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389\": rpc error: code = NotFound desc = could not find container \"b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389\": container with ID starting with b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389 not found: ID does not exist" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.421774 4846 scope.go:117] "RemoveContainer" containerID="0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.421995 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341"} err="failed to get container status \"0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341\": rpc error: code = NotFound desc = could not find container \"0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341\": container with ID starting with 0fdac965392a1e74bdb8aa39e207f0a5f122e5845dc0b323cb66cf6232cef341 not found: ID does not exist" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.422010 4846 scope.go:117] "RemoveContainer" containerID="b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.422165 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389"} err="failed to get container status \"b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389\": rpc error: code = NotFound desc = could not find container \"b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389\": container with ID starting with b36708e0ab1d4fec2abb82a7fac50c1f01cfbf5198a5843c3799e61a7ca56389 not found: ID does not exist" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.422176 4846 scope.go:117] "RemoveContainer" containerID="e9afbbe94cd0b66d2701c6bfe73f4de7c7680ba854f612982eafd6a0db29d1a5" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.437117 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-c48q4"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.446998 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74b99f9475-c48q4"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.447582 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" containerName="galera" containerID="cri-o://481554f18c651c06a0af22ab57eb2b59897fae18ef156d9d05b24af20a349d9a" gracePeriod=30 Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.453568 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.453598 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.453608 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.453639 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/797602a8-7d50-4ab4-8ba5-4afff6a34409-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.453652 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n9nc\" (UniqueName: \"kubernetes.io/projected/797602a8-7d50-4ab4-8ba5-4afff6a34409-kube-api-access-5n9nc\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.453678 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.462296 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-b2hcc"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.470856 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-b2hcc"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.481689 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.481750 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.502214 4846 scope.go:117] "RemoveContainer" containerID="af94415cb893888a9711b1b30d691047d195ad2f337e8d9c9b9a3cdc61e8c9db" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.523917 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.533747 4846 scope.go:117] "RemoveContainer" containerID="ebe246dd795f71a1ad40ab382816989fc7965fac63ac05fe50ffa37d6e3182a0" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.554674 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.572772 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "797602a8-7d50-4ab4-8ba5-4afff6a34409" (UID: "797602a8-7d50-4ab4-8ba5-4afff6a34409"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.610909 4846 scope.go:117] "RemoveContainer" containerID="0448b20de6b4b9dfd9b17913d840a0f18245b4ede9e0f0f9853c064551787006" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.641932 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.656291 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-combined-ca-bundle\") pod \"ba822b20-88d5-404b-850c-7f9441933baf\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.656338 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config\") pod \"ba822b20-88d5-404b-850c-7f9441933baf\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.656468 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config-secret\") pod \"ba822b20-88d5-404b-850c-7f9441933baf\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.656527 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt6kt\" (UniqueName: \"kubernetes.io/projected/ba822b20-88d5-404b-850c-7f9441933baf-kube-api-access-vt6kt\") pod \"ba822b20-88d5-404b-850c-7f9441933baf\" (UID: \"ba822b20-88d5-404b-850c-7f9441933baf\") " Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.657041 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.661691 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba822b20-88d5-404b-850c-7f9441933baf-kube-api-access-vt6kt" (OuterVolumeSpecName: "kube-api-access-vt6kt") pod "ba822b20-88d5-404b-850c-7f9441933baf" (UID: "ba822b20-88d5-404b-850c-7f9441933baf"). InnerVolumeSpecName "kube-api-access-vt6kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.716778 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba822b20-88d5-404b-850c-7f9441933baf" (UID: "ba822b20-88d5-404b-850c-7f9441933baf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.729300 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ba822b20-88d5-404b-850c-7f9441933baf" (UID: "ba822b20-88d5-404b-850c-7f9441933baf"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.751510 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ba822b20-88d5-404b-850c-7f9441933baf" (UID: "ba822b20-88d5-404b-850c-7f9441933baf"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.759593 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.759865 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.759964 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ba822b20-88d5-404b-850c-7f9441933baf-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.760048 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt6kt\" (UniqueName: \"kubernetes.io/projected/ba822b20-88d5-404b-850c-7f9441933baf-kube-api-access-vt6kt\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.806323 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "797602a8-7d50-4ab4-8ba5-4afff6a34409" (UID: "797602a8-7d50-4ab4-8ba5-4afff6a34409"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.861585 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/797602a8-7d50-4ab4-8ba5-4afff6a34409-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.894929 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a0f2-account-create-update-g95vn"] Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.963585 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.963672 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data podName:327c8143-b3ed-4431-a206-d2019ddcb606 nodeName:}" failed. No retries permitted until 2026-02-02 12:33:02.963655167 +0000 UTC m=+1414.192242030 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data") pod "rabbitmq-cell1-server-0" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606") : configmap "rabbitmq-cell1-config-data" not found Feb 02 12:32:58 crc kubenswrapper[4846]: I0202 12:32:58.967864 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-cd5d-account-create-update-v79s9"] Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.977389 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:32:58 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: if [ -n "placement" ]; then Feb 02 12:32:58 crc kubenswrapper[4846]: GRANT_DATABASE="placement" Feb 02 12:32:58 crc kubenswrapper[4846]: else Feb 02 12:32:58 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:32:58 crc kubenswrapper[4846]: fi Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:32:58 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:32:58 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:32:58 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:32:58 crc kubenswrapper[4846]: # support updates Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.979204 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"placement-db-secret\\\" not found\"" pod="openstack/placement-a0f2-account-create-update-g95vn" podUID="24671593-6e2f-44b9-909f-66b7a794c372" Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.980839 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:32:58 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: if [ -n "cinder" ]; then Feb 02 12:32:58 crc kubenswrapper[4846]: GRANT_DATABASE="cinder" Feb 02 12:32:58 crc kubenswrapper[4846]: else Feb 02 12:32:58 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:32:58 crc kubenswrapper[4846]: fi Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:32:58 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:32:58 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:32:58 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:32:58 crc kubenswrapper[4846]: # support updates Feb 02 12:32:58 crc kubenswrapper[4846]: Feb 02 12:32:58 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:32:58 crc kubenswrapper[4846]: E0202 12:32:58.982363 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"cinder-db-secret\\\" not found\"" pod="openstack/cinder-cd5d-account-create-update-v79s9" podUID="db660713-b11b-416f-9fb6-075facb7ecf0" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:58.993659 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="fa1b6480-161c-433d-975b-83079f2c979a" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.201:6080/vnc_lite.html\": dial tcp 10.217.0.201:6080: connect: connection refused" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:58.998090 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f94a-account-create-update-zg8j8" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.041704 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-33ff-account-create-update-vz5ls"] Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.052198 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_797602a8-7d50-4ab4-8ba5-4afff6a34409/ovsdbserver-sb/0.log" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.052439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"797602a8-7d50-4ab4-8ba5-4afff6a34409","Type":"ContainerDied","Data":"b4e9070cc9cc12c08941b4b81e99ada4d05d93e6997c9d731c3f64fef3b9183c"} Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.052485 4846 scope.go:117] "RemoveContainer" containerID="8cd3e31565aad9a3a479cdc1ab4236b7a0085ec750392646a09a6087a0894db0" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.052742 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.060350 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerID="66365bb3d3091805961a6da62564f6a7bf8e51dc20054684c80e03009a5d18dd" exitCode=143 Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.060431 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e2d348d-dbbf-4046-b9fb-94f8aab64573","Type":"ContainerDied","Data":"66365bb3d3091805961a6da62564f6a7bf8e51dc20054684c80e03009a5d18dd"} Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.067413 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0f2-account-create-update-g95vn" event={"ID":"24671593-6e2f-44b9-909f-66b7a794c372","Type":"ContainerStarted","Data":"f2547c641705890efbad2385a672a907ef7af23af6a26c3887bab778ddb7184d"} Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.071740 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cd5d-account-create-update-v79s9" event={"ID":"db660713-b11b-416f-9fb6-075facb7ecf0","Type":"ContainerStarted","Data":"cbb64985f175d499483f9119475d235ddd1f1e53765727c08160fad1904feab2"} Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.073251 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f94a-account-create-update-zg8j8" event={"ID":"b6f74db8-a48c-4358-81d0-41391a207008","Type":"ContainerDied","Data":"791a74da2fcdd8a47c73f68a10370c78c2e796b5930e01db19b535181176d76b"} Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.073304 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f94a-account-create-update-zg8j8" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.074863 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rr425" event={"ID":"71584556-0f8c-4781-bbd6-2e3f6913f0e1","Type":"ContainerStarted","Data":"babc16a10569ff1d331b10fa3f112c4f0c1c3761793bc6bca6f0a2b9212490c2"} Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.077271 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2sw2\" (UniqueName: \"kubernetes.io/projected/b6f74db8-a48c-4358-81d0-41391a207008-kube-api-access-q2sw2\") pod \"b6f74db8-a48c-4358-81d0-41391a207008\" (UID: \"b6f74db8-a48c-4358-81d0-41391a207008\") " Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.077507 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f74db8-a48c-4358-81d0-41391a207008-operator-scripts\") pod \"b6f74db8-a48c-4358-81d0-41391a207008\" (UID: \"b6f74db8-a48c-4358-81d0-41391a207008\") " Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.081162 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6f74db8-a48c-4358-81d0-41391a207008-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6f74db8-a48c-4358-81d0-41391a207008" (UID: "b6f74db8-a48c-4358-81d0-41391a207008"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.081354 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-798b-account-create-update-7g5s7"] Feb 02 12:32:59 crc kubenswrapper[4846]: E0202 12:32:59.107129 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:32:59 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: if [ -n "neutron" ]; then Feb 02 12:32:59 crc kubenswrapper[4846]: GRANT_DATABASE="neutron" Feb 02 12:32:59 crc kubenswrapper[4846]: else Feb 02 12:32:59 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:32:59 crc kubenswrapper[4846]: fi Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:32:59 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:32:59 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:32:59 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:32:59 crc kubenswrapper[4846]: # support updates Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.107423 4846 generic.go:334] "Generic (PLEG): container finished" podID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerID="ceb10096d86d9be3fc00e39352a6fcbaf81ecf7bd346d3e24a99c69fe60cd9f2" exitCode=143 Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.107525 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c452f22-eef8-4e05-a1dd-f6c841833ab4","Type":"ContainerDied","Data":"ceb10096d86d9be3fc00e39352a6fcbaf81ecf7bd346d3e24a99c69fe60cd9f2"} Feb 02 12:32:59 crc kubenswrapper[4846]: E0202 12:32:59.108297 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"neutron-db-secret\\\" not found\"" pod="openstack/neutron-33ff-account-create-update-vz5ls" podUID="318e1583-4759-4158-9683-14524ad771c2" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.109157 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-f9gxt"] Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.122813 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f74db8-a48c-4358-81d0-41391a207008-kube-api-access-q2sw2" (OuterVolumeSpecName: "kube-api-access-q2sw2") pod "b6f74db8-a48c-4358-81d0-41391a207008" (UID: "b6f74db8-a48c-4358-81d0-41391a207008"). InnerVolumeSpecName "kube-api-access-q2sw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.123017 4846 scope.go:117] "RemoveContainer" containerID="ff0b192a1def5dc4bf6d75c57899e36756b63762315f1372301913fb0bf56835" Feb 02 12:32:59 crc kubenswrapper[4846]: E0202 12:32:59.123841 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:32:59 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: if [ -n "nova_api" ]; then Feb 02 12:32:59 crc kubenswrapper[4846]: GRANT_DATABASE="nova_api" Feb 02 12:32:59 crc kubenswrapper[4846]: else Feb 02 12:32:59 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:32:59 crc kubenswrapper[4846]: fi Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:32:59 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:32:59 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:32:59 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:32:59 crc kubenswrapper[4846]: # support updates Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:32:59 crc kubenswrapper[4846]: E0202 12:32:59.125047 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"nova-api-db-secret\\\" not found\"" pod="openstack/nova-api-798b-account-create-update-7g5s7" podUID="73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.126310 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 12:32:59 crc kubenswrapper[4846]: E0202 12:32:59.173493 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:32:59 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: if [ -n "nova_cell0" ]; then Feb 02 12:32:59 crc kubenswrapper[4846]: GRANT_DATABASE="nova_cell0" Feb 02 12:32:59 crc kubenswrapper[4846]: else Feb 02 12:32:59 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:32:59 crc kubenswrapper[4846]: fi Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:32:59 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:32:59 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:32:59 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:32:59 crc kubenswrapper[4846]: # support updates Feb 02 12:32:59 crc kubenswrapper[4846]: Feb 02 12:32:59 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:32:59 crc kubenswrapper[4846]: E0202 12:32:59.181543 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"nova-cell0-db-secret\\\" not found\"" pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" podUID="1d163475-4fe6-4992-90ae-4cfc97a1e6cd" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.181746 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6f74db8-a48c-4358-81d0-41391a207008-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:32:59 crc kubenswrapper[4846]: I0202 12:32:59.181765 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2sw2\" (UniqueName: \"kubernetes.io/projected/b6f74db8-a48c-4358-81d0-41391a207008-kube-api-access-q2sw2\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.198221 4846 generic.go:334] "Generic (PLEG): container finished" podID="fa1b6480-161c-433d-975b-83079f2c979a" containerID="c50d7d730873e682af6be480671b01b55a5ad561b1a9bbf213963fb50c01e8d1" exitCode=0 Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.198305 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fa1b6480-161c-433d-975b-83079f2c979a","Type":"ContainerDied","Data":"c50d7d730873e682af6be480671b01b55a5ad561b1a9bbf213963fb50c01e8d1"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.224220 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.230519 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.271274 4846 scope.go:117] "RemoveContainer" containerID="b4630774f6b8174e6eaa2407dc38ee14e1e53c6776c23ce799141cd61d5f6c49" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.409612 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-z5lnl"] Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.410378 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd01ba51-8a4c-4b74-a7ab-382600f93e68" containerName="openstack-network-exporter" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410406 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd01ba51-8a4c-4b74-a7ab-382600f93e68" containerName="openstack-network-exporter" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.410439 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bda23288-2fee-45ad-81e2-a14654c16589" containerName="ovn-controller" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410447 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bda23288-2fee-45ad-81e2-a14654c16589" containerName="ovn-controller" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.410487 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36061d30-6265-47ca-8a87-5275f41d03ed" containerName="openstack-network-exporter" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410496 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="36061d30-6265-47ca-8a87-5275f41d03ed" containerName="openstack-network-exporter" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.410507 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerName="ovsdbserver-sb" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410513 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerName="ovsdbserver-sb" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.410525 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerName="openstack-network-exporter" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410532 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerName="openstack-network-exporter" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.410546 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36061d30-6265-47ca-8a87-5275f41d03ed" containerName="ovsdbserver-nb" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410568 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="36061d30-6265-47ca-8a87-5275f41d03ed" containerName="ovsdbserver-nb" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.410583 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7187614-7a06-41a1-9835-d309ceb1ac9f" containerName="dnsmasq-dns" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410588 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7187614-7a06-41a1-9835-d309ceb1ac9f" containerName="dnsmasq-dns" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.410595 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7187614-7a06-41a1-9835-d309ceb1ac9f" containerName="init" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410603 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7187614-7a06-41a1-9835-d309ceb1ac9f" containerName="init" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410856 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerName="openstack-network-exporter" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410897 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd01ba51-8a4c-4b74-a7ab-382600f93e68" containerName="openstack-network-exporter" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410907 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7187614-7a06-41a1-9835-d309ceb1ac9f" containerName="dnsmasq-dns" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410918 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bda23288-2fee-45ad-81e2-a14654c16589" containerName="ovn-controller" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410928 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="36061d30-6265-47ca-8a87-5275f41d03ed" containerName="ovsdbserver-nb" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410938 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="36061d30-6265-47ca-8a87-5275f41d03ed" containerName="openstack-network-exporter" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.410944 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="797602a8-7d50-4ab4-8ba5-4afff6a34409" containerName="ovsdbserver-sb" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.411801 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-z5lnl" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.415885 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.419978 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-z5lnl"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.488411 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgttm\" (UniqueName: \"kubernetes.io/projected/473fcbe1-028d-47ad-8c91-18b00e845232-kube-api-access-jgttm\") pod \"root-account-create-update-z5lnl\" (UID: \"473fcbe1-028d-47ad-8c91-18b00e845232\") " pod="openstack/root-account-create-update-z5lnl" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.488476 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/473fcbe1-028d-47ad-8c91-18b00e845232-operator-scripts\") pod \"root-account-create-update-z5lnl\" (UID: \"473fcbe1-028d-47ad-8c91-18b00e845232\") " pod="openstack/root-account-create-update-z5lnl" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.526911 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="203788b8-50cf-402c-afb3-a546b4753707" path="/var/lib/kubelet/pods/203788b8-50cf-402c-afb3-a546b4753707/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.527696 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36061d30-6265-47ca-8a87-5275f41d03ed" path="/var/lib/kubelet/pods/36061d30-6265-47ca-8a87-5275f41d03ed/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.528828 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="797602a8-7d50-4ab4-8ba5-4afff6a34409" path="/var/lib/kubelet/pods/797602a8-7d50-4ab4-8ba5-4afff6a34409/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.530377 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acd7f59a-b591-4a5c-b183-7006ae325ed8" path="/var/lib/kubelet/pods/acd7f59a-b591-4a5c-b183-7006ae325ed8/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.530989 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7187614-7a06-41a1-9835-d309ceb1ac9f" path="/var/lib/kubelet/pods/b7187614-7a06-41a1-9835-d309ceb1ac9f/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.532937 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba822b20-88d5-404b-850c-7f9441933baf" path="/var/lib/kubelet/pods/ba822b20-88d5-404b-850c-7f9441933baf/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.533616 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bda23288-2fee-45ad-81e2-a14654c16589" path="/var/lib/kubelet/pods/bda23288-2fee-45ad-81e2-a14654c16589/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.536158 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1269aae-c267-46d0-8adb-cdbe01430a31" path="/var/lib/kubelet/pods/c1269aae-c267-46d0-8adb-cdbe01430a31/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.536900 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c146f388-9b5f-4340-a68d-4a6f32d634cc" path="/var/lib/kubelet/pods/c146f388-9b5f-4340-a68d-4a6f32d634cc/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.546882 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db0a7be1-2077-4c1b-967c-4b767e5d6a51" path="/var/lib/kubelet/pods/db0a7be1-2077-4c1b-967c-4b767e5d6a51/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.565231 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db37c129-8a93-4f2b-a21b-abe8565d1fee" path="/var/lib/kubelet/pods/db37c129-8a93-4f2b-a21b-abe8565d1fee/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.567709 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd01ba51-8a4c-4b74-a7ab-382600f93e68" path="/var/lib/kubelet/pods/fd01ba51-8a4c-4b74-a7ab-382600f93e68/volumes" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.590930 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/473fcbe1-028d-47ad-8c91-18b00e845232-operator-scripts\") pod \"root-account-create-update-z5lnl\" (UID: \"473fcbe1-028d-47ad-8c91-18b00e845232\") " pod="openstack/root-account-create-update-z5lnl" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.591101 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgttm\" (UniqueName: \"kubernetes.io/projected/473fcbe1-028d-47ad-8c91-18b00e845232-kube-api-access-jgttm\") pod \"root-account-create-update-z5lnl\" (UID: \"473fcbe1-028d-47ad-8c91-18b00e845232\") " pod="openstack/root-account-create-update-z5lnl" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.596321 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/473fcbe1-028d-47ad-8c91-18b00e845232-operator-scripts\") pod \"root-account-create-update-z5lnl\" (UID: \"473fcbe1-028d-47ad-8c91-18b00e845232\") " pod="openstack/root-account-create-update-z5lnl" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.659323 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.669161 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgttm\" (UniqueName: \"kubernetes.io/projected/473fcbe1-028d-47ad-8c91-18b00e845232-kube-api-access-jgttm\") pod \"root-account-create-update-z5lnl\" (UID: \"473fcbe1-028d-47ad-8c91-18b00e845232\") " pod="openstack/root-account-create-update-z5lnl" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.763999 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f94a-account-create-update-zg8j8"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.770304 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-f94a-account-create-update-zg8j8"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.776759 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-68f9885b6f-c5hz4"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.779287 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-68f9885b6f-c5hz4" podUID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" containerName="proxy-httpd" containerID="cri-o://a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce" gracePeriod=30 Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.779386 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-68f9885b6f-c5hz4" podUID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" containerName="proxy-server" containerID="cri-o://fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d" gracePeriod=30 Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.797993 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-vencrypt-tls-certs\") pod \"fa1b6480-161c-433d-975b-83079f2c979a\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.798101 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-config-data\") pod \"fa1b6480-161c-433d-975b-83079f2c979a\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.798174 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-combined-ca-bundle\") pod \"fa1b6480-161c-433d-975b-83079f2c979a\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.798429 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-nova-novncproxy-tls-certs\") pod \"fa1b6480-161c-433d-975b-83079f2c979a\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.799342 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfbtg\" (UniqueName: \"kubernetes.io/projected/fa1b6480-161c-433d-975b-83079f2c979a-kube-api-access-hfbtg\") pod \"fa1b6480-161c-433d-975b-83079f2c979a\" (UID: \"fa1b6480-161c-433d-975b-83079f2c979a\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.817990 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa1b6480-161c-433d-975b-83079f2c979a-kube-api-access-hfbtg" (OuterVolumeSpecName: "kube-api-access-hfbtg") pod "fa1b6480-161c-433d-975b-83079f2c979a" (UID: "fa1b6480-161c-433d-975b-83079f2c979a"). InnerVolumeSpecName "kube-api-access-hfbtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.822365 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.822598 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.822804 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.822832 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.827349 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.828940 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.829697 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-config-data" (OuterVolumeSpecName: "config-data") pod "fa1b6480-161c-433d-975b-83079f2c979a" (UID: "fa1b6480-161c-433d-975b-83079f2c979a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.831199 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:32:59.831506 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovs-vswitchd" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.832031 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa1b6480-161c-433d-975b-83079f2c979a" (UID: "fa1b6480-161c-433d-975b-83079f2c979a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.878113 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-z5lnl" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.883903 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "fa1b6480-161c-433d-975b-83079f2c979a" (UID: "fa1b6480-161c-433d-975b-83079f2c979a"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.890722 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "fa1b6480-161c-433d-975b-83079f2c979a" (UID: "fa1b6480-161c-433d-975b-83079f2c979a"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.903716 4846 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.903743 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfbtg\" (UniqueName: \"kubernetes.io/projected/fa1b6480-161c-433d-975b-83079f2c979a-kube-api-access-hfbtg\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.903755 4846 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.903765 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:32:59.903777 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1b6480-161c-433d-975b-83079f2c979a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.005205 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.005266 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data podName:9cec353a-92f9-4ed4-8ec9-15271cbc6912 nodeName:}" failed. No retries permitted until 2026-02-02 12:33:04.005251287 +0000 UTC m=+1415.233838150 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data") pod "rabbitmq-server-0" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912") : configmap "rabbitmq-config-data" not found Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.125456 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda322cd4_f770_42d7_af2f_e4ed438da7dd.slice/crio-77c5ca7cfef457fcaadda875671e3c5d3ab8c7606c3409257026722b9bd9fa0f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef1ed9cc_3a1a_45cb_ba25_2db484fe129b.slice/crio-conmon-a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef1ed9cc_3a1a_45cb_ba25_2db484fe129b.slice/crio-a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce.scope\": RecentStats: unable to find data in memory cache]" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.211945 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rr425" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.220360 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8 is running failed: container process not found" containerID="24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.221819 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cd5d-account-create-update-v79s9" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.221976 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8 is running failed: container process not found" containerID="24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.225750 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0f2-account-create-update-g95vn" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.225818 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8 is running failed: container process not found" containerID="24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.225849 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7f25197c-8c4e-41ef-b7ef-1ed0f141599b" containerName="nova-scheduler-scheduler" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.275523 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-798b-account-create-update-7g5s7" event={"ID":"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb","Type":"ContainerStarted","Data":"b51ae7d316572bd9bb8b2caef81d62cbbf1773f72fbdde87be3f35e6b6d49d9f"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.306673 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-33ff-account-create-update-vz5ls" event={"ID":"318e1583-4759-4158-9683-14524ad771c2","Type":"ContainerStarted","Data":"cf7929b90b2dd2c0b5b32b34bc23a56b137004cd57f29f4d886f2aed29775599"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.308579 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rr425" event={"ID":"71584556-0f8c-4781-bbd6-2e3f6913f0e1","Type":"ContainerDied","Data":"babc16a10569ff1d331b10fa3f112c4f0c1c3761793bc6bca6f0a2b9212490c2"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.308713 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rr425" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.309179 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71584556-0f8c-4781-bbd6-2e3f6913f0e1-operator-scripts\") pod \"71584556-0f8c-4781-bbd6-2e3f6913f0e1\" (UID: \"71584556-0f8c-4781-bbd6-2e3f6913f0e1\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.309241 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gswh5\" (UniqueName: \"kubernetes.io/projected/db660713-b11b-416f-9fb6-075facb7ecf0-kube-api-access-gswh5\") pod \"db660713-b11b-416f-9fb6-075facb7ecf0\" (UID: \"db660713-b11b-416f-9fb6-075facb7ecf0\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.309320 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db660713-b11b-416f-9fb6-075facb7ecf0-operator-scripts\") pod \"db660713-b11b-416f-9fb6-075facb7ecf0\" (UID: \"db660713-b11b-416f-9fb6-075facb7ecf0\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.309449 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm67h\" (UniqueName: \"kubernetes.io/projected/71584556-0f8c-4781-bbd6-2e3f6913f0e1-kube-api-access-bm67h\") pod \"71584556-0f8c-4781-bbd6-2e3f6913f0e1\" (UID: \"71584556-0f8c-4781-bbd6-2e3f6913f0e1\") " Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.309957 4846 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.310005 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:04.309992274 +0000 UTC m=+1415.538579137 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scripts" not found Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.310676 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db660713-b11b-416f-9fb6-075facb7ecf0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db660713-b11b-416f-9fb6-075facb7ecf0" (UID: "db660713-b11b-416f-9fb6-075facb7ecf0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.311818 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71584556-0f8c-4781-bbd6-2e3f6913f0e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "71584556-0f8c-4781-bbd6-2e3f6913f0e1" (UID: "71584556-0f8c-4781-bbd6-2e3f6913f0e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.311948 4846 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.312007 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:04.311986612 +0000 UTC m=+1415.540573545 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-config-data" not found Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.312066 4846 secret.go:188] Couldn't get secret openstack/cinder-scheduler-config-data: secret "cinder-scheduler-config-data" not found Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.312107 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:04.312094866 +0000 UTC m=+1415.540681839 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scheduler-config-data" not found Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.318358 4846 generic.go:334] "Generic (PLEG): container finished" podID="7f25197c-8c4e-41ef-b7ef-1ed0f141599b" containerID="24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8" exitCode=0 Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.318448 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7f25197c-8c4e-41ef-b7ef-1ed0f141599b","Type":"ContainerDied","Data":"24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.322874 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71584556-0f8c-4781-bbd6-2e3f6913f0e1-kube-api-access-bm67h" (OuterVolumeSpecName: "kube-api-access-bm67h") pod "71584556-0f8c-4781-bbd6-2e3f6913f0e1" (UID: "71584556-0f8c-4781-bbd6-2e3f6913f0e1"). InnerVolumeSpecName "kube-api-access-bm67h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.323366 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db660713-b11b-416f-9fb6-075facb7ecf0-kube-api-access-gswh5" (OuterVolumeSpecName: "kube-api-access-gswh5") pod "db660713-b11b-416f-9fb6-075facb7ecf0" (UID: "db660713-b11b-416f-9fb6-075facb7ecf0"). InnerVolumeSpecName "kube-api-access-gswh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.325011 4846 generic.go:334] "Generic (PLEG): container finished" podID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" containerID="a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce" exitCode=0 Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.325132 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68f9885b6f-c5hz4" event={"ID":"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b","Type":"ContainerDied","Data":"a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.329487 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" event={"ID":"1d163475-4fe6-4992-90ae-4cfc97a1e6cd","Type":"ContainerStarted","Data":"42b243132b2d1d210226bd08fb9e824d701a669be01a1db187b2d404a2b2f1de"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.344223 4846 generic.go:334] "Generic (PLEG): container finished" podID="da322cd4-f770-42d7-af2f-e4ed438da7dd" containerID="77c5ca7cfef457fcaadda875671e3c5d3ab8c7606c3409257026722b9bd9fa0f" exitCode=0 Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.344487 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9447b957-4tn8l" event={"ID":"da322cd4-f770-42d7-af2f-e4ed438da7dd","Type":"ContainerDied","Data":"77c5ca7cfef457fcaadda875671e3c5d3ab8c7606c3409257026722b9bd9fa0f"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.349963 4846 generic.go:334] "Generic (PLEG): container finished" podID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" containerID="319bed97e3756574fc101f5461b63927d20d9217f036ffcd758a126ac4c88d35" exitCode=0 Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.350029 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" event={"ID":"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3","Type":"ContainerDied","Data":"319bed97e3756574fc101f5461b63927d20d9217f036ffcd758a126ac4c88d35"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.377661 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0f2-account-create-update-g95vn" event={"ID":"24671593-6e2f-44b9-909f-66b7a794c372","Type":"ContainerDied","Data":"f2547c641705890efbad2385a672a907ef7af23af6a26c3887bab778ddb7184d"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.377726 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0f2-account-create-update-g95vn" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.392063 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="9cec353a-92f9-4ed4-8ec9-15271cbc6912" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.398480 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cd5d-account-create-update-v79s9" event={"ID":"db660713-b11b-416f-9fb6-075facb7ecf0","Type":"ContainerDied","Data":"cbb64985f175d499483f9119475d235ddd1f1e53765727c08160fad1904feab2"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.398980 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cd5d-account-create-update-v79s9" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.415825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m88rs\" (UniqueName: \"kubernetes.io/projected/24671593-6e2f-44b9-909f-66b7a794c372-kube-api-access-m88rs\") pod \"24671593-6e2f-44b9-909f-66b7a794c372\" (UID: \"24671593-6e2f-44b9-909f-66b7a794c372\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.415881 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24671593-6e2f-44b9-909f-66b7a794c372-operator-scripts\") pod \"24671593-6e2f-44b9-909f-66b7a794c372\" (UID: \"24671593-6e2f-44b9-909f-66b7a794c372\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.417298 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db660713-b11b-416f-9fb6-075facb7ecf0-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.417329 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm67h\" (UniqueName: \"kubernetes.io/projected/71584556-0f8c-4781-bbd6-2e3f6913f0e1-kube-api-access-bm67h\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.417344 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71584556-0f8c-4781-bbd6-2e3f6913f0e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.417358 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gswh5\" (UniqueName: \"kubernetes.io/projected/db660713-b11b-416f-9fb6-075facb7ecf0-kube-api-access-gswh5\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.417830 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24671593-6e2f-44b9-909f-66b7a794c372-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "24671593-6e2f-44b9-909f-66b7a794c372" (UID: "24671593-6e2f-44b9-909f-66b7a794c372"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.419158 4846 generic.go:334] "Generic (PLEG): container finished" podID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" containerID="642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210" exitCode=0 Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.419227 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"444e13f0-bc18-4bc9-8d92-af6a379e22ab","Type":"ContainerDied","Data":"642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.421263 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fa1b6480-161c-433d-975b-83079f2c979a","Type":"ContainerDied","Data":"aa453eed133ed193ddd2753ec424f485f7ac743d86340f632580643174b56da3"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.421306 4846 scope.go:117] "RemoveContainer" containerID="c50d7d730873e682af6be480671b01b55a5ad561b1a9bbf213963fb50c01e8d1" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.421471 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.429751 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24671593-6e2f-44b9-909f-66b7a794c372-kube-api-access-m88rs" (OuterVolumeSpecName: "kube-api-access-m88rs") pod "24671593-6e2f-44b9-909f-66b7a794c372" (UID: "24671593-6e2f-44b9-909f-66b7a794c372"). InnerVolumeSpecName "kube-api-access-m88rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.430232 4846 generic.go:334] "Generic (PLEG): container finished" podID="c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" containerID="481554f18c651c06a0af22ab57eb2b59897fae18ef156d9d05b24af20a349d9a" exitCode=0 Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.430278 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54","Type":"ContainerDied","Data":"481554f18c651c06a0af22ab57eb2b59897fae18ef156d9d05b24af20a349d9a"} Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.432042 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c2bc-account-create-update-qz9kb" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.505500 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b89c77647ebfc096292368b4c128270a751702e9c4f3ecf26b326d816dff5fbc" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.519481 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m88rs\" (UniqueName: \"kubernetes.io/projected/24671593-6e2f-44b9-909f-66b7a794c372-kube-api-access-m88rs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.519526 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24671593-6e2f-44b9-909f-66b7a794c372-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.526491 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b89c77647ebfc096292368b4c128270a751702e9c4f3ecf26b326d816dff5fbc" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.535444 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b89c77647ebfc096292368b4c128270a751702e9c4f3ecf26b326d816dff5fbc" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 02 12:33:00 crc kubenswrapper[4846]: E0202 12:33:00.535498 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="d5f78273-c1e2-452f-8bcb-a51c9bc238e4" containerName="nova-cell1-conductor-conductor" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.546410 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.164:8776/healthcheck\": read tcp 10.217.0.2:52348->10.217.0.164:8776: read: connection reset by peer" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.561722 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-cd5d-account-create-update-v79s9"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.568389 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-cd5d-account-create-update-v79s9"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.576008 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.582558 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.624833 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3915f1bc-20d3-4e7d-b627-46406d0d0461-operator-scripts\") pod \"3915f1bc-20d3-4e7d-b627-46406d0d0461\" (UID: \"3915f1bc-20d3-4e7d-b627-46406d0d0461\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.625099 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdgzg\" (UniqueName: \"kubernetes.io/projected/3915f1bc-20d3-4e7d-b627-46406d0d0461-kube-api-access-bdgzg\") pod \"3915f1bc-20d3-4e7d-b627-46406d0d0461\" (UID: \"3915f1bc-20d3-4e7d-b627-46406d0d0461\") " Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.634977 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3915f1bc-20d3-4e7d-b627-46406d0d0461-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3915f1bc-20d3-4e7d-b627-46406d0d0461" (UID: "3915f1bc-20d3-4e7d-b627-46406d0d0461"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.655871 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3915f1bc-20d3-4e7d-b627-46406d0d0461-kube-api-access-bdgzg" (OuterVolumeSpecName: "kube-api-access-bdgzg") pod "3915f1bc-20d3-4e7d-b627-46406d0d0461" (UID: "3915f1bc-20d3-4e7d-b627-46406d0d0461"). InnerVolumeSpecName "kube-api-access-bdgzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.710502 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rr425"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.730014 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-rr425"] Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.731049 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdgzg\" (UniqueName: \"kubernetes.io/projected/3915f1bc-20d3-4e7d-b627-46406d0d0461-kube-api-access-bdgzg\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.731070 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3915f1bc-20d3-4e7d-b627-46406d0d0461-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.751882 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f6447845b-zlnf5" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:43736->10.217.0.166:9311: read: connection reset by peer" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.752207 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f6447845b-zlnf5" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.166:9311/healthcheck\": read tcp 10.217.0.2:43734->10.217.0.166:9311: read: connection reset by peer" Feb 02 12:33:00 crc kubenswrapper[4846]: I0202 12:33:00.803312 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="327c8143-b3ed-4431-a206-d2019ddcb606" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.030836 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": read tcp 10.217.0.2:46632->10.217.0.208:8775: read: connection reset by peer" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.031168 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": read tcp 10.217.0.2:46640->10.217.0.208:8775: read: connection reset by peer" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.085468 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a0f2-account-create-update-g95vn"] Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.094478 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a0f2-account-create-update-g95vn"] Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.101649 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.159838 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.221309 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.259917 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.266045 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m2bw\" (UniqueName: \"kubernetes.io/projected/da322cd4-f770-42d7-af2f-e4ed438da7dd-kube-api-access-9m2bw\") pod \"da322cd4-f770-42d7-af2f-e4ed438da7dd\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.269025 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data\") pod \"da322cd4-f770-42d7-af2f-e4ed438da7dd\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.269203 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-combined-ca-bundle\") pod \"da322cd4-f770-42d7-af2f-e4ed438da7dd\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.272086 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data-custom\") pod \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.272241 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-combined-ca-bundle\") pod \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.319369 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da322cd4-f770-42d7-af2f-e4ed438da7dd-logs\") pod \"da322cd4-f770-42d7-af2f-e4ed438da7dd\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.319436 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w9ln\" (UniqueName: \"kubernetes.io/projected/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-kube-api-access-4w9ln\") pod \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.319650 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data\") pod \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.319685 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-logs\") pod \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\" (UID: \"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.319967 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data-custom\") pod \"da322cd4-f770-42d7-af2f-e4ed438da7dd\" (UID: \"da322cd4-f770-42d7-af2f-e4ed438da7dd\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.327368 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da322cd4-f770-42d7-af2f-e4ed438da7dd-kube-api-access-9m2bw" (OuterVolumeSpecName: "kube-api-access-9m2bw") pod "da322cd4-f770-42d7-af2f-e4ed438da7dd" (UID: "da322cd4-f770-42d7-af2f-e4ed438da7dd"). InnerVolumeSpecName "kube-api-access-9m2bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.328923 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "da322cd4-f770-42d7-af2f-e4ed438da7dd" (UID: "da322cd4-f770-42d7-af2f-e4ed438da7dd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.329205 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da322cd4-f770-42d7-af2f-e4ed438da7dd-logs" (OuterVolumeSpecName: "logs") pod "da322cd4-f770-42d7-af2f-e4ed438da7dd" (UID: "da322cd4-f770-42d7-af2f-e4ed438da7dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.331880 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-kube-api-access-4w9ln" (OuterVolumeSpecName: "kube-api-access-4w9ln") pod "2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" (UID: "2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3"). InnerVolumeSpecName "kube-api-access-4w9ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.335838 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-logs" (OuterVolumeSpecName: "logs") pod "2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" (UID: "2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.353737 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-798b-account-create-update-7g5s7" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.373358 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-33ff-account-create-update-vz5ls" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.395770 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" (UID: "2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.414862 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.424059 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.424107 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-galera-tls-certs\") pod \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.424181 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-combined-ca-bundle\") pod \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.424661 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6m4b\" (UniqueName: \"kubernetes.io/projected/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kube-api-access-x6m4b\") pod \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.424705 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kolla-config\") pod \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.424731 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7prsb\" (UniqueName: \"kubernetes.io/projected/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-kube-api-access-7prsb\") pod \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.424756 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-combined-ca-bundle\") pod \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.425078 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-generated\") pod \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.425527 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-default\") pod \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.426237 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" (UID: "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.429390 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" (UID: "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.429998 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-operator-scripts\") pod \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\" (UID: \"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.430158 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-config-data\") pod \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\" (UID: \"7f25197c-8c4e-41ef-b7ef-1ed0f141599b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.430653 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" (UID: "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.431022 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" (UID: "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.431383 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kube-api-access-x6m4b" (OuterVolumeSpecName: "kube-api-access-x6m4b") pod "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" (UID: "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54"). InnerVolumeSpecName "kube-api-access-x6m4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.432928 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da322cd4-f770-42d7-af2f-e4ed438da7dd-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.433259 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w9ln\" (UniqueName: \"kubernetes.io/projected/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-kube-api-access-4w9ln\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.433367 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.433470 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.433568 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.433688 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m2bw\" (UniqueName: \"kubernetes.io/projected/da322cd4-f770-42d7-af2f-e4ed438da7dd-kube-api-access-9m2bw\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.433784 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6m4b\" (UniqueName: \"kubernetes.io/projected/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kube-api-access-x6m4b\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.433895 4846 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-kolla-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.433984 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.434077 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-generated\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.434165 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-config-data-default\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.450426 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-kube-api-access-7prsb" (OuterVolumeSpecName: "kube-api-access-7prsb") pod "7f25197c-8c4e-41ef-b7ef-1ed0f141599b" (UID: "7f25197c-8c4e-41ef-b7ef-1ed0f141599b"). InnerVolumeSpecName "kube-api-access-7prsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.457078 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-798b-account-create-update-7g5s7" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.460579 4846 generic.go:334] "Generic (PLEG): container finished" podID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerID="afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2" exitCode=0 Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.487968 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da322cd4-f770-42d7-af2f-e4ed438da7dd" (UID: "da322cd4-f770-42d7-af2f-e4ed438da7dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.489006 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.489979 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.491094 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f25197c-8c4e-41ef-b7ef-1ed0f141599b" (UID: "7f25197c-8c4e-41ef-b7ef-1ed0f141599b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.507061 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24671593-6e2f-44b9-909f-66b7a794c372" path="/var/lib/kubelet/pods/24671593-6e2f-44b9-909f-66b7a794c372/volumes" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.507895 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71584556-0f8c-4781-bbd6-2e3f6913f0e1" path="/var/lib/kubelet/pods/71584556-0f8c-4781-bbd6-2e3f6913f0e1/volumes" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.508348 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6f74db8-a48c-4358-81d0-41391a207008" path="/var/lib/kubelet/pods/b6f74db8-a48c-4358-81d0-41391a207008/volumes" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.519773 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" (UID: "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.520716 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" (UID: "2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.525837 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db660713-b11b-416f-9fb6-075facb7ecf0" path="/var/lib/kubelet/pods/db660713-b11b-416f-9fb6-075facb7ecf0/volumes" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.526534 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa1b6480-161c-433d-975b-83079f2c979a" path="/var/lib/kubelet/pods/fa1b6480-161c-433d-975b-83079f2c979a/volumes" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.537570 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-run-httpd\") pod \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.545374 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" (UID: "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.552325 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.554739 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-config-data\") pod \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.554806 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-combined-ca-bundle\") pod \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.554844 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-internal-tls-certs\") pod \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.554879 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-etc-swift\") pod \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.554932 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d4jf\" (UniqueName: \"kubernetes.io/projected/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-kube-api-access-7d4jf\") pod \"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb\" (UID: \"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.554983 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-public-tls-certs\") pod \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555060 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/318e1583-4759-4158-9683-14524ad771c2-operator-scripts\") pod \"318e1583-4759-4158-9683-14524ad771c2\" (UID: \"318e1583-4759-4158-9683-14524ad771c2\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555120 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44x6w\" (UniqueName: \"kubernetes.io/projected/318e1583-4759-4158-9683-14524ad771c2-kube-api-access-44x6w\") pod \"318e1583-4759-4158-9683-14524ad771c2\" (UID: \"318e1583-4759-4158-9683-14524ad771c2\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555173 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-operator-scripts\") pod \"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb\" (UID: \"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555196 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvlnm\" (UniqueName: \"kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-kube-api-access-dvlnm\") pod \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555257 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-log-httpd\") pod \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\" (UID: \"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555886 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555909 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555920 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7prsb\" (UniqueName: \"kubernetes.io/projected/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-kube-api-access-7prsb\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555930 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555938 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.555948 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.556956 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb" (UID: "73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.561103 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data" (OuterVolumeSpecName: "config-data") pod "da322cd4-f770-42d7-af2f-e4ed438da7dd" (UID: "da322cd4-f770-42d7-af2f-e4ed438da7dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.575560 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/318e1583-4759-4158-9683-14524ad771c2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "318e1583-4759-4158-9683-14524ad771c2" (UID: "318e1583-4759-4158-9683-14524ad771c2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.576013 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.583050 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" (UID: "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.584803 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/318e1583-4759-4158-9683-14524ad771c2-kube-api-access-44x6w" (OuterVolumeSpecName: "kube-api-access-44x6w") pod "318e1583-4759-4158-9683-14524ad771c2" (UID: "318e1583-4759-4158-9683-14524ad771c2"). InnerVolumeSpecName "kube-api-access-44x6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.589101 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-kube-api-access-7d4jf" (OuterVolumeSpecName: "kube-api-access-7d4jf") pod "73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb" (UID: "73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb"). InnerVolumeSpecName "kube-api-access-7d4jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.589500 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.589843 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" (UID: "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.607522 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-kube-api-access-dvlnm" (OuterVolumeSpecName: "kube-api-access-dvlnm") pod "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" (UID: "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b"). InnerVolumeSpecName "kube-api-access-dvlnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.653247 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-798b-account-create-update-7g5s7" event={"ID":"73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb","Type":"ContainerDied","Data":"b51ae7d316572bd9bb8b2caef81d62cbbf1773f72fbdde87be3f35e6b6d49d9f"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.653287 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"532236a1-a04d-4b99-9131-b1d16a3d5ed4","Type":"ContainerDied","Data":"afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.653305 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"532236a1-a04d-4b99-9131-b1d16a3d5ed4","Type":"ContainerDied","Data":"6cd6eb31b01b4a0397af1e0fddba60ae869a8fe15c7977a5da77283e4cba9946"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.653315 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c75e5c92-9e5b-4d18-bb6a-d183c18d8a54","Type":"ContainerDied","Data":"cac3239c4d40ee1f49d1191feea2cd4016924088a291fff1445d44df7bc5e639"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.653325 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-df8f98698-ddglj" event={"ID":"2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3","Type":"ContainerDied","Data":"d9e9e45c9a9509e1b26760b56d70a5a4d9c8a2fe419074443d10faf93fcdd5d2"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.653336 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" event={"ID":"1d163475-4fe6-4992-90ae-4cfc97a1e6cd","Type":"ContainerDied","Data":"42b243132b2d1d210226bd08fb9e824d701a669be01a1db187b2d404a2b2f1de"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.653353 4846 scope.go:117] "RemoveContainer" containerID="afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657157 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpgqz\" (UniqueName: \"kubernetes.io/projected/532236a1-a04d-4b99-9131-b1d16a3d5ed4-kube-api-access-bpgqz\") pod \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657249 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-internal-tls-certs\") pod \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657296 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/532236a1-a04d-4b99-9131-b1d16a3d5ed4-logs\") pod \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657310 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/532236a1-a04d-4b99-9131-b1d16a3d5ed4-etc-machine-id\") pod \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657332 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-scripts\") pod \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657417 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-combined-ca-bundle\") pod \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657462 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data\") pod \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657515 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-public-tls-certs\") pod \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657544 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data-custom\") pod \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\" (UID: \"532236a1-a04d-4b99-9131-b1d16a3d5ed4\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657918 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657930 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d4jf\" (UniqueName: \"kubernetes.io/projected/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-kube-api-access-7d4jf\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657940 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/318e1583-4759-4158-9683-14524ad771c2-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657950 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44x6w\" (UniqueName: \"kubernetes.io/projected/318e1583-4759-4158-9683-14524ad771c2-kube-api-access-44x6w\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657959 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657969 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvlnm\" (UniqueName: \"kubernetes.io/projected/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-kube-api-access-dvlnm\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657977 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.657988 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da322cd4-f770-42d7-af2f-e4ed438da7dd-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.685665 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/532236a1-a04d-4b99-9131-b1d16a3d5ed4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "532236a1-a04d-4b99-9131-b1d16a3d5ed4" (UID: "532236a1-a04d-4b99-9131-b1d16a3d5ed4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.732897 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9447b957-4tn8l" event={"ID":"da322cd4-f770-42d7-af2f-e4ed438da7dd","Type":"ContainerDied","Data":"70d06d2b49e642d48d21ff3ba2fbf33bfa031c755653a0cdd03546e05a3866d0"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.733012 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5d9447b957-4tn8l" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.737131 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/532236a1-a04d-4b99-9131-b1d16a3d5ed4-logs" (OuterVolumeSpecName: "logs") pod "532236a1-a04d-4b99-9131-b1d16a3d5ed4" (UID: "532236a1-a04d-4b99-9131-b1d16a3d5ed4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.753064 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/532236a1-a04d-4b99-9131-b1d16a3d5ed4-kube-api-access-bpgqz" (OuterVolumeSpecName: "kube-api-access-bpgqz") pod "532236a1-a04d-4b99-9131-b1d16a3d5ed4" (UID: "532236a1-a04d-4b99-9131-b1d16a3d5ed4"). InnerVolumeSpecName "kube-api-access-bpgqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.766291 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-internal-tls-certs\") pod \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.766349 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-httpd-run\") pod \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.766419 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-config-data\") pod \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.766479 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.766526 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-scripts\") pod \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.766564 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-logs\") pod \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.766580 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhmqx\" (UniqueName: \"kubernetes.io/projected/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-kube-api-access-dhmqx\") pod \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.766615 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x59n9\" (UniqueName: \"kubernetes.io/projected/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-kube-api-access-x59n9\") pod \"1d163475-4fe6-4992-90ae-4cfc97a1e6cd\" (UID: \"1d163475-4fe6-4992-90ae-4cfc97a1e6cd\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.766686 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-combined-ca-bundle\") pod \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\" (UID: \"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.766716 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-operator-scripts\") pod \"1d163475-4fe6-4992-90ae-4cfc97a1e6cd\" (UID: \"1d163475-4fe6-4992-90ae-4cfc97a1e6cd\") " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.767116 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpgqz\" (UniqueName: \"kubernetes.io/projected/532236a1-a04d-4b99-9131-b1d16a3d5ed4-kube-api-access-bpgqz\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.767131 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/532236a1-a04d-4b99-9131-b1d16a3d5ed4-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.767143 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/532236a1-a04d-4b99-9131-b1d16a3d5ed4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.779423 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-logs" (OuterVolumeSpecName: "logs") pod "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" (UID: "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.787272 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "532236a1-a04d-4b99-9131-b1d16a3d5ed4" (UID: "532236a1-a04d-4b99-9131-b1d16a3d5ed4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.787358 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" (UID: "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.788041 4846 generic.go:334] "Generic (PLEG): container finished" podID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" containerID="265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f" exitCode=0 Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.788096 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9","Type":"ContainerDied","Data":"265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.788123 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"be88d7e8-f24f-4abf-b26a-e9698b1cd8f9","Type":"ContainerDied","Data":"4bbaef0fedc86edabb38183c4145e745121eaa5ceefb10d69f194519fd2e79eb"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.788176 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.789319 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" (UID: "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.790921 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d163475-4fe6-4992-90ae-4cfc97a1e6cd" (UID: "1d163475-4fe6-4992-90ae-4cfc97a1e6cd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.791159 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-scripts" (OuterVolumeSpecName: "scripts") pod "532236a1-a04d-4b99-9131-b1d16a3d5ed4" (UID: "532236a1-a04d-4b99-9131-b1d16a3d5ed4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: E0202 12:33:01.793857 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a60702e8aab62dc62e3d8fd855c2cb15d42dc1d569a8a5f56df75bb229c8672e" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Feb 02 12:33:01 crc kubenswrapper[4846]: E0202 12:33:01.803977 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a60702e8aab62dc62e3d8fd855c2cb15d42dc1d569a8a5f56df75bb229c8672e" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Feb 02 12:33:01 crc kubenswrapper[4846]: E0202 12:33:01.808560 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a60702e8aab62dc62e3d8fd855c2cb15d42dc1d569a8a5f56df75bb229c8672e" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Feb 02 12:33:01 crc kubenswrapper[4846]: E0202 12:33:01.808612 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerName="ovn-northd" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.809323 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-33ff-account-create-update-vz5ls" event={"ID":"318e1583-4759-4158-9683-14524ad771c2","Type":"ContainerDied","Data":"cf7929b90b2dd2c0b5b32b34bc23a56b137004cd57f29f4d886f2aed29775599"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.809475 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-33ff-account-create-update-vz5ls" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.818019 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-kube-api-access-dhmqx" (OuterVolumeSpecName: "kube-api-access-dhmqx") pod "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" (UID: "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9"). InnerVolumeSpecName "kube-api-access-dhmqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.821882 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-kube-api-access-x59n9" (OuterVolumeSpecName: "kube-api-access-x59n9") pod "1d163475-4fe6-4992-90ae-4cfc97a1e6cd" (UID: "1d163475-4fe6-4992-90ae-4cfc97a1e6cd"). InnerVolumeSpecName "kube-api-access-x59n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.833915 4846 generic.go:334] "Generic (PLEG): container finished" podID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerID="99b0a8947603d649bb9d1f3decf69b9ee0004e7f9c373151604124884a5d8445" exitCode=0 Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.834035 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c452f22-eef8-4e05-a1dd-f6c841833ab4","Type":"ContainerDied","Data":"99b0a8947603d649bb9d1f3decf69b9ee0004e7f9c373151604124884a5d8445"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.842246 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.842601 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7f25197c-8c4e-41ef-b7ef-1ed0f141599b","Type":"ContainerDied","Data":"585dc97a1f45b5b4dabb82cd33a5932a469364646b448d0b8364bf69db252fa3"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.843857 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-scripts" (OuterVolumeSpecName: "scripts") pod "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" (UID: "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.855012 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-config-data" (OuterVolumeSpecName: "config-data") pod "7f25197c-8c4e-41ef-b7ef-1ed0f141599b" (UID: "7f25197c-8c4e-41ef-b7ef-1ed0f141599b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.855117 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" (UID: "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.858185 4846 generic.go:334] "Generic (PLEG): container finished" podID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" containerID="fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d" exitCode=0 Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.858266 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68f9885b6f-c5hz4" event={"ID":"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b","Type":"ContainerDied","Data":"fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.858308 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68f9885b6f-c5hz4" event={"ID":"ef1ed9cc-3a1a-45cb-ba25-2db484fe129b","Type":"ContainerDied","Data":"efff0ccd3dbde79a2cc68d31b0063ea27185ac79942fa5cddcebdbd453a3635e"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.858403 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-68f9885b6f-c5hz4" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.864594 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c2bc-account-create-update-qz9kb" event={"ID":"3915f1bc-20d3-4e7d-b627-46406d0d0461","Type":"ContainerDied","Data":"327a79df04532b0e25d4d566df398944e9876a11015e055cd7ba2ea7dc899bbb"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.864723 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c2bc-account-create-update-qz9kb" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875735 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875764 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875776 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhmqx\" (UniqueName: \"kubernetes.io/projected/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-kube-api-access-dhmqx\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875786 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875797 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x59n9\" (UniqueName: \"kubernetes.io/projected/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-kube-api-access-x59n9\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875807 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d163475-4fe6-4992-90ae-4cfc97a1e6cd-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875817 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f25197c-8c4e-41ef-b7ef-1ed0f141599b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875827 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875837 4846 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875847 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.875871 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.889779 4846 generic.go:334] "Generic (PLEG): container finished" podID="525708d2-15d1-4bfb-8258-bf216ebc4bab" containerID="f7527fb5dfdb0a8afb41a261ae73e292db1a4a3cb90a3972583cde5f54856e3b" exitCode=0 Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.889869 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c75698f54-h7x5d" event={"ID":"525708d2-15d1-4bfb-8258-bf216ebc4bab","Type":"ContainerDied","Data":"f7527fb5dfdb0a8afb41a261ae73e292db1a4a3cb90a3972583cde5f54856e3b"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.894477 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerID="ba2153f30138c7492692991e0000a0e9232db386da8169c4c360432f8ac0b9a2" exitCode=0 Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.894552 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e2d348d-dbbf-4046-b9fb-94f8aab64573","Type":"ContainerDied","Data":"ba2153f30138c7492692991e0000a0e9232db386da8169c4c360432f8ac0b9a2"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.896196 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" (UID: "c75e5c92-9e5b-4d18-bb6a-d183c18d8a54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.907184 4846 generic.go:334] "Generic (PLEG): container finished" podID="271eeb3d-6d83-49c6-9d80-75b2276a304a" containerID="dbd55a0c1904df42f4b3eb5b57e2aa4fadf58f7e576128bd91743f3ff542702f" exitCode=0 Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.907269 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"271eeb3d-6d83-49c6-9d80-75b2276a304a","Type":"ContainerDied","Data":"dbd55a0c1904df42f4b3eb5b57e2aa4fadf58f7e576128bd91743f3ff542702f"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.911511 4846 generic.go:334] "Generic (PLEG): container finished" podID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerID="b358fa2e3b3d3040b293911322aa36237a5340d9d50ce3922695a3ff266a9b44" exitCode=0 Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.911562 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6447845b-zlnf5" event={"ID":"69522beb-f7d5-4694-ada8-0d17d7d3f16b","Type":"ContainerDied","Data":"b358fa2e3b3d3040b293911322aa36237a5340d9d50ce3922695a3ff266a9b44"} Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.912096 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-config-data" (OuterVolumeSpecName: "config-data") pod "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" (UID: "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.955184 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" (UID: "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.990711 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.990751 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:01 crc kubenswrapper[4846]: I0202 12:33:01.990764 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.046755 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.092099 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.155932 4846 scope.go:117] "RemoveContainer" containerID="428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.160651 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.163996 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.173830 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="ceilometer-central-agent" containerID="cri-o://e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea" gracePeriod=30 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.174274 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.174366 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="proxy-httpd" containerID="cri-o://c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e" gracePeriod=30 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.174407 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="sg-core" containerID="cri-o://821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91" gracePeriod=30 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.174450 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="ceilometer-notification-agent" containerID="cri-o://b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f" gracePeriod=30 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.194639 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmmpv\" (UniqueName: \"kubernetes.io/projected/69522beb-f7d5-4694-ada8-0d17d7d3f16b-kube-api-access-hmmpv\") pod \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.194741 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data-custom\") pod \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.194858 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69522beb-f7d5-4694-ada8-0d17d7d3f16b-logs\") pod \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.194958 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-public-tls-certs\") pod \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.194984 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-combined-ca-bundle\") pod \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.195013 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data\") pod \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.196788 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-internal-tls-certs\") pod \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\" (UID: \"69522beb-f7d5-4694-ada8-0d17d7d3f16b\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.198986 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69522beb-f7d5-4694-ada8-0d17d7d3f16b-logs" (OuterVolumeSpecName: "logs") pod "69522beb-f7d5-4694-ada8-0d17d7d3f16b" (UID: "69522beb-f7d5-4694-ada8-0d17d7d3f16b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.210439 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="40dea2331c92832ae0ee4b74bc35edbdedf985815110294b594bde6cb57fb223" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.211680 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data" (OuterVolumeSpecName: "config-data") pod "2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" (UID: "2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.224519 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.224645 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.224658 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69522beb-f7d5-4694-ada8-0d17d7d3f16b-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.229797 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="40dea2331c92832ae0ee4b74bc35edbdedf985815110294b594bde6cb57fb223" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.243793 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" (UID: "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.244189 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="40dea2331c92832ae0ee4b74bc35edbdedf985815110294b594bde6cb57fb223" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.244228 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="c7dd2f87-cea2-4575-ad0c-483944433a55" containerName="nova-cell0-conductor-conductor" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.255883 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.261388 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.261572 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="9a52309a-c6e4-481a-9fba-69de4c57e6b1" containerName="kube-state-metrics" containerID="cri-o://218c4aa156faa3e69e561faabd794bbb67a15863d0a9d773a8b44e2e31cb4a4c" gracePeriod=30 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.264761 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "532236a1-a04d-4b99-9131-b1d16a3d5ed4" (UID: "532236a1-a04d-4b99-9131-b1d16a3d5ed4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.282236 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5b76-account-create-update-gm2qj"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.287878 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69522beb-f7d5-4694-ada8-0d17d7d3f16b-kube-api-access-hmmpv" (OuterVolumeSpecName: "kube-api-access-hmmpv") pod "69522beb-f7d5-4694-ada8-0d17d7d3f16b" (UID: "69522beb-f7d5-4694-ada8-0d17d7d3f16b"). InnerVolumeSpecName "kube-api-access-hmmpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.288021 4846 scope.go:117] "RemoveContainer" containerID="afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.289061 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2\": container with ID starting with afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2 not found: ID does not exist" containerID="afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.289089 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2"} err="failed to get container status \"afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2\": rpc error: code = NotFound desc = could not find container \"afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2\": container with ID starting with afda29ffff158e1fe45c09447ad73ac2a8b1fe4c7d1e0534df183d9a56fd47b2 not found: ID does not exist" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.289108 4846 scope.go:117] "RemoveContainer" containerID="428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.289897 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "69522beb-f7d5-4694-ada8-0d17d7d3f16b" (UID: "69522beb-f7d5-4694-ada8-0d17d7d3f16b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.289613 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646\": container with ID starting with 428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646 not found: ID does not exist" containerID="428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.291680 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646"} err="failed to get container status \"428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646\": rpc error: code = NotFound desc = could not find container \"428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646\": container with ID starting with 428ca40ab70b2455aae33655594412b13e6f6f8b0102c5a77f77b4d0225b7646 not found: ID does not exist" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.291696 4846 scope.go:117] "RemoveContainer" containerID="481554f18c651c06a0af22ab57eb2b59897fae18ef156d9d05b24af20a349d9a" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.321725 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.325460 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-combined-ca-bundle\") pod \"525708d2-15d1-4bfb-8258-bf216ebc4bab\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.325504 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/525708d2-15d1-4bfb-8258-bf216ebc4bab-logs\") pod \"525708d2-15d1-4bfb-8258-bf216ebc4bab\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.325536 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-internal-tls-certs\") pod \"525708d2-15d1-4bfb-8258-bf216ebc4bab\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.325686 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-config-data\") pod \"525708d2-15d1-4bfb-8258-bf216ebc4bab\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.325754 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ps9lh\" (UniqueName: \"kubernetes.io/projected/525708d2-15d1-4bfb-8258-bf216ebc4bab-kube-api-access-ps9lh\") pod \"525708d2-15d1-4bfb-8258-bf216ebc4bab\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.325823 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-public-tls-certs\") pod \"525708d2-15d1-4bfb-8258-bf216ebc4bab\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.325929 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-scripts\") pod \"525708d2-15d1-4bfb-8258-bf216ebc4bab\" (UID: \"525708d2-15d1-4bfb-8258-bf216ebc4bab\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.327713 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/525708d2-15d1-4bfb-8258-bf216ebc4bab-logs" (OuterVolumeSpecName: "logs") pod "525708d2-15d1-4bfb-8258-bf216ebc4bab" (UID: "525708d2-15d1-4bfb-8258-bf216ebc4bab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.331827 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.331853 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.331863 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmmpv\" (UniqueName: \"kubernetes.io/projected/69522beb-f7d5-4694-ada8-0d17d7d3f16b-kube-api-access-hmmpv\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.331873 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.331883 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/525708d2-15d1-4bfb-8258-bf216ebc4bab-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.349864 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "532236a1-a04d-4b99-9131-b1d16a3d5ed4" (UID: "532236a1-a04d-4b99-9131-b1d16a3d5ed4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.349911 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-scripts" (OuterVolumeSpecName: "scripts") pod "525708d2-15d1-4bfb-8258-bf216ebc4bab" (UID: "525708d2-15d1-4bfb-8258-bf216ebc4bab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.351578 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.351866 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="a7820c87-e5cc-42fd-b3f4-551cd39d3a15" containerName="memcached" containerID="cri-o://20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997" gracePeriod=30 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380056 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5b76-account-create-update-9h9tt"] Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380410 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerName="cinder-api" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380432 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerName="cinder-api" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380451 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" containerName="proxy-httpd" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380458 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" containerName="proxy-httpd" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380477 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="525708d2-15d1-4bfb-8258-bf216ebc4bab" containerName="placement-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380485 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="525708d2-15d1-4bfb-8258-bf216ebc4bab" containerName="placement-log" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380503 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380509 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-log" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380523 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerName="barbican-api-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380528 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerName="barbican-api-log" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380542 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" containerName="glance-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380547 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" containerName="glance-log" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380558 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="525708d2-15d1-4bfb-8258-bf216ebc4bab" containerName="placement-api" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380564 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="525708d2-15d1-4bfb-8258-bf216ebc4bab" containerName="placement-api" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380577 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da322cd4-f770-42d7-af2f-e4ed438da7dd" containerName="barbican-worker-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380582 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="da322cd4-f770-42d7-af2f-e4ed438da7dd" containerName="barbican-worker-log" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380591 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-metadata" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380596 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-metadata" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380605 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da322cd4-f770-42d7-af2f-e4ed438da7dd" containerName="barbican-worker" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380692 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="da322cd4-f770-42d7-af2f-e4ed438da7dd" containerName="barbican-worker" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380708 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" containerName="barbican-keystone-listener-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380714 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" containerName="barbican-keystone-listener-log" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380725 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" containerName="barbican-keystone-listener" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380731 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" containerName="barbican-keystone-listener" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380742 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" containerName="glance-httpd" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380747 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" containerName="glance-httpd" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380755 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f25197c-8c4e-41ef-b7ef-1ed0f141599b" containerName="nova-scheduler-scheduler" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380761 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f25197c-8c4e-41ef-b7ef-1ed0f141599b" containerName="nova-scheduler-scheduler" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380770 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerName="cinder-api-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380776 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerName="cinder-api-log" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380784 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerName="barbican-api" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380791 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerName="barbican-api" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380798 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" containerName="galera" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380803 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" containerName="galera" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380813 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1b6480-161c-433d-975b-83079f2c979a" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380818 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1b6480-161c-433d-975b-83079f2c979a" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380831 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" containerName="mysql-bootstrap" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380836 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" containerName="mysql-bootstrap" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.380851 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" containerName="proxy-server" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.380857 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" containerName="proxy-server" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381008 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerName="cinder-api" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381022 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" containerName="cinder-api-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381029 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerName="barbican-api" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381039 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" containerName="barbican-keystone-listener" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381047 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" containerName="barbican-api-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381054 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" containerName="proxy-server" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381061 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-metadata" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381068 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="da322cd4-f770-42d7-af2f-e4ed438da7dd" containerName="barbican-worker" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381080 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa1b6480-161c-433d-975b-83079f2c979a" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381090 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="525708d2-15d1-4bfb-8258-bf216ebc4bab" containerName="placement-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381102 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" containerName="glance-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381114 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" containerName="nova-metadata-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381125 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="525708d2-15d1-4bfb-8258-bf216ebc4bab" containerName="placement-api" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381139 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" containerName="barbican-keystone-listener-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381151 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="da322cd4-f770-42d7-af2f-e4ed438da7dd" containerName="barbican-worker-log" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381167 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" containerName="proxy-httpd" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381181 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" containerName="glance-httpd" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381192 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f25197c-8c4e-41ef-b7ef-1ed0f141599b" containerName="nova-scheduler-scheduler" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381199 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" containerName="galera" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.381837 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.387936 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.393194 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/525708d2-15d1-4bfb-8258-bf216ebc4bab-kube-api-access-ps9lh" (OuterVolumeSpecName: "kube-api-access-ps9lh") pod "525708d2-15d1-4bfb-8258-bf216ebc4bab" (UID: "525708d2-15d1-4bfb-8258-bf216ebc4bab"). InnerVolumeSpecName "kube-api-access-ps9lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.394375 4846 scope.go:117] "RemoveContainer" containerID="eb2f98a1c13c7bf89f3e7b636cf147de5beecff45f5771d9fc6fe245fe4a6f37" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.420800 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5b76-account-create-update-gm2qj"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.433225 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e2d348d-dbbf-4046-b9fb-94f8aab64573-logs\") pod \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.433352 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-config-data\") pod \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.433434 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-nova-metadata-tls-certs\") pod \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.433473 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhbpb\" (UniqueName: \"kubernetes.io/projected/5e2d348d-dbbf-4046-b9fb-94f8aab64573-kube-api-access-xhbpb\") pod \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.433608 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-combined-ca-bundle\") pod \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\" (UID: \"5e2d348d-dbbf-4046-b9fb-94f8aab64573\") " Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.434611 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts\") pod \"keystone-5b76-account-create-update-9h9tt\" (UID: \"7ad24791-f83d-4b1b-a19a-3e3b669b598d\") " pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.434652 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e2d348d-dbbf-4046-b9fb-94f8aab64573-logs" (OuterVolumeSpecName: "logs") pod "5e2d348d-dbbf-4046-b9fb-94f8aab64573" (UID: "5e2d348d-dbbf-4046-b9fb-94f8aab64573"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.435190 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqrpl\" (UniqueName: \"kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl\") pod \"keystone-5b76-account-create-update-9h9tt\" (UID: \"7ad24791-f83d-4b1b-a19a-3e3b669b598d\") " pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.435291 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.435307 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.435317 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e2d348d-dbbf-4046-b9fb-94f8aab64573-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.435326 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ps9lh\" (UniqueName: \"kubernetes.io/projected/525708d2-15d1-4bfb-8258-bf216ebc4bab-kube-api-access-ps9lh\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.445731 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" (UID: "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.502243 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b76-account-create-update-9h9tt"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.503826 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-m2dtn"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.506849 4846 scope.go:117] "RemoveContainer" containerID="319bed97e3756574fc101f5461b63927d20d9217f036ffcd758a126ac4c88d35" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.517330 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xf7xw"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.527991 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-m2dtn"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.537213 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqrpl\" (UniqueName: \"kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl\") pod \"keystone-5b76-account-create-update-9h9tt\" (UID: \"7ad24791-f83d-4b1b-a19a-3e3b669b598d\") " pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.537285 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts\") pod \"keystone-5b76-account-create-update-9h9tt\" (UID: \"7ad24791-f83d-4b1b-a19a-3e3b669b598d\") " pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.537366 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.537439 4846 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.537526 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts podName:7ad24791-f83d-4b1b-a19a-3e3b669b598d nodeName:}" failed. No retries permitted until 2026-02-02 12:33:03.03747294 +0000 UTC m=+1414.266059803 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts") pod "keystone-5b76-account-create-update-9h9tt" (UID: "7ad24791-f83d-4b1b-a19a-3e3b669b598d") : configmap "openstack-scripts" not found Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.541058 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xf7xw"] Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.542427 4846 projected.go:194] Error preparing data for projected volume kube-api-access-tqrpl for pod openstack/keystone-5b76-account-create-update-9h9tt: failed to fetch token: serviceaccounts "galera-openstack" not found Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.542477 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl podName:7ad24791-f83d-4b1b-a19a-3e3b669b598d nodeName:}" failed. No retries permitted until 2026-02-02 12:33:03.042461272 +0000 UTC m=+1414.271048135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-tqrpl" (UniqueName: "kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl") pod "keystone-5b76-account-create-update-9h9tt" (UID: "7ad24791-f83d-4b1b-a19a-3e3b669b598d") : failed to fetch token: serviceaccounts "galera-openstack" not found Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.547639 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.563996 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6fb558fdfb-5z9p5"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.564213 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-6fb558fdfb-5z9p5" podUID="c858d5df-042d-4211-898b-1ad4ef1fea47" containerName="keystone-api" containerID="cri-o://c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced" gracePeriod=30 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.567300 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2d348d-dbbf-4046-b9fb-94f8aab64573-kube-api-access-xhbpb" (OuterVolumeSpecName: "kube-api-access-xhbpb") pod "5e2d348d-dbbf-4046-b9fb-94f8aab64573" (UID: "5e2d348d-dbbf-4046-b9fb-94f8aab64573"). InnerVolumeSpecName "kube-api-access-xhbpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.617212 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c2bc-account-create-update-qz9kb"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.642267 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhbpb\" (UniqueName: \"kubernetes.io/projected/5e2d348d-dbbf-4046-b9fb-94f8aab64573-kube-api-access-xhbpb\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.666752 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-config-data" (OuterVolumeSpecName: "config-data") pod "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" (UID: "be88d7e8-f24f-4abf-b26a-e9698b1cd8f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.667523 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "532236a1-a04d-4b99-9131-b1d16a3d5ed4" (UID: "532236a1-a04d-4b99-9131-b1d16a3d5ed4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.672130 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-config-data" (OuterVolumeSpecName: "config-data") pod "5e2d348d-dbbf-4046-b9fb-94f8aab64573" (UID: "5e2d348d-dbbf-4046-b9fb-94f8aab64573"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.674219 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c2bc-account-create-update-qz9kb"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.683752 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.693383 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-z5lnl"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.703294 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.716019 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-xvdwm"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.742530 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.744435 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e2d348d-dbbf-4046-b9fb-94f8aab64573" (UID: "5e2d348d-dbbf-4046-b9fb-94f8aab64573"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.744471 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.744543 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.744562 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.750444 4846 scope.go:117] "RemoveContainer" containerID="199fdc9812e4067e4c230bdb3772de61e8c1276ecc5f01f740fc1af2fa9d74c4" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.766918 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-xvdwm"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.770252 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69522beb-f7d5-4694-ada8-0d17d7d3f16b" (UID: "69522beb-f7d5-4694-ada8-0d17d7d3f16b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.772446 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.785978 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-5d9447b957-4tn8l"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.812900 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5b76-account-create-update-9h9tt"] Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.813598 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-tqrpl operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone-5b76-account-create-update-9h9tt" podUID="7ad24791-f83d-4b1b-a19a-3e3b669b598d" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.820786 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" (UID: "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.821664 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-5d9447b957-4tn8l"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.824553 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" (UID: "ef1ed9cc-3a1a-45cb-ba25-2db484fe129b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.826806 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data" (OuterVolumeSpecName: "config-data") pod "532236a1-a04d-4b99-9131-b1d16a3d5ed4" (UID: "532236a1-a04d-4b99-9131-b1d16a3d5ed4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.847270 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.847307 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.847320 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/532236a1-a04d-4b99-9131-b1d16a3d5ed4-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.847333 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.847345 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.872432 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-798b-account-create-update-7g5s7"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.883091 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-798b-account-create-update-7g5s7"] Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.899065 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 12:33:02 crc kubenswrapper[4846]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:0f7943e02fbdd3daec1d3db72fa9396bf37ad3fdd6b0f3119c90e29629e095ed,Command:[/bin/sh -c #!/bin/bash Feb 02 12:33:02 crc kubenswrapper[4846]: Feb 02 12:33:02 crc kubenswrapper[4846]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 02 12:33:02 crc kubenswrapper[4846]: Feb 02 12:33:02 crc kubenswrapper[4846]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 02 12:33:02 crc kubenswrapper[4846]: Feb 02 12:33:02 crc kubenswrapper[4846]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 02 12:33:02 crc kubenswrapper[4846]: Feb 02 12:33:02 crc kubenswrapper[4846]: if [ -n "" ]; then Feb 02 12:33:02 crc kubenswrapper[4846]: GRANT_DATABASE="" Feb 02 12:33:02 crc kubenswrapper[4846]: else Feb 02 12:33:02 crc kubenswrapper[4846]: GRANT_DATABASE="*" Feb 02 12:33:02 crc kubenswrapper[4846]: fi Feb 02 12:33:02 crc kubenswrapper[4846]: Feb 02 12:33:02 crc kubenswrapper[4846]: # going for maximum compatibility here: Feb 02 12:33:02 crc kubenswrapper[4846]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 02 12:33:02 crc kubenswrapper[4846]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 02 12:33:02 crc kubenswrapper[4846]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 02 12:33:02 crc kubenswrapper[4846]: # support updates Feb 02 12:33:02 crc kubenswrapper[4846]: Feb 02 12:33:02 crc kubenswrapper[4846]: $MYSQL_CMD < logger="UnhandledError" Feb 02 12:33:02 crc kubenswrapper[4846]: E0202 12:33:02.900394 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"openstack-mariadb-root-db-secret\\\" not found\"" pod="openstack/root-account-create-update-z5lnl" podUID="473fcbe1-028d-47ad-8c91-18b00e845232" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.911652 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-33ff-account-create-update-vz5ls"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.917241 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-33ff-account-create-update-vz5ls"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.928639 4846 generic.go:334] "Generic (PLEG): container finished" podID="9a52309a-c6e4-481a-9fba-69de4c57e6b1" containerID="218c4aa156faa3e69e561faabd794bbb67a15863d0a9d773a8b44e2e31cb4a4c" exitCode=2 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.928654 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9a52309a-c6e4-481a-9fba-69de4c57e6b1","Type":"ContainerDied","Data":"218c4aa156faa3e69e561faabd794bbb67a15863d0a9d773a8b44e2e31cb4a4c"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.930768 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"271eeb3d-6d83-49c6-9d80-75b2276a304a","Type":"ContainerDied","Data":"d8cfb3823d5db38975c107d55a8870660a3a63bc0167de20ac845189945afadc"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.930793 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8cfb3823d5db38975c107d55a8870660a3a63bc0167de20ac845189945afadc" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.931881 4846 generic.go:334] "Generic (PLEG): container finished" podID="d5f78273-c1e2-452f-8bcb-a51c9bc238e4" containerID="b89c77647ebfc096292368b4c128270a751702e9c4f3ecf26b326d816dff5fbc" exitCode=0 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.931917 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d5f78273-c1e2-452f-8bcb-a51c9bc238e4","Type":"ContainerDied","Data":"b89c77647ebfc096292368b4c128270a751702e9c4f3ecf26b326d816dff5fbc"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.931931 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d5f78273-c1e2-452f-8bcb-a51c9bc238e4","Type":"ContainerDied","Data":"8a05e0a9e87160b63b680e8161dfe05acfe7a085f7a20c66e6b6cf52a774d301"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.931941 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a05e0a9e87160b63b680e8161dfe05acfe7a085f7a20c66e6b6cf52a774d301" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.932859 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "69522beb-f7d5-4694-ada8-0d17d7d3f16b" (UID: "69522beb-f7d5-4694-ada8-0d17d7d3f16b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.942608 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-df8f98698-ddglj"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.942848 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3c452f22-eef8-4e05-a1dd-f6c841833ab4","Type":"ContainerDied","Data":"a0d977765916383faa5992a6fbbed67a8ead03c2da665089b66deda768013551"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.942876 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0d977765916383faa5992a6fbbed67a8ead03c2da665089b66deda768013551" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.949126 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.949175 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-df8f98698-ddglj"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.952065 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c75698f54-h7x5d" event={"ID":"525708d2-15d1-4bfb-8258-bf216ebc4bab","Type":"ContainerDied","Data":"359771b6d080e100102b52000fea7f8597862be6e2195eb43d0b8aab5071233c"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.952190 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c75698f54-h7x5d" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.958197 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-z5lnl"] Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.961268 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e2d348d-dbbf-4046-b9fb-94f8aab64573","Type":"ContainerDied","Data":"9e37fc284b289789211d3d4abe18eeeb82d5df4b59f640e5b8743b4bd5f0b315"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.961369 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.963915 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.978615 4846 generic.go:334] "Generic (PLEG): container finished" podID="c7dd2f87-cea2-4575-ad0c-483944433a55" containerID="40dea2331c92832ae0ee4b74bc35edbdedf985815110294b594bde6cb57fb223" exitCode=0 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.978710 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c7dd2f87-cea2-4575-ad0c-483944433a55","Type":"ContainerDied","Data":"40dea2331c92832ae0ee4b74bc35edbdedf985815110294b594bde6cb57fb223"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.979281 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5e2d348d-dbbf-4046-b9fb-94f8aab64573" (UID: "5e2d348d-dbbf-4046-b9fb-94f8aab64573"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.984584 4846 generic.go:334] "Generic (PLEG): container finished" podID="92819684-ecb5-4c4d-950c-edacf38e269b" containerID="c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e" exitCode=0 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.984655 4846 generic.go:334] "Generic (PLEG): container finished" podID="92819684-ecb5-4c4d-950c-edacf38e269b" containerID="821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91" exitCode=2 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.984667 4846 generic.go:334] "Generic (PLEG): container finished" podID="92819684-ecb5-4c4d-950c-edacf38e269b" containerID="e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea" exitCode=0 Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.984714 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92819684-ecb5-4c4d-950c-edacf38e269b","Type":"ContainerDied","Data":"c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.984744 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92819684-ecb5-4c4d-950c-edacf38e269b","Type":"ContainerDied","Data":"821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.984757 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92819684-ecb5-4c4d-950c-edacf38e269b","Type":"ContainerDied","Data":"e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea"} Feb 02 12:33:02 crc kubenswrapper[4846]: I0202 12:33:02.987092 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-z5lnl" event={"ID":"473fcbe1-028d-47ad-8c91-18b00e845232","Type":"ContainerStarted","Data":"1679e6ab821caff4f71f3187a54c2be0303243fecf4dc974a939837ddde53f34"} Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.013279 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.013733 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f6447845b-zlnf5" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.014104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6447845b-zlnf5" event={"ID":"69522beb-f7d5-4694-ada8-0d17d7d3f16b","Type":"ContainerDied","Data":"4c0d152aa8838d6ae8e216419456abc7969ff129785bcc1f4be20cac7b2e770f"} Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.014134 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ad5a-account-create-update-f9gxt" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.015591 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-config-data" (OuterVolumeSpecName: "config-data") pod "525708d2-15d1-4bfb-8258-bf216ebc4bab" (UID: "525708d2-15d1-4bfb-8258-bf216ebc4bab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.017375 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data" (OuterVolumeSpecName: "config-data") pod "69522beb-f7d5-4694-ada8-0d17d7d3f16b" (UID: "69522beb-f7d5-4694-ada8-0d17d7d3f16b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.022498 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "525708d2-15d1-4bfb-8258-bf216ebc4bab" (UID: "525708d2-15d1-4bfb-8258-bf216ebc4bab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.024729 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "525708d2-15d1-4bfb-8258-bf216ebc4bab" (UID: "525708d2-15d1-4bfb-8258-bf216ebc4bab"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.029716 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "69522beb-f7d5-4694-ada8-0d17d7d3f16b" (UID: "69522beb-f7d5-4694-ada8-0d17d7d3f16b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.057380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqrpl\" (UniqueName: \"kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl\") pod \"keystone-5b76-account-create-update-9h9tt\" (UID: \"7ad24791-f83d-4b1b-a19a-3e3b669b598d\") " pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.057456 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts\") pod \"keystone-5b76-account-create-update-9h9tt\" (UID: \"7ad24791-f83d-4b1b-a19a-3e3b669b598d\") " pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.057540 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e2d348d-dbbf-4046-b9fb-94f8aab64573-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.057553 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.057565 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69522beb-f7d5-4694-ada8-0d17d7d3f16b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.057575 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.057589 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.057600 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: E0202 12:33:03.057709 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Feb 02 12:33:03 crc kubenswrapper[4846]: E0202 12:33:03.057764 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data podName:327c8143-b3ed-4431-a206-d2019ddcb606 nodeName:}" failed. No retries permitted until 2026-02-02 12:33:11.057746807 +0000 UTC m=+1422.286333670 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data") pod "rabbitmq-cell1-server-0" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606") : configmap "rabbitmq-cell1-config-data" not found Feb 02 12:33:03 crc kubenswrapper[4846]: E0202 12:33:03.058315 4846 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Feb 02 12:33:03 crc kubenswrapper[4846]: E0202 12:33:03.058380 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts podName:7ad24791-f83d-4b1b-a19a-3e3b669b598d nodeName:}" failed. No retries permitted until 2026-02-02 12:33:04.058368792 +0000 UTC m=+1415.286955655 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts") pod "keystone-5b76-account-create-update-9h9tt" (UID: "7ad24791-f83d-4b1b-a19a-3e3b669b598d") : configmap "openstack-scripts" not found Feb 02 12:33:03 crc kubenswrapper[4846]: E0202 12:33:03.061942 4846 projected.go:194] Error preparing data for projected volume kube-api-access-tqrpl for pod openstack/keystone-5b76-account-create-update-9h9tt: failed to fetch token: serviceaccounts "galera-openstack" not found Feb 02 12:33:03 crc kubenswrapper[4846]: E0202 12:33:03.062023 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl podName:7ad24791-f83d-4b1b-a19a-3e3b669b598d nodeName:}" failed. No retries permitted until 2026-02-02 12:33:04.061999741 +0000 UTC m=+1415.290586674 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-tqrpl" (UniqueName: "kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl") pod "keystone-5b76-account-create-update-9h9tt" (UID: "7ad24791-f83d-4b1b-a19a-3e3b669b598d") : failed to fetch token: serviceaccounts "galera-openstack" not found Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.062539 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" containerName="galera" containerID="cri-o://e484c823759e2a5e61def70b96f67d07c7974e29d5c4f9fc17eee40d86e83c71" gracePeriod=30 Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.101017 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.111034 4846 scope.go:117] "RemoveContainer" containerID="77c5ca7cfef457fcaadda875671e3c5d3ab8c7606c3409257026722b9bd9fa0f" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.117173 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.124353 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.130377 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.135000 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.141915 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "525708d2-15d1-4bfb-8258-bf216ebc4bab" (UID: "525708d2-15d1-4bfb-8258-bf216ebc4bab"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.151745 4846 scope.go:117] "RemoveContainer" containerID="c6820cbfbd66e9d71a05716d17fadd1549cc9619a0b848aee68552415ffa96a1" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.164324 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.164992 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-config-data\") pod \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165046 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jrtf\" (UniqueName: \"kubernetes.io/projected/271eeb3d-6d83-49c6-9d80-75b2276a304a-kube-api-access-7jrtf\") pod \"271eeb3d-6d83-49c6-9d80-75b2276a304a\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165076 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-combined-ca-bundle\") pod \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165106 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpqfm\" (UniqueName: \"kubernetes.io/projected/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-kube-api-access-cpqfm\") pod \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\" (UID: \"d5f78273-c1e2-452f-8bcb-a51c9bc238e4\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165164 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-scripts\") pod \"271eeb3d-6d83-49c6-9d80-75b2276a304a\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165192 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-public-tls-certs\") pod \"271eeb3d-6d83-49c6-9d80-75b2276a304a\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165219 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-logs\") pod \"271eeb3d-6d83-49c6-9d80-75b2276a304a\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165256 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-config-data\") pod \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165272 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-config-data\") pod \"271eeb3d-6d83-49c6-9d80-75b2276a304a\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165287 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c452f22-eef8-4e05-a1dd-f6c841833ab4-logs\") pod \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165305 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"271eeb3d-6d83-49c6-9d80-75b2276a304a\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165377 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-httpd-run\") pod \"271eeb3d-6d83-49c6-9d80-75b2276a304a\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165401 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkjz7\" (UniqueName: \"kubernetes.io/projected/3c452f22-eef8-4e05-a1dd-f6c841833ab4-kube-api-access-xkjz7\") pod \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165418 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-combined-ca-bundle\") pod \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165445 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-combined-ca-bundle\") pod \"271eeb3d-6d83-49c6-9d80-75b2276a304a\" (UID: \"271eeb3d-6d83-49c6-9d80-75b2276a304a\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165468 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-public-tls-certs\") pod \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.165497 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-internal-tls-certs\") pod \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\" (UID: \"3c452f22-eef8-4e05-a1dd-f6c841833ab4\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.169682 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/525708d2-15d1-4bfb-8258-bf216ebc4bab-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.182383 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c452f22-eef8-4e05-a1dd-f6c841833ab4-logs" (OuterVolumeSpecName: "logs") pod "3c452f22-eef8-4e05-a1dd-f6c841833ab4" (UID: "3c452f22-eef8-4e05-a1dd-f6c841833ab4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.183229 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "271eeb3d-6d83-49c6-9d80-75b2276a304a" (UID: "271eeb3d-6d83-49c6-9d80-75b2276a304a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.183657 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-logs" (OuterVolumeSpecName: "logs") pod "271eeb3d-6d83-49c6-9d80-75b2276a304a" (UID: "271eeb3d-6d83-49c6-9d80-75b2276a304a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.201770 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-kube-api-access-cpqfm" (OuterVolumeSpecName: "kube-api-access-cpqfm") pod "d5f78273-c1e2-452f-8bcb-a51c9bc238e4" (UID: "d5f78273-c1e2-452f-8bcb-a51c9bc238e4"). InnerVolumeSpecName "kube-api-access-cpqfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.206053 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c452f22-eef8-4e05-a1dd-f6c841833ab4-kube-api-access-xkjz7" (OuterVolumeSpecName: "kube-api-access-xkjz7") pod "3c452f22-eef8-4e05-a1dd-f6c841833ab4" (UID: "3c452f22-eef8-4e05-a1dd-f6c841833ab4"). InnerVolumeSpecName "kube-api-access-xkjz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.206559 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.209392 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.214886 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-scripts" (OuterVolumeSpecName: "scripts") pod "271eeb3d-6d83-49c6-9d80-75b2276a304a" (UID: "271eeb3d-6d83-49c6-9d80-75b2276a304a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.215000 4846 scope.go:117] "RemoveContainer" containerID="265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.218310 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/271eeb3d-6d83-49c6-9d80-75b2276a304a-kube-api-access-7jrtf" (OuterVolumeSpecName: "kube-api-access-7jrtf") pod "271eeb3d-6d83-49c6-9d80-75b2276a304a" (UID: "271eeb3d-6d83-49c6-9d80-75b2276a304a"). InnerVolumeSpecName "kube-api-access-7jrtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.225110 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-68f9885b6f-c5hz4"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.228952 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "271eeb3d-6d83-49c6-9d80-75b2276a304a" (UID: "271eeb3d-6d83-49c6-9d80-75b2276a304a"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.229119 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.241704 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-68f9885b6f-c5hz4"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.290645 4846 scope.go:117] "RemoveContainer" containerID="7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.295560 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jrtf\" (UniqueName: \"kubernetes.io/projected/271eeb3d-6d83-49c6-9d80-75b2276a304a-kube-api-access-7jrtf\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.300085 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpqfm\" (UniqueName: \"kubernetes.io/projected/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-kube-api-access-cpqfm\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.300192 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.300248 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.300275 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c452f22-eef8-4e05-a1dd-f6c841833ab4-logs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.300319 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.303195 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/271eeb3d-6d83-49c6-9d80-75b2276a304a-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.303222 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkjz7\" (UniqueName: \"kubernetes.io/projected/3c452f22-eef8-4e05-a1dd-f6c841833ab4-kube-api-access-xkjz7\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.338795 4846 scope.go:117] "RemoveContainer" containerID="265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f" Feb 02 12:33:03 crc kubenswrapper[4846]: E0202 12:33:03.339979 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f\": container with ID starting with 265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f not found: ID does not exist" containerID="265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.340023 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f"} err="failed to get container status \"265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f\": rpc error: code = NotFound desc = could not find container \"265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f\": container with ID starting with 265bd542a9bccc71f3fe24bb23a24e3b79b237bd018ddec3ed2667f4eb80177f not found: ID does not exist" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.340050 4846 scope.go:117] "RemoveContainer" containerID="7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686" Feb 02 12:33:03 crc kubenswrapper[4846]: E0202 12:33:03.340501 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686\": container with ID starting with 7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686 not found: ID does not exist" containerID="7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.340531 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686"} err="failed to get container status \"7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686\": rpc error: code = NotFound desc = could not find container \"7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686\": container with ID starting with 7dd59b58f5b096c8564d821ea06410bc2b28a0ee055a3b1145753772f4f20686 not found: ID does not exist" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.340555 4846 scope.go:117] "RemoveContainer" containerID="24fc757cb85af638648f312e50d71e7b54a3b6acbf927e832da107f02fc692d8" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.344631 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c452f22-eef8-4e05-a1dd-f6c841833ab4" (UID: "3c452f22-eef8-4e05-a1dd-f6c841833ab4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.357839 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "271eeb3d-6d83-49c6-9d80-75b2276a304a" (UID: "271eeb3d-6d83-49c6-9d80-75b2276a304a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.365931 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5f78273-c1e2-452f-8bcb-a51c9bc238e4" (UID: "d5f78273-c1e2-452f-8bcb-a51c9bc238e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.366067 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-config-data" (OuterVolumeSpecName: "config-data") pod "d5f78273-c1e2-452f-8bcb-a51c9bc238e4" (UID: "d5f78273-c1e2-452f-8bcb-a51c9bc238e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.395019 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-f9gxt"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.403995 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-config\") pod \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.404113 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-combined-ca-bundle\") pod \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.404197 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbjp4\" (UniqueName: \"kubernetes.io/projected/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-api-access-hbjp4\") pod \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.404318 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-certs\") pod \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\" (UID: \"9a52309a-c6e4-481a-9fba-69de4c57e6b1\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.409786 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.409882 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5f78273-c1e2-452f-8bcb-a51c9bc238e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.409897 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.409909 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.418020 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-config-data" (OuterVolumeSpecName: "config-data") pod "3c452f22-eef8-4e05-a1dd-f6c841833ab4" (UID: "3c452f22-eef8-4e05-a1dd-f6c841833ab4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.430979 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ad5a-account-create-update-f9gxt"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.435163 4846 scope.go:117] "RemoveContainer" containerID="fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.452180 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-api-access-hbjp4" (OuterVolumeSpecName: "kube-api-access-hbjp4") pod "9a52309a-c6e4-481a-9fba-69de4c57e6b1" (UID: "9a52309a-c6e4-481a-9fba-69de4c57e6b1"). InnerVolumeSpecName "kube-api-access-hbjp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.453197 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.458669 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-z5lnl" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.477100 4846 scope.go:117] "RemoveContainer" containerID="a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.494797 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "9a52309a-c6e4-481a-9fba-69de4c57e6b1" (UID: "9a52309a-c6e4-481a-9fba-69de4c57e6b1"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.499796 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-config-data" (OuterVolumeSpecName: "config-data") pod "271eeb3d-6d83-49c6-9d80-75b2276a304a" (UID: "271eeb3d-6d83-49c6-9d80-75b2276a304a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.506504 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3c452f22-eef8-4e05-a1dd-f6c841833ab4" (UID: "3c452f22-eef8-4e05-a1dd-f6c841833ab4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.509188 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d858780-98c0-4167-9bc5-8a4cc0e054da" path="/var/lib/kubelet/pods/0d858780-98c0-4167-9bc5-8a4cc0e054da/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.520493 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d163475-4fe6-4992-90ae-4cfc97a1e6cd" path="/var/lib/kubelet/pods/1d163475-4fe6-4992-90ae-4cfc97a1e6cd/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.523217 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3" path="/var/lib/kubelet/pods/2bcbe7a8-5e9a-4adb-a94d-5f08234b05c3/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.524211 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.524237 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbjp4\" (UniqueName: \"kubernetes.io/projected/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-api-access-hbjp4\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.524250 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.524259 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.524269 4846 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.526576 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="318e1583-4759-4158-9683-14524ad771c2" path="/var/lib/kubelet/pods/318e1583-4759-4158-9683-14524ad771c2/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.527064 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3915f1bc-20d3-4e7d-b627-46406d0d0461" path="/var/lib/kubelet/pods/3915f1bc-20d3-4e7d-b627-46406d0d0461/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.527562 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="532236a1-a04d-4b99-9131-b1d16a3d5ed4" path="/var/lib/kubelet/pods/532236a1-a04d-4b99-9131-b1d16a3d5ed4/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.528755 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="534b394f-d62c-4c1d-8528-3a8aa37d5d7c" path="/var/lib/kubelet/pods/534b394f-d62c-4c1d-8528-3a8aa37d5d7c/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.530255 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5812b4fc-8c66-418f-8886-0ff26018b905" path="/var/lib/kubelet/pods/5812b4fc-8c66-418f-8886-0ff26018b905/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.530914 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb" path="/var/lib/kubelet/pods/73a3070f-a62c-4c04-9c00-ccdf1c5fc6cb/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.531317 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f25197c-8c4e-41ef-b7ef-1ed0f141599b" path="/var/lib/kubelet/pods/7f25197c-8c4e-41ef-b7ef-1ed0f141599b/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.532144 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757" path="/var/lib/kubelet/pods/b5fa8c0f-b2ae-49c3-a2d6-6ff48830d757/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.534353 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be88d7e8-f24f-4abf-b26a-e9698b1cd8f9" path="/var/lib/kubelet/pods/be88d7e8-f24f-4abf-b26a-e9698b1cd8f9/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.534906 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.536694 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c75e5c92-9e5b-4d18-bb6a-d183c18d8a54" path="/var/lib/kubelet/pods/c75e5c92-9e5b-4d18-bb6a-d183c18d8a54/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.540821 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da322cd4-f770-42d7-af2f-e4ed438da7dd" path="/var/lib/kubelet/pods/da322cd4-f770-42d7-af2f-e4ed438da7dd/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.541537 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef1ed9cc-3a1a-45cb-ba25-2db484fe129b" path="/var/lib/kubelet/pods/ef1ed9cc-3a1a-45cb-ba25-2db484fe129b/volumes" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.552179 4846 scope.go:117] "RemoveContainer" containerID="fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d" Feb 02 12:33:03 crc kubenswrapper[4846]: E0202 12:33:03.552854 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d\": container with ID starting with fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d not found: ID does not exist" containerID="fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.552888 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d"} err="failed to get container status \"fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d\": rpc error: code = NotFound desc = could not find container \"fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d\": container with ID starting with fd1ec95710d32536e5d02ceb5a86154e247f25651f1ce46d9ee97ccb6cd3553d not found: ID does not exist" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.552908 4846 scope.go:117] "RemoveContainer" containerID="a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce" Feb 02 12:33:03 crc kubenswrapper[4846]: E0202 12:33:03.553582 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce\": container with ID starting with a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce not found: ID does not exist" containerID="a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.553639 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce"} err="failed to get container status \"a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce\": rpc error: code = NotFound desc = could not find container \"a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce\": container with ID starting with a7e171296fa79829c4087198852d001506bdbbf6fb31552efef14c171e7f71ce not found: ID does not exist" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.553653 4846 scope.go:117] "RemoveContainer" containerID="f7527fb5dfdb0a8afb41a261ae73e292db1a4a3cb90a3972583cde5f54856e3b" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.556713 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "9a52309a-c6e4-481a-9fba-69de4c57e6b1" (UID: "9a52309a-c6e4-481a-9fba-69de4c57e6b1"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.562131 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.562156 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.562170 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c75698f54-h7x5d"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.562180 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c75698f54-h7x5d"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.562192 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-f6447845b-zlnf5"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.562204 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-f6447845b-zlnf5"] Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.576041 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "271eeb3d-6d83-49c6-9d80-75b2276a304a" (UID: "271eeb3d-6d83-49c6-9d80-75b2276a304a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.582543 4846 scope.go:117] "RemoveContainer" containerID="4f2c0bed69df185a3aca7bd0bcabf6f5afe33072cfbe8448804ba379ea4ff18e" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.589459 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a52309a-c6e4-481a-9fba-69de4c57e6b1" (UID: "9a52309a-c6e4-481a-9fba-69de4c57e6b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.606983 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3c452f22-eef8-4e05-a1dd-f6c841833ab4" (UID: "3c452f22-eef8-4e05-a1dd-f6c841833ab4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.625287 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6xjw\" (UniqueName: \"kubernetes.io/projected/c7dd2f87-cea2-4575-ad0c-483944433a55-kube-api-access-f6xjw\") pod \"c7dd2f87-cea2-4575-ad0c-483944433a55\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.625360 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-config-data\") pod \"c7dd2f87-cea2-4575-ad0c-483944433a55\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.625592 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/473fcbe1-028d-47ad-8c91-18b00e845232-operator-scripts\") pod \"473fcbe1-028d-47ad-8c91-18b00e845232\" (UID: \"473fcbe1-028d-47ad-8c91-18b00e845232\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.625646 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-combined-ca-bundle\") pod \"c7dd2f87-cea2-4575-ad0c-483944433a55\" (UID: \"c7dd2f87-cea2-4575-ad0c-483944433a55\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.625674 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgttm\" (UniqueName: \"kubernetes.io/projected/473fcbe1-028d-47ad-8c91-18b00e845232-kube-api-access-jgttm\") pod \"473fcbe1-028d-47ad-8c91-18b00e845232\" (UID: \"473fcbe1-028d-47ad-8c91-18b00e845232\") " Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.626278 4846 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.626320 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/271eeb3d-6d83-49c6-9d80-75b2276a304a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.626332 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.626343 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a52309a-c6e4-481a-9fba-69de4c57e6b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.626354 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c452f22-eef8-4e05-a1dd-f6c841833ab4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.627858 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/473fcbe1-028d-47ad-8c91-18b00e845232-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "473fcbe1-028d-47ad-8c91-18b00e845232" (UID: "473fcbe1-028d-47ad-8c91-18b00e845232"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.629976 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7dd2f87-cea2-4575-ad0c-483944433a55-kube-api-access-f6xjw" (OuterVolumeSpecName: "kube-api-access-f6xjw") pod "c7dd2f87-cea2-4575-ad0c-483944433a55" (UID: "c7dd2f87-cea2-4575-ad0c-483944433a55"). InnerVolumeSpecName "kube-api-access-f6xjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.638923 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/473fcbe1-028d-47ad-8c91-18b00e845232-kube-api-access-jgttm" (OuterVolumeSpecName: "kube-api-access-jgttm") pod "473fcbe1-028d-47ad-8c91-18b00e845232" (UID: "473fcbe1-028d-47ad-8c91-18b00e845232"). InnerVolumeSpecName "kube-api-access-jgttm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.662412 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7dd2f87-cea2-4575-ad0c-483944433a55" (UID: "c7dd2f87-cea2-4575-ad0c-483944433a55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.662807 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-config-data" (OuterVolumeSpecName: "config-data") pod "c7dd2f87-cea2-4575-ad0c-483944433a55" (UID: "c7dd2f87-cea2-4575-ad0c-483944433a55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.667587 4846 scope.go:117] "RemoveContainer" containerID="ba2153f30138c7492692991e0000a0e9232db386da8169c4c360432f8ac0b9a2" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.728388 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.728411 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/473fcbe1-028d-47ad-8c91-18b00e845232-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.728419 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7dd2f87-cea2-4575-ad0c-483944433a55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.728428 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgttm\" (UniqueName: \"kubernetes.io/projected/473fcbe1-028d-47ad-8c91-18b00e845232-kube-api-access-jgttm\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.728437 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6xjw\" (UniqueName: \"kubernetes.io/projected/c7dd2f87-cea2-4575-ad0c-483944433a55-kube-api-access-f6xjw\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.896280 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.909998 4846 scope.go:117] "RemoveContainer" containerID="66365bb3d3091805961a6da62564f6a7bf8e51dc20054684c80e03009a5d18dd" Feb 02 12:33:03 crc kubenswrapper[4846]: I0202 12:33:03.938368 4846 scope.go:117] "RemoveContainer" containerID="b358fa2e3b3d3040b293911322aa36237a5340d9d50ce3922695a3ff266a9b44" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.045133 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-config-data\") pod \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.045593 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-memcached-tls-certs\") pod \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.045682 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-combined-ca-bundle\") pod \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.045739 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7vcm\" (UniqueName: \"kubernetes.io/projected/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kube-api-access-h7vcm\") pod \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.045771 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kolla-config\") pod \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\" (UID: \"a7820c87-e5cc-42fd-b3f4-551cd39d3a15\") " Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.046312 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.046371 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data podName:9cec353a-92f9-4ed4-8ec9-15271cbc6912 nodeName:}" failed. No retries permitted until 2026-02-02 12:33:12.0463521 +0000 UTC m=+1423.274938963 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data") pod "rabbitmq-server-0" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912") : configmap "rabbitmq-config-data" not found Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.048052 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-config-data" (OuterVolumeSpecName: "config-data") pod "a7820c87-e5cc-42fd-b3f4-551cd39d3a15" (UID: "a7820c87-e5cc-42fd-b3f4-551cd39d3a15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.048650 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "a7820c87-e5cc-42fd-b3f4-551cd39d3a15" (UID: "a7820c87-e5cc-42fd-b3f4-551cd39d3a15"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.058824 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kube-api-access-h7vcm" (OuterVolumeSpecName: "kube-api-access-h7vcm") pod "a7820c87-e5cc-42fd-b3f4-551cd39d3a15" (UID: "a7820c87-e5cc-42fd-b3f4-551cd39d3a15"). InnerVolumeSpecName "kube-api-access-h7vcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.081453 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-z5lnl" event={"ID":"473fcbe1-028d-47ad-8c91-18b00e845232","Type":"ContainerDied","Data":"1679e6ab821caff4f71f3187a54c2be0303243fecf4dc974a939837ddde53f34"} Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.081707 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-z5lnl" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.087589 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e/ovn-northd/0.log" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.087660 4846 generic.go:334] "Generic (PLEG): container finished" podID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerID="a60702e8aab62dc62e3d8fd855c2cb15d42dc1d569a8a5f56df75bb229c8672e" exitCode=139 Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.088144 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e","Type":"ContainerDied","Data":"a60702e8aab62dc62e3d8fd855c2cb15d42dc1d569a8a5f56df75bb229c8672e"} Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.091180 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7820c87-e5cc-42fd-b3f4-551cd39d3a15" (UID: "a7820c87-e5cc-42fd-b3f4-551cd39d3a15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.092505 4846 scope.go:117] "RemoveContainer" containerID="f357c7a0e61fbd03ef490e5638f0b35530ac99e278d44d9f752d0369a28df54b" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.094188 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c7dd2f87-cea2-4575-ad0c-483944433a55","Type":"ContainerDied","Data":"fa693a993222ca3fa10756cb1e809d64ccf4f9a8cd192182b044fb70d408eab3"} Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.094267 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.103037 4846 generic.go:334] "Generic (PLEG): container finished" podID="a7820c87-e5cc-42fd-b3f4-551cd39d3a15" containerID="20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997" exitCode=0 Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.103324 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a7820c87-e5cc-42fd-b3f4-551cd39d3a15","Type":"ContainerDied","Data":"20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997"} Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.103434 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a7820c87-e5cc-42fd-b3f4-551cd39d3a15","Type":"ContainerDied","Data":"af4efb7632ab326628f579b99842b6efa732df21f2bd455ab4c2227d445e0985"} Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.103605 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.117677 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9a52309a-c6e4-481a-9fba-69de4c57e6b1","Type":"ContainerDied","Data":"72249adb66feaa7262b83c820c56d294b09b1b83a53e9d1b6f4f6ee1495786c2"} Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.117775 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.118095 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "a7820c87-e5cc-42fd-b3f4-551cd39d3a15" (UID: "a7820c87-e5cc-42fd-b3f4-551cd39d3a15"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.123798 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.124784 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.125549 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.128750 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.147509 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqrpl\" (UniqueName: \"kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl\") pod \"keystone-5b76-account-create-update-9h9tt\" (UID: \"7ad24791-f83d-4b1b-a19a-3e3b669b598d\") " pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.147562 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts\") pod \"keystone-5b76-account-create-update-9h9tt\" (UID: \"7ad24791-f83d-4b1b-a19a-3e3b669b598d\") " pod="openstack/keystone-5b76-account-create-update-9h9tt" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.147693 4846 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.147706 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.147715 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7vcm\" (UniqueName: \"kubernetes.io/projected/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kube-api-access-h7vcm\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.147726 4846 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-kolla-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.147734 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7820c87-e5cc-42fd-b3f4-551cd39d3a15-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.147792 4846 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.147840 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts podName:7ad24791-f83d-4b1b-a19a-3e3b669b598d nodeName:}" failed. No retries permitted until 2026-02-02 12:33:06.147826546 +0000 UTC m=+1417.376413409 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts") pod "keystone-5b76-account-create-update-9h9tt" (UID: "7ad24791-f83d-4b1b-a19a-3e3b669b598d") : configmap "openstack-scripts" not found Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.151493 4846 projected.go:194] Error preparing data for projected volume kube-api-access-tqrpl for pod openstack/keystone-5b76-account-create-update-9h9tt: failed to fetch token: serviceaccounts "galera-openstack" not found Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.151563 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl podName:7ad24791-f83d-4b1b-a19a-3e3b669b598d nodeName:}" failed. No retries permitted until 2026-02-02 12:33:06.151544367 +0000 UTC m=+1417.380131230 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-tqrpl" (UniqueName: "kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl") pod "keystone-5b76-account-create-update-9h9tt" (UID: "7ad24791-f83d-4b1b-a19a-3e3b669b598d") : failed to fetch token: serviceaccounts "galera-openstack" not found Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.155104 4846 scope.go:117] "RemoveContainer" containerID="40dea2331c92832ae0ee4b74bc35edbdedf985815110294b594bde6cb57fb223" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.232703 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.234568 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e/ovn-northd/0.log" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.234854 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.236580 4846 scope.go:117] "RemoveContainer" containerID="20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.239036 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.249355 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-rundir\") pod \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.249419 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wsbq\" (UniqueName: \"kubernetes.io/projected/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-kube-api-access-5wsbq\") pod \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.249453 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-northd-tls-certs\") pod \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.249471 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-config\") pod \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.249498 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-combined-ca-bundle\") pod \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.249582 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-scripts\") pod \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.249668 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-metrics-certs-tls-certs\") pod \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\" (UID: \"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.252537 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" (UID: "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.252927 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-scripts" (OuterVolumeSpecName: "scripts") pod "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" (UID: "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.261934 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-config" (OuterVolumeSpecName: "config") pod "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" (UID: "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.265225 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.265254 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.265265 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-rundir\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.269123 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-kube-api-access-5wsbq" (OuterVolumeSpecName: "kube-api-access-5wsbq") pod "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" (UID: "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e"). InnerVolumeSpecName "kube-api-access-5wsbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.276556 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-z5lnl"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.290772 4846 scope.go:117] "RemoveContainer" containerID="20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997" Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.291297 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997\": container with ID starting with 20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997 not found: ID does not exist" containerID="20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.291341 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997"} err="failed to get container status \"20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997\": rpc error: code = NotFound desc = could not find container \"20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997\": container with ID starting with 20a79194c98e732f1452a69c60eec16def8a2e98bd5748405e1e03b1c7b08997 not found: ID does not exist" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.291369 4846 scope.go:117] "RemoveContainer" containerID="218c4aa156faa3e69e561faabd794bbb67a15863d0a9d773a8b44e2e31cb4a4c" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.293154 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-z5lnl"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.297578 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" (UID: "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.320302 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.332849 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.349440 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5b76-account-create-update-9h9tt"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.366951 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5b76-account-create-update-9h9tt"] Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.386592 4846 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.386717 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:12.386690859 +0000 UTC m=+1423.615277722 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scripts" not found Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.389828 4846 secret.go:188] Couldn't get secret openstack/cinder-scheduler-config-data: secret "cinder-scheduler-config-data" not found Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.389908 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:12.389890157 +0000 UTC m=+1423.618477020 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scheduler-config-data" not found Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.390035 4846 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.390121 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:12.390087991 +0000 UTC m=+1423.618674854 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-config-data" not found Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.391302 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.400760 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wsbq\" (UniqueName: \"kubernetes.io/projected/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-kube-api-access-5wsbq\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.409812 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.401810 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" (UID: "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.416940 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.422269 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.431269 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.441577 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" (UID: "a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.446234 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.452475 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.459660 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.471070 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.494749 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512308 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-confd\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512357 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512391 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-tls\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512487 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/327c8143-b3ed-4431-a206-d2019ddcb606-pod-info\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512590 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-erlang-cookie\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512644 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5l9t\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-kube-api-access-f5l9t\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512681 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/327c8143-b3ed-4431-a206-d2019ddcb606-erlang-cookie-secret\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512703 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-plugins-conf\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512728 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-plugins\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512836 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-server-conf\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.512875 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data\") pod \"327c8143-b3ed-4431-a206-d2019ddcb606\" (UID: \"327c8143-b3ed-4431-a206-d2019ddcb606\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.513383 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqrpl\" (UniqueName: \"kubernetes.io/projected/7ad24791-f83d-4b1b-a19a-3e3b669b598d-kube-api-access-tqrpl\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.513433 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad24791-f83d-4b1b-a19a-3e3b669b598d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.513448 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.513463 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.513944 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.515123 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.515573 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.518306 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-kube-api-access-f5l9t" (OuterVolumeSpecName: "kube-api-access-f5l9t") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "kube-api-access-f5l9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.518360 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.519252 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.523076 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/327c8143-b3ed-4431-a206-d2019ddcb606-pod-info" (OuterVolumeSpecName: "pod-info") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.524181 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/327c8143-b3ed-4431-a206-d2019ddcb606-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.558063 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-server-conf" (OuterVolumeSpecName: "server-conf") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.561209 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data" (OuterVolumeSpecName: "config-data") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.589975 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "327c8143-b3ed-4431-a206-d2019ddcb606" (UID: "327c8143-b3ed-4431-a206-d2019ddcb606"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.614956 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.614991 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.615027 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.615041 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.615052 4846 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/327c8143-b3ed-4431-a206-d2019ddcb606-pod-info\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.615064 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.615074 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5l9t\" (UniqueName: \"kubernetes.io/projected/327c8143-b3ed-4431-a206-d2019ddcb606-kube-api-access-f5l9t\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.615084 4846 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.615095 4846 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/327c8143-b3ed-4431-a206-d2019ddcb606-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.615105 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/327c8143-b3ed-4431-a206-d2019ddcb606-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.615114 4846 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/327c8143-b3ed-4431-a206-d2019ddcb606-server-conf\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.631812 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.716491 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.798398 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.826049 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.826634 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.827096 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.827129 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.831811 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.833332 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.834564 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:04 crc kubenswrapper[4846]: E0202 12:33:04.834605 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovs-vswitchd" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919328 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-plugins-conf\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-confd\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919481 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9cec353a-92f9-4ed4-8ec9-15271cbc6912-pod-info\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919522 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919543 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-tls\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919572 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-server-conf\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919607 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjtpb\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-kube-api-access-tjtpb\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919684 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-erlang-cookie\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919728 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-plugins\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919764 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919792 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9cec353a-92f9-4ed4-8ec9-15271cbc6912-erlang-cookie-secret\") pod \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\" (UID: \"9cec353a-92f9-4ed4-8ec9-15271cbc6912\") " Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.919933 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.920199 4846 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.920475 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.921013 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.923514 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cec353a-92f9-4ed4-8ec9-15271cbc6912-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.924020 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-kube-api-access-tjtpb" (OuterVolumeSpecName: "kube-api-access-tjtpb") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "kube-api-access-tjtpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.927813 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.931504 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9cec353a-92f9-4ed4-8ec9-15271cbc6912-pod-info" (OuterVolumeSpecName: "pod-info") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.937740 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.945921 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data" (OuterVolumeSpecName: "config-data") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.963065 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-server-conf" (OuterVolumeSpecName: "server-conf") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:04 crc kubenswrapper[4846]: I0202 12:33:04.997343 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9cec353a-92f9-4ed4-8ec9-15271cbc6912" (UID: "9cec353a-92f9-4ed4-8ec9-15271cbc6912"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.021715 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.021767 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.021788 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.021799 4846 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9cec353a-92f9-4ed4-8ec9-15271cbc6912-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.021808 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.021818 4846 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9cec353a-92f9-4ed4-8ec9-15271cbc6912-pod-info\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.021827 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.021835 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.021846 4846 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9cec353a-92f9-4ed4-8ec9-15271cbc6912-server-conf\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.021855 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjtpb\" (UniqueName: \"kubernetes.io/projected/9cec353a-92f9-4ed4-8ec9-15271cbc6912-kube-api-access-tjtpb\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.044638 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.123244 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.142106 4846 generic.go:334] "Generic (PLEG): container finished" podID="9cec353a-92f9-4ed4-8ec9-15271cbc6912" containerID="488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2" exitCode=0 Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.142179 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9cec353a-92f9-4ed4-8ec9-15271cbc6912","Type":"ContainerDied","Data":"488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2"} Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.142212 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9cec353a-92f9-4ed4-8ec9-15271cbc6912","Type":"ContainerDied","Data":"4d3faa4533f7f892e5650a5fff8a4ddfbb90fa8747bd72cc9430ed290fd0d96a"} Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.142235 4846 scope.go:117] "RemoveContainer" containerID="488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.142438 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.157811 4846 generic.go:334] "Generic (PLEG): container finished" podID="327c8143-b3ed-4431-a206-d2019ddcb606" containerID="ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56" exitCode=0 Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.157909 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"327c8143-b3ed-4431-a206-d2019ddcb606","Type":"ContainerDied","Data":"ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56"} Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.157946 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"327c8143-b3ed-4431-a206-d2019ddcb606","Type":"ContainerDied","Data":"0acf417de61fa10c68d806b0e4b27aef887c00dc4b801daf9610cf72400b0adf"} Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.157920 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.163387 4846 generic.go:334] "Generic (PLEG): container finished" podID="b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" containerID="e484c823759e2a5e61def70b96f67d07c7974e29d5c4f9fc17eee40d86e83c71" exitCode=0 Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.163473 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a","Type":"ContainerDied","Data":"e484c823759e2a5e61def70b96f67d07c7974e29d5c4f9fc17eee40d86e83c71"} Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.171664 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e/ovn-northd/0.log" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.171702 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e","Type":"ContainerDied","Data":"be0202f872b03082898f8eac24324b031c5a25cacce1fb82f674633ee4244ce3"} Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.171778 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.188644 4846 scope.go:117] "RemoveContainer" containerID="0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.193284 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6497c67f7-24wxw" podUID="5a1c1394-3713-41ee-948e-3800e5d51020" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.167:9696/\": dial tcp 10.217.0.167:9696: connect: connection refused" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.208783 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.222892 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.233148 4846 scope.go:117] "RemoveContainer" containerID="488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2" Feb 02 12:33:05 crc kubenswrapper[4846]: E0202 12:33:05.233977 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2\": container with ID starting with 488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2 not found: ID does not exist" containerID="488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.234032 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2"} err="failed to get container status \"488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2\": rpc error: code = NotFound desc = could not find container \"488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2\": container with ID starting with 488810b80e229697844a66a71b7e9f1f00e426cb92168e34ffbfc0215e3085a2 not found: ID does not exist" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.234067 4846 scope.go:117] "RemoveContainer" containerID="0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f" Feb 02 12:33:05 crc kubenswrapper[4846]: E0202 12:33:05.234407 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f\": container with ID starting with 0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f not found: ID does not exist" containerID="0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.234448 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f"} err="failed to get container status \"0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f\": rpc error: code = NotFound desc = could not find container \"0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f\": container with ID starting with 0b918655f8f6931166b18f5e1de988412763f8659ee38050800bcbace2d2389f not found: ID does not exist" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.234476 4846 scope.go:117] "RemoveContainer" containerID="ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.239713 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.246633 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.251793 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.256860 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.258010 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.261791 4846 scope.go:117] "RemoveContainer" containerID="f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.291958 4846 scope.go:117] "RemoveContainer" containerID="ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56" Feb 02 12:33:05 crc kubenswrapper[4846]: E0202 12:33:05.292382 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56\": container with ID starting with ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56 not found: ID does not exist" containerID="ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.292427 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56"} err="failed to get container status \"ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56\": rpc error: code = NotFound desc = could not find container \"ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56\": container with ID starting with ef097544298b2db51cab809f1471d986eba12b63e52a2cd9f51b0572a0c99d56 not found: ID does not exist" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.292454 4846 scope.go:117] "RemoveContainer" containerID="f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb" Feb 02 12:33:05 crc kubenswrapper[4846]: E0202 12:33:05.293084 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb\": container with ID starting with f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb not found: ID does not exist" containerID="f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.293122 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb"} err="failed to get container status \"f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb\": rpc error: code = NotFound desc = could not find container \"f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb\": container with ID starting with f529d8d0a7eea5f1def171c964e6b04a897292f90f046254f9b2f132ce6817bb not found: ID does not exist" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.293149 4846 scope.go:117] "RemoveContainer" containerID="a6d2a885c9e46febf0a5624587138a27dc292d388c2adfac2d4f55ce324f5505" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.318651 4846 scope.go:117] "RemoveContainer" containerID="a60702e8aab62dc62e3d8fd855c2cb15d42dc1d569a8a5f56df75bb229c8672e" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.330351 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-generated\") pod \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.330408 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-operator-scripts\") pod \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.330434 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-default\") pod \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.330456 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.330481 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-galera-tls-certs\") pod \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.330536 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-combined-ca-bundle\") pod \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.330553 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk6x9\" (UniqueName: \"kubernetes.io/projected/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kube-api-access-fk6x9\") pod \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.330602 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kolla-config\") pod \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\" (UID: \"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.331383 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" (UID: "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.331639 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" (UID: "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.331686 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" (UID: "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.331884 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" (UID: "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.333989 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kube-api-access-fk6x9" (OuterVolumeSpecName: "kube-api-access-fk6x9") pod "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" (UID: "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a"). InnerVolumeSpecName "kube-api-access-fk6x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.338844 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "mysql-db") pod "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" (UID: "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.358193 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" (UID: "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.371355 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" (UID: "b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.431725 4846 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kolla-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.431753 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-generated\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.431764 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.431773 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-config-data-default\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.431800 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.431809 4846 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.431817 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.431825 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk6x9\" (UniqueName: \"kubernetes.io/projected/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a-kube-api-access-fk6x9\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.432590 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="271eeb3d-6d83-49c6-9d80-75b2276a304a" path="/var/lib/kubelet/pods/271eeb3d-6d83-49c6-9d80-75b2276a304a/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.433648 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="327c8143-b3ed-4431-a206-d2019ddcb606" path="/var/lib/kubelet/pods/327c8143-b3ed-4431-a206-d2019ddcb606/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.434763 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" path="/var/lib/kubelet/pods/3c452f22-eef8-4e05-a1dd-f6c841833ab4/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.435352 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="473fcbe1-028d-47ad-8c91-18b00e845232" path="/var/lib/kubelet/pods/473fcbe1-028d-47ad-8c91-18b00e845232/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.435741 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="525708d2-15d1-4bfb-8258-bf216ebc4bab" path="/var/lib/kubelet/pods/525708d2-15d1-4bfb-8258-bf216ebc4bab/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.436332 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2d348d-dbbf-4046-b9fb-94f8aab64573" path="/var/lib/kubelet/pods/5e2d348d-dbbf-4046-b9fb-94f8aab64573/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.437454 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69522beb-f7d5-4694-ada8-0d17d7d3f16b" path="/var/lib/kubelet/pods/69522beb-f7d5-4694-ada8-0d17d7d3f16b/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.437929 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ad24791-f83d-4b1b-a19a-3e3b669b598d" path="/var/lib/kubelet/pods/7ad24791-f83d-4b1b-a19a-3e3b669b598d/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.438296 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a52309a-c6e4-481a-9fba-69de4c57e6b1" path="/var/lib/kubelet/pods/9a52309a-c6e4-481a-9fba-69de4c57e6b1/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.439942 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cec353a-92f9-4ed4-8ec9-15271cbc6912" path="/var/lib/kubelet/pods/9cec353a-92f9-4ed4-8ec9-15271cbc6912/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.441199 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" path="/var/lib/kubelet/pods/a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.441912 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7820c87-e5cc-42fd-b3f4-551cd39d3a15" path="/var/lib/kubelet/pods/a7820c87-e5cc-42fd-b3f4-551cd39d3a15/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.442466 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7dd2f87-cea2-4575-ad0c-483944433a55" path="/var/lib/kubelet/pods/c7dd2f87-cea2-4575-ad0c-483944433a55/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.443431 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5f78273-c1e2-452f-8bcb-a51c9bc238e4" path="/var/lib/kubelet/pods/d5f78273-c1e2-452f-8bcb-a51c9bc238e4/volumes" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.459881 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.532885 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.891962 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.938738 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-scripts\") pod \"92819684-ecb5-4c4d-950c-edacf38e269b\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.938840 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-combined-ca-bundle\") pod \"92819684-ecb5-4c4d-950c-edacf38e269b\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.938859 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnhmc\" (UniqueName: \"kubernetes.io/projected/92819684-ecb5-4c4d-950c-edacf38e269b-kube-api-access-xnhmc\") pod \"92819684-ecb5-4c4d-950c-edacf38e269b\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.938894 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-ceilometer-tls-certs\") pod \"92819684-ecb5-4c4d-950c-edacf38e269b\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.938918 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-log-httpd\") pod \"92819684-ecb5-4c4d-950c-edacf38e269b\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.938955 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-run-httpd\") pod \"92819684-ecb5-4c4d-950c-edacf38e269b\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.938981 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-config-data\") pod \"92819684-ecb5-4c4d-950c-edacf38e269b\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.939019 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-sg-core-conf-yaml\") pod \"92819684-ecb5-4c4d-950c-edacf38e269b\" (UID: \"92819684-ecb5-4c4d-950c-edacf38e269b\") " Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.939999 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "92819684-ecb5-4c4d-950c-edacf38e269b" (UID: "92819684-ecb5-4c4d-950c-edacf38e269b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.940347 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "92819684-ecb5-4c4d-950c-edacf38e269b" (UID: "92819684-ecb5-4c4d-950c-edacf38e269b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.947945 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92819684-ecb5-4c4d-950c-edacf38e269b-kube-api-access-xnhmc" (OuterVolumeSpecName: "kube-api-access-xnhmc") pod "92819684-ecb5-4c4d-950c-edacf38e269b" (UID: "92819684-ecb5-4c4d-950c-edacf38e269b"). InnerVolumeSpecName "kube-api-access-xnhmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.949847 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-scripts" (OuterVolumeSpecName: "scripts") pod "92819684-ecb5-4c4d-950c-edacf38e269b" (UID: "92819684-ecb5-4c4d-950c-edacf38e269b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:05 crc kubenswrapper[4846]: I0202 12:33:05.971746 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "92819684-ecb5-4c4d-950c-edacf38e269b" (UID: "92819684-ecb5-4c4d-950c-edacf38e269b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.015543 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "92819684-ecb5-4c4d-950c-edacf38e269b" (UID: "92819684-ecb5-4c4d-950c-edacf38e269b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.016741 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92819684-ecb5-4c4d-950c-edacf38e269b" (UID: "92819684-ecb5-4c4d-950c-edacf38e269b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.041812 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.041841 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.041850 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.041858 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnhmc\" (UniqueName: \"kubernetes.io/projected/92819684-ecb5-4c4d-950c-edacf38e269b-kube-api-access-xnhmc\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.041887 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.041896 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.041904 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92819684-ecb5-4c4d-950c-edacf38e269b-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.074799 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-config-data" (OuterVolumeSpecName: "config-data") pod "92819684-ecb5-4c4d-950c-edacf38e269b" (UID: "92819684-ecb5-4c4d-950c-edacf38e269b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.095069 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.161830 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-internal-tls-certs\") pod \"c858d5df-042d-4211-898b-1ad4ef1fea47\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.161893 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-credential-keys\") pod \"c858d5df-042d-4211-898b-1ad4ef1fea47\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.161932 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-public-tls-certs\") pod \"c858d5df-042d-4211-898b-1ad4ef1fea47\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.161973 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-config-data\") pod \"c858d5df-042d-4211-898b-1ad4ef1fea47\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.162356 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-scripts\") pod \"c858d5df-042d-4211-898b-1ad4ef1fea47\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.162899 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-fernet-keys\") pod \"c858d5df-042d-4211-898b-1ad4ef1fea47\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.162933 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cglmv\" (UniqueName: \"kubernetes.io/projected/c858d5df-042d-4211-898b-1ad4ef1fea47-kube-api-access-cglmv\") pod \"c858d5df-042d-4211-898b-1ad4ef1fea47\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.162982 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-combined-ca-bundle\") pod \"c858d5df-042d-4211-898b-1ad4ef1fea47\" (UID: \"c858d5df-042d-4211-898b-1ad4ef1fea47\") " Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.163761 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92819684-ecb5-4c4d-950c-edacf38e269b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.168342 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c858d5df-042d-4211-898b-1ad4ef1fea47" (UID: "c858d5df-042d-4211-898b-1ad4ef1fea47"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.168768 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c858d5df-042d-4211-898b-1ad4ef1fea47-kube-api-access-cglmv" (OuterVolumeSpecName: "kube-api-access-cglmv") pod "c858d5df-042d-4211-898b-1ad4ef1fea47" (UID: "c858d5df-042d-4211-898b-1ad4ef1fea47"). InnerVolumeSpecName "kube-api-access-cglmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.193109 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c858d5df-042d-4211-898b-1ad4ef1fea47" (UID: "c858d5df-042d-4211-898b-1ad4ef1fea47"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.196728 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-scripts" (OuterVolumeSpecName: "scripts") pod "c858d5df-042d-4211-898b-1ad4ef1fea47" (UID: "c858d5df-042d-4211-898b-1ad4ef1fea47"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.198279 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c858d5df-042d-4211-898b-1ad4ef1fea47" (UID: "c858d5df-042d-4211-898b-1ad4ef1fea47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.203181 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a","Type":"ContainerDied","Data":"d616050f9a822b3fb0ebb372d1934eca8cdf9b1fa2823a510a1a7315eb62b648"} Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.203228 4846 scope.go:117] "RemoveContainer" containerID="e484c823759e2a5e61def70b96f67d07c7974e29d5c4f9fc17eee40d86e83c71" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.203338 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.209157 4846 generic.go:334] "Generic (PLEG): container finished" podID="c858d5df-042d-4211-898b-1ad4ef1fea47" containerID="c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced" exitCode=0 Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.209223 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6fb558fdfb-5z9p5" event={"ID":"c858d5df-042d-4211-898b-1ad4ef1fea47","Type":"ContainerDied","Data":"c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced"} Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.209247 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6fb558fdfb-5z9p5" event={"ID":"c858d5df-042d-4211-898b-1ad4ef1fea47","Type":"ContainerDied","Data":"86f259fd640dbbb237ee8cd278959e04a228664bd7b954c37f50421a3305dd96"} Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.209320 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6fb558fdfb-5z9p5" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.212656 4846 generic.go:334] "Generic (PLEG): container finished" podID="92819684-ecb5-4c4d-950c-edacf38e269b" containerID="b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f" exitCode=0 Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.212725 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92819684-ecb5-4c4d-950c-edacf38e269b","Type":"ContainerDied","Data":"b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f"} Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.212747 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92819684-ecb5-4c4d-950c-edacf38e269b","Type":"ContainerDied","Data":"7812dfc323a65a36cf25f3ba6d08f1f304a1dc41c215d04527f26995552b4c29"} Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.212771 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.222597 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-config-data" (OuterVolumeSpecName: "config-data") pod "c858d5df-042d-4211-898b-1ad4ef1fea47" (UID: "c858d5df-042d-4211-898b-1ad4ef1fea47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.228901 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.233176 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.237109 4846 scope.go:117] "RemoveContainer" containerID="2849d11fa15affc0fa45246ebd204917e430eb9bef2190522d139fa1508c3879" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.249912 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.257437 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c858d5df-042d-4211-898b-1ad4ef1fea47" (UID: "c858d5df-042d-4211-898b-1ad4ef1fea47"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.262257 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.265913 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.266163 4846 scope.go:117] "RemoveContainer" containerID="c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.266166 4846 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.266284 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.266298 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.266310 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.266323 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.266333 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cglmv\" (UniqueName: \"kubernetes.io/projected/c858d5df-042d-4211-898b-1ad4ef1fea47-kube-api-access-cglmv\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.268721 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c858d5df-042d-4211-898b-1ad4ef1fea47" (UID: "c858d5df-042d-4211-898b-1ad4ef1fea47"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.296494 4846 scope.go:117] "RemoveContainer" containerID="c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced" Feb 02 12:33:06 crc kubenswrapper[4846]: E0202 12:33:06.296911 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced\": container with ID starting with c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced not found: ID does not exist" containerID="c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.296947 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced"} err="failed to get container status \"c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced\": rpc error: code = NotFound desc = could not find container \"c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced\": container with ID starting with c12a221429998bf5e5218d8ab0a416ab8d9dd1495dccaf0453bdbd3278466ced not found: ID does not exist" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.296971 4846 scope.go:117] "RemoveContainer" containerID="c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.325947 4846 scope.go:117] "RemoveContainer" containerID="821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.348336 4846 scope.go:117] "RemoveContainer" containerID="b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.367391 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c858d5df-042d-4211-898b-1ad4ef1fea47-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.368780 4846 scope.go:117] "RemoveContainer" containerID="e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.390075 4846 scope.go:117] "RemoveContainer" containerID="c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e" Feb 02 12:33:06 crc kubenswrapper[4846]: E0202 12:33:06.390590 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e\": container with ID starting with c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e not found: ID does not exist" containerID="c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.390651 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e"} err="failed to get container status \"c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e\": rpc error: code = NotFound desc = could not find container \"c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e\": container with ID starting with c6cb861e2c0a22c50bca996fadf561d74ecb50b41f2c1ee302d4db8aaf20d11e not found: ID does not exist" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.390680 4846 scope.go:117] "RemoveContainer" containerID="821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91" Feb 02 12:33:06 crc kubenswrapper[4846]: E0202 12:33:06.391182 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91\": container with ID starting with 821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91 not found: ID does not exist" containerID="821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.391203 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91"} err="failed to get container status \"821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91\": rpc error: code = NotFound desc = could not find container \"821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91\": container with ID starting with 821c5a261142d2270ac550b15f62ba672ef7abc82f5c4802966b56d541c2cd91 not found: ID does not exist" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.391215 4846 scope.go:117] "RemoveContainer" containerID="b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f" Feb 02 12:33:06 crc kubenswrapper[4846]: E0202 12:33:06.391554 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f\": container with ID starting with b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f not found: ID does not exist" containerID="b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.391643 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f"} err="failed to get container status \"b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f\": rpc error: code = NotFound desc = could not find container \"b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f\": container with ID starting with b53bf3648137604f9a32ecefd89911564190b11ae597d4c3b49735775d6aed5f not found: ID does not exist" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.391729 4846 scope.go:117] "RemoveContainer" containerID="e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea" Feb 02 12:33:06 crc kubenswrapper[4846]: E0202 12:33:06.392162 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea\": container with ID starting with e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea not found: ID does not exist" containerID="e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.392197 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea"} err="failed to get container status \"e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea\": rpc error: code = NotFound desc = could not find container \"e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea\": container with ID starting with e38588cfcf48ab5f94521a835423508fd2a191efa8392c9cab2683c3ae0b1bea not found: ID does not exist" Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.537478 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6fb558fdfb-5z9p5"] Feb 02 12:33:06 crc kubenswrapper[4846]: I0202 12:33:06.541721 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6fb558fdfb-5z9p5"] Feb 02 12:33:07 crc kubenswrapper[4846]: I0202 12:33:07.449221 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" path="/var/lib/kubelet/pods/92819684-ecb5-4c4d-950c-edacf38e269b/volumes" Feb 02 12:33:07 crc kubenswrapper[4846]: I0202 12:33:07.451927 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" path="/var/lib/kubelet/pods/b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a/volumes" Feb 02 12:33:07 crc kubenswrapper[4846]: I0202 12:33:07.453698 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c858d5df-042d-4211-898b-1ad4ef1fea47" path="/var/lib/kubelet/pods/c858d5df-042d-4211-898b-1ad4ef1fea47/volumes" Feb 02 12:33:08 crc kubenswrapper[4846]: I0202 12:33:08.615101 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="a7820c87-e5cc-42fd-b3f4-551cd39d3a15" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.105:11211: i/o timeout" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.124132 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.220829 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-public-tls-certs\") pod \"5a1c1394-3713-41ee-948e-3800e5d51020\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.220922 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-internal-tls-certs\") pod \"5a1c1394-3713-41ee-948e-3800e5d51020\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.220977 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-config\") pod \"5a1c1394-3713-41ee-948e-3800e5d51020\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.221011 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-combined-ca-bundle\") pod \"5a1c1394-3713-41ee-948e-3800e5d51020\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.221038 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-ovndb-tls-certs\") pod \"5a1c1394-3713-41ee-948e-3800e5d51020\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.221106 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-httpd-config\") pod \"5a1c1394-3713-41ee-948e-3800e5d51020\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.221135 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql92s\" (UniqueName: \"kubernetes.io/projected/5a1c1394-3713-41ee-948e-3800e5d51020-kube-api-access-ql92s\") pod \"5a1c1394-3713-41ee-948e-3800e5d51020\" (UID: \"5a1c1394-3713-41ee-948e-3800e5d51020\") " Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.226598 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5a1c1394-3713-41ee-948e-3800e5d51020" (UID: "5a1c1394-3713-41ee-948e-3800e5d51020"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.238829 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a1c1394-3713-41ee-948e-3800e5d51020-kube-api-access-ql92s" (OuterVolumeSpecName: "kube-api-access-ql92s") pod "5a1c1394-3713-41ee-948e-3800e5d51020" (UID: "5a1c1394-3713-41ee-948e-3800e5d51020"). InnerVolumeSpecName "kube-api-access-ql92s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.264089 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5a1c1394-3713-41ee-948e-3800e5d51020" (UID: "5a1c1394-3713-41ee-948e-3800e5d51020"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.266196 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-config" (OuterVolumeSpecName: "config") pod "5a1c1394-3713-41ee-948e-3800e5d51020" (UID: "5a1c1394-3713-41ee-948e-3800e5d51020"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.270818 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a1c1394-3713-41ee-948e-3800e5d51020" (UID: "5a1c1394-3713-41ee-948e-3800e5d51020"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.273085 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5a1c1394-3713-41ee-948e-3800e5d51020" (UID: "5a1c1394-3713-41ee-948e-3800e5d51020"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.278310 4846 generic.go:334] "Generic (PLEG): container finished" podID="5a1c1394-3713-41ee-948e-3800e5d51020" containerID="38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd" exitCode=0 Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.278406 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6497c67f7-24wxw" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.278452 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6497c67f7-24wxw" event={"ID":"5a1c1394-3713-41ee-948e-3800e5d51020","Type":"ContainerDied","Data":"38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd"} Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.278652 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6497c67f7-24wxw" event={"ID":"5a1c1394-3713-41ee-948e-3800e5d51020","Type":"ContainerDied","Data":"64b75209ff5acfa45f93f1ea20834dcdc3631d0b716bc99475e3d63d64e1ad21"} Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.278679 4846 scope.go:117] "RemoveContainer" containerID="f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.287169 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5a1c1394-3713-41ee-948e-3800e5d51020" (UID: "5a1c1394-3713-41ee-948e-3800e5d51020"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.303034 4846 scope.go:117] "RemoveContainer" containerID="38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.320284 4846 scope.go:117] "RemoveContainer" containerID="f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292" Feb 02 12:33:09 crc kubenswrapper[4846]: E0202 12:33:09.320841 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292\": container with ID starting with f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292 not found: ID does not exist" containerID="f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.320871 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292"} err="failed to get container status \"f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292\": rpc error: code = NotFound desc = could not find container \"f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292\": container with ID starting with f709c0cf649ffbfa64472d17e3f87f2e7b446ed716028e7c8a9ef2d443efb292 not found: ID does not exist" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.320907 4846 scope.go:117] "RemoveContainer" containerID="38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd" Feb 02 12:33:09 crc kubenswrapper[4846]: E0202 12:33:09.321198 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd\": container with ID starting with 38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd not found: ID does not exist" containerID="38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.321230 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd"} err="failed to get container status \"38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd\": rpc error: code = NotFound desc = could not find container \"38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd\": container with ID starting with 38c858bcf6a7b97b1804c69987e4a04a63be90335f383e4f8159d1c3d0725cbd not found: ID does not exist" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.323216 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.323246 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql92s\" (UniqueName: \"kubernetes.io/projected/5a1c1394-3713-41ee-948e-3800e5d51020-kube-api-access-ql92s\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.323261 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.323301 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.323315 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-config\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.323329 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.323342 4846 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a1c1394-3713-41ee-948e-3800e5d51020-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.605756 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6497c67f7-24wxw"] Feb 02 12:33:09 crc kubenswrapper[4846]: I0202 12:33:09.627570 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6497c67f7-24wxw"] Feb 02 12:33:09 crc kubenswrapper[4846]: E0202 12:33:09.823105 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:09 crc kubenswrapper[4846]: E0202 12:33:09.823677 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:09 crc kubenswrapper[4846]: E0202 12:33:09.823987 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:09 crc kubenswrapper[4846]: E0202 12:33:09.824018 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" Feb 02 12:33:09 crc kubenswrapper[4846]: E0202 12:33:09.824347 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:09 crc kubenswrapper[4846]: E0202 12:33:09.826101 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:09 crc kubenswrapper[4846]: E0202 12:33:09.829826 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:09 crc kubenswrapper[4846]: E0202 12:33:09.829864 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovs-vswitchd" Feb 02 12:33:11 crc kubenswrapper[4846]: I0202 12:33:11.435000 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a1c1394-3713-41ee-948e-3800e5d51020" path="/var/lib/kubelet/pods/5a1c1394-3713-41ee-948e-3800e5d51020/volumes" Feb 02 12:33:12 crc kubenswrapper[4846]: E0202 12:33:12.476748 4846 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Feb 02 12:33:12 crc kubenswrapper[4846]: E0202 12:33:12.477228 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:28.477194164 +0000 UTC m=+1439.705781187 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scripts" not found Feb 02 12:33:12 crc kubenswrapper[4846]: E0202 12:33:12.476765 4846 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Feb 02 12:33:12 crc kubenswrapper[4846]: E0202 12:33:12.476765 4846 secret.go:188] Couldn't get secret openstack/cinder-scheduler-config-data: secret "cinder-scheduler-config-data" not found Feb 02 12:33:12 crc kubenswrapper[4846]: E0202 12:33:12.477361 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:28.477331767 +0000 UTC m=+1439.705918630 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-config-data" not found Feb 02 12:33:12 crc kubenswrapper[4846]: E0202 12:33:12.477498 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom podName:444e13f0-bc18-4bc9-8d92-af6a379e22ab nodeName:}" failed. No retries permitted until 2026-02-02 12:33:28.47747978 +0000 UTC m=+1439.706066643 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom") pod "cinder-scheduler-0" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab") : secret "cinder-scheduler-config-data" not found Feb 02 12:33:14 crc kubenswrapper[4846]: E0202 12:33:14.822590 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:14 crc kubenswrapper[4846]: E0202 12:33:14.823752 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:14 crc kubenswrapper[4846]: E0202 12:33:14.823757 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:14 crc kubenswrapper[4846]: E0202 12:33:14.824216 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:14 crc kubenswrapper[4846]: E0202 12:33:14.824255 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" Feb 02 12:33:14 crc kubenswrapper[4846]: E0202 12:33:14.826888 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:14 crc kubenswrapper[4846]: E0202 12:33:14.828797 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:14 crc kubenswrapper[4846]: E0202 12:33:14.828833 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovs-vswitchd" Feb 02 12:33:19 crc kubenswrapper[4846]: E0202 12:33:19.823432 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:19 crc kubenswrapper[4846]: E0202 12:33:19.824711 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:19 crc kubenswrapper[4846]: E0202 12:33:19.825058 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:19 crc kubenswrapper[4846]: E0202 12:33:19.825077 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:19 crc kubenswrapper[4846]: E0202 12:33:19.825143 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" Feb 02 12:33:19 crc kubenswrapper[4846]: E0202 12:33:19.826458 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:19 crc kubenswrapper[4846]: E0202 12:33:19.827881 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:19 crc kubenswrapper[4846]: E0202 12:33:19.827933 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovs-vswitchd" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875138 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wqsff"] Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875773 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerName="openstack-network-exporter" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875786 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerName="openstack-network-exporter" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875795 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c858d5df-042d-4211-898b-1ad4ef1fea47" containerName="keystone-api" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875801 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c858d5df-042d-4211-898b-1ad4ef1fea47" containerName="keystone-api" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875808 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerName="nova-api-log" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875814 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerName="nova-api-log" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875824 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cec353a-92f9-4ed4-8ec9-15271cbc6912" containerName="setup-container" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875830 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cec353a-92f9-4ed4-8ec9-15271cbc6912" containerName="setup-container" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875844 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" containerName="mysql-bootstrap" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875850 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" containerName="mysql-bootstrap" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875859 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cec353a-92f9-4ed4-8ec9-15271cbc6912" containerName="rabbitmq" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875865 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cec353a-92f9-4ed4-8ec9-15271cbc6912" containerName="rabbitmq" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875876 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="327c8143-b3ed-4431-a206-d2019ddcb606" containerName="setup-container" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875883 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="327c8143-b3ed-4431-a206-d2019ddcb606" containerName="setup-container" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875895 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="ceilometer-central-agent" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875901 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="ceilometer-central-agent" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875913 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5f78273-c1e2-452f-8bcb-a51c9bc238e4" containerName="nova-cell1-conductor-conductor" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875920 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5f78273-c1e2-452f-8bcb-a51c9bc238e4" containerName="nova-cell1-conductor-conductor" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875929 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerName="ovn-northd" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875938 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerName="ovn-northd" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875956 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="327c8143-b3ed-4431-a206-d2019ddcb606" containerName="rabbitmq" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875964 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="327c8143-b3ed-4431-a206-d2019ddcb606" containerName="rabbitmq" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.875980 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerName="nova-api-api" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.875987 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerName="nova-api-api" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876000 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271eeb3d-6d83-49c6-9d80-75b2276a304a" containerName="glance-httpd" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876008 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="271eeb3d-6d83-49c6-9d80-75b2276a304a" containerName="glance-httpd" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876020 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a52309a-c6e4-481a-9fba-69de4c57e6b1" containerName="kube-state-metrics" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876030 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a52309a-c6e4-481a-9fba-69de4c57e6b1" containerName="kube-state-metrics" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876044 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1c1394-3713-41ee-948e-3800e5d51020" containerName="neutron-httpd" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876052 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1c1394-3713-41ee-948e-3800e5d51020" containerName="neutron-httpd" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876065 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="ceilometer-notification-agent" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876075 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="ceilometer-notification-agent" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876086 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="proxy-httpd" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876092 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="proxy-httpd" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876100 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271eeb3d-6d83-49c6-9d80-75b2276a304a" containerName="glance-log" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876105 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="271eeb3d-6d83-49c6-9d80-75b2276a304a" containerName="glance-log" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876119 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="sg-core" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876126 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="sg-core" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876137 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1c1394-3713-41ee-948e-3800e5d51020" containerName="neutron-api" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876142 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1c1394-3713-41ee-948e-3800e5d51020" containerName="neutron-api" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876154 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7dd2f87-cea2-4575-ad0c-483944433a55" containerName="nova-cell0-conductor-conductor" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876160 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7dd2f87-cea2-4575-ad0c-483944433a55" containerName="nova-cell0-conductor-conductor" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876174 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7820c87-e5cc-42fd-b3f4-551cd39d3a15" containerName="memcached" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876182 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7820c87-e5cc-42fd-b3f4-551cd39d3a15" containerName="memcached" Feb 02 12:33:22 crc kubenswrapper[4846]: E0202 12:33:22.876193 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" containerName="galera" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876201 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" containerName="galera" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876338 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerName="openstack-network-exporter" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876353 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cec353a-92f9-4ed4-8ec9-15271cbc6912" containerName="rabbitmq" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876361 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c858d5df-042d-4211-898b-1ad4ef1fea47" containerName="keystone-api" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876371 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerName="nova-api-log" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876379 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="327c8143-b3ed-4431-a206-d2019ddcb606" containerName="rabbitmq" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876389 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7dd2f87-cea2-4575-ad0c-483944433a55" containerName="nova-cell0-conductor-conductor" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876397 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7820c87-e5cc-42fd-b3f4-551cd39d3a15" containerName="memcached" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876408 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1c1394-3713-41ee-948e-3800e5d51020" containerName="neutron-api" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876414 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="ceilometer-notification-agent" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876423 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ec2404-1bf0-4fc1-bc66-1b2023c7cd9e" containerName="ovn-northd" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876428 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="ceilometer-central-agent" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876442 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5f78273-c1e2-452f-8bcb-a51c9bc238e4" containerName="nova-cell1-conductor-conductor" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876449 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="sg-core" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876457 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="271eeb3d-6d83-49c6-9d80-75b2276a304a" containerName="glance-httpd" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876467 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="271eeb3d-6d83-49c6-9d80-75b2276a304a" containerName="glance-log" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876477 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1c1394-3713-41ee-948e-3800e5d51020" containerName="neutron-httpd" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876488 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a52309a-c6e4-481a-9fba-69de4c57e6b1" containerName="kube-state-metrics" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876500 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c452f22-eef8-4e05-a1dd-f6c841833ab4" containerName="nova-api-api" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876513 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1b78cf8-e00b-4c2b-ba6f-28bd24eb0f5a" containerName="galera" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.876522 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="92819684-ecb5-4c4d-950c-edacf38e269b" containerName="proxy-httpd" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.877650 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.887947 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wqsff"] Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.943193 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-catalog-content\") pod \"redhat-operators-wqsff\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.943253 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-utilities\") pod \"redhat-operators-wqsff\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:22 crc kubenswrapper[4846]: I0202 12:33:22.943325 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9zhd\" (UniqueName: \"kubernetes.io/projected/775f235b-175e-4a8b-88cd-3e3f735949ec-kube-api-access-v9zhd\") pod \"redhat-operators-wqsff\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:23 crc kubenswrapper[4846]: I0202 12:33:23.044319 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9zhd\" (UniqueName: \"kubernetes.io/projected/775f235b-175e-4a8b-88cd-3e3f735949ec-kube-api-access-v9zhd\") pod \"redhat-operators-wqsff\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:23 crc kubenswrapper[4846]: I0202 12:33:23.044488 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-catalog-content\") pod \"redhat-operators-wqsff\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:23 crc kubenswrapper[4846]: I0202 12:33:23.044531 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-utilities\") pod \"redhat-operators-wqsff\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:23 crc kubenswrapper[4846]: I0202 12:33:23.045120 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-utilities\") pod \"redhat-operators-wqsff\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:23 crc kubenswrapper[4846]: I0202 12:33:23.045123 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-catalog-content\") pod \"redhat-operators-wqsff\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:23 crc kubenswrapper[4846]: I0202 12:33:23.063486 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9zhd\" (UniqueName: \"kubernetes.io/projected/775f235b-175e-4a8b-88cd-3e3f735949ec-kube-api-access-v9zhd\") pod \"redhat-operators-wqsff\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:23 crc kubenswrapper[4846]: I0202 12:33:23.204075 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:23 crc kubenswrapper[4846]: I0202 12:33:23.650114 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wqsff"] Feb 02 12:33:24 crc kubenswrapper[4846]: I0202 12:33:24.435986 4846 generic.go:334] "Generic (PLEG): container finished" podID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerID="7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa" exitCode=0 Feb 02 12:33:24 crc kubenswrapper[4846]: I0202 12:33:24.436047 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wqsff" event={"ID":"775f235b-175e-4a8b-88cd-3e3f735949ec","Type":"ContainerDied","Data":"7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa"} Feb 02 12:33:24 crc kubenswrapper[4846]: I0202 12:33:24.436395 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wqsff" event={"ID":"775f235b-175e-4a8b-88cd-3e3f735949ec","Type":"ContainerStarted","Data":"fdf1aa4ad9ec307a910916aca0263be6f42f902d267e0d848db37cae7cc15e1d"} Feb 02 12:33:24 crc kubenswrapper[4846]: I0202 12:33:24.437926 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 12:33:24 crc kubenswrapper[4846]: E0202 12:33:24.822724 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:24 crc kubenswrapper[4846]: E0202 12:33:24.823562 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:24 crc kubenswrapper[4846]: E0202 12:33:24.824043 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 02 12:33:24 crc kubenswrapper[4846]: E0202 12:33:24.824083 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" Feb 02 12:33:24 crc kubenswrapper[4846]: E0202 12:33:24.824717 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:24 crc kubenswrapper[4846]: E0202 12:33:24.826809 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:24 crc kubenswrapper[4846]: E0202 12:33:24.828367 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 02 12:33:24 crc kubenswrapper[4846]: E0202 12:33:24.828428 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-m79vg" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovs-vswitchd" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.453705 4846 generic.go:334] "Generic (PLEG): container finished" podID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerID="94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473" exitCode=0 Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.453776 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wqsff" event={"ID":"775f235b-175e-4a8b-88cd-3e3f735949ec","Type":"ContainerDied","Data":"94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473"} Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.456424 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m79vg_53dd5f18-d111-4359-8170-102a8b30d0d9/ovs-vswitchd/0.log" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.459355 4846 generic.go:334] "Generic (PLEG): container finished" podID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" exitCode=137 Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.459392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m79vg" event={"ID":"53dd5f18-d111-4359-8170-102a8b30d0d9","Type":"ContainerDied","Data":"c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a"} Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.730288 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m79vg_53dd5f18-d111-4359-8170-102a8b30d0d9/ovs-vswitchd/0.log" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.731151 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.805687 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-run\") pod \"53dd5f18-d111-4359-8170-102a8b30d0d9\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.805765 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb7p8\" (UniqueName: \"kubernetes.io/projected/53dd5f18-d111-4359-8170-102a8b30d0d9-kube-api-access-lb7p8\") pod \"53dd5f18-d111-4359-8170-102a8b30d0d9\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.805825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53dd5f18-d111-4359-8170-102a8b30d0d9-scripts\") pod \"53dd5f18-d111-4359-8170-102a8b30d0d9\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.805845 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-lib\") pod \"53dd5f18-d111-4359-8170-102a8b30d0d9\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.805861 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-run" (OuterVolumeSpecName: "var-run") pod "53dd5f18-d111-4359-8170-102a8b30d0d9" (UID: "53dd5f18-d111-4359-8170-102a8b30d0d9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.805931 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-lib" (OuterVolumeSpecName: "var-lib") pod "53dd5f18-d111-4359-8170-102a8b30d0d9" (UID: "53dd5f18-d111-4359-8170-102a8b30d0d9"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.806293 4846 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-lib\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.806313 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-run\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.807044 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53dd5f18-d111-4359-8170-102a8b30d0d9-scripts" (OuterVolumeSpecName: "scripts") pod "53dd5f18-d111-4359-8170-102a8b30d0d9" (UID: "53dd5f18-d111-4359-8170-102a8b30d0d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.811268 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53dd5f18-d111-4359-8170-102a8b30d0d9-kube-api-access-lb7p8" (OuterVolumeSpecName: "kube-api-access-lb7p8") pod "53dd5f18-d111-4359-8170-102a8b30d0d9" (UID: "53dd5f18-d111-4359-8170-102a8b30d0d9"). InnerVolumeSpecName "kube-api-access-lb7p8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.907150 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-etc-ovs\") pod \"53dd5f18-d111-4359-8170-102a8b30d0d9\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.907489 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-log\") pod \"53dd5f18-d111-4359-8170-102a8b30d0d9\" (UID: \"53dd5f18-d111-4359-8170-102a8b30d0d9\") " Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.907412 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "53dd5f18-d111-4359-8170-102a8b30d0d9" (UID: "53dd5f18-d111-4359-8170-102a8b30d0d9"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.907643 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-log" (OuterVolumeSpecName: "var-log") pod "53dd5f18-d111-4359-8170-102a8b30d0d9" (UID: "53dd5f18-d111-4359-8170-102a8b30d0d9"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.908009 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb7p8\" (UniqueName: \"kubernetes.io/projected/53dd5f18-d111-4359-8170-102a8b30d0d9-kube-api-access-lb7p8\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.908031 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53dd5f18-d111-4359-8170-102a8b30d0d9-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.908044 4846 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-etc-ovs\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:26 crc kubenswrapper[4846]: I0202 12:33:26.908073 4846 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/53dd5f18-d111-4359-8170-102a8b30d0d9-var-log\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.180524 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.210741 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2ctx\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-kube-api-access-b2ctx\") pod \"309e8cc0-3b70-4961-b69c-1abf373f3c61\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.210797 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309e8cc0-3b70-4961-b69c-1abf373f3c61-combined-ca-bundle\") pod \"309e8cc0-3b70-4961-b69c-1abf373f3c61\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.210826 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift\") pod \"309e8cc0-3b70-4961-b69c-1abf373f3c61\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.210855 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-cache\") pod \"309e8cc0-3b70-4961-b69c-1abf373f3c61\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.210880 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"309e8cc0-3b70-4961-b69c-1abf373f3c61\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.210907 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-lock\") pod \"309e8cc0-3b70-4961-b69c-1abf373f3c61\" (UID: \"309e8cc0-3b70-4961-b69c-1abf373f3c61\") " Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.211565 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-lock" (OuterVolumeSpecName: "lock") pod "309e8cc0-3b70-4961-b69c-1abf373f3c61" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.211855 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-cache" (OuterVolumeSpecName: "cache") pod "309e8cc0-3b70-4961-b69c-1abf373f3c61" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.215849 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "309e8cc0-3b70-4961-b69c-1abf373f3c61" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.217306 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "swift") pod "309e8cc0-3b70-4961-b69c-1abf373f3c61" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.230881 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-kube-api-access-b2ctx" (OuterVolumeSpecName: "kube-api-access-b2ctx") pod "309e8cc0-3b70-4961-b69c-1abf373f3c61" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61"). InnerVolumeSpecName "kube-api-access-b2ctx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.321841 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2ctx\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-kube-api-access-b2ctx\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.321873 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/309e8cc0-3b70-4961-b69c-1abf373f3c61-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.321882 4846 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-cache\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.321902 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.321910 4846 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/309e8cc0-3b70-4961-b69c-1abf373f3c61-lock\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.383254 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.423316 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.469375 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m79vg_53dd5f18-d111-4359-8170-102a8b30d0d9/ovs-vswitchd/0.log" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.469999 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m79vg" event={"ID":"53dd5f18-d111-4359-8170-102a8b30d0d9","Type":"ContainerDied","Data":"80ba3f6f664654c9815bb893f341567415125f9d92fad00f7248ac217eba3452"} Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.470038 4846 scope.go:117] "RemoveContainer" containerID="c0ee8b47cb907fc18adbd7c1432d7146c455530249af82e0ea3bbb36b9734a4a" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.470063 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-m79vg" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.475521 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wqsff" event={"ID":"775f235b-175e-4a8b-88cd-3e3f735949ec","Type":"ContainerStarted","Data":"9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6"} Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.497136 4846 generic.go:334] "Generic (PLEG): container finished" podID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerID="c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455" exitCode=137 Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.497225 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455"} Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.497304 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.497328 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"309e8cc0-3b70-4961-b69c-1abf373f3c61","Type":"ContainerDied","Data":"a3a0e95c20fcca3b1fd57686cceb9b66f09a011e6eaf01e68d9543c52646d542"} Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.497503 4846 scope.go:117] "RemoveContainer" containerID="7a740ed13f137826ca022377207bbdf38cfe88459269d833b4cfd3f48d422481" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.502678 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wqsff" podStartSLOduration=3.111252852 podStartE2EDuration="5.502665904s" podCreationTimestamp="2026-02-02 12:33:22 +0000 UTC" firstStartedPulling="2026-02-02 12:33:24.437658738 +0000 UTC m=+1435.666245611" lastFinishedPulling="2026-02-02 12:33:26.8290718 +0000 UTC m=+1438.057658663" observedRunningTime="2026-02-02 12:33:27.498177094 +0000 UTC m=+1438.726763957" watchObservedRunningTime="2026-02-02 12:33:27.502665904 +0000 UTC m=+1438.731252767" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.526495 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-m79vg"] Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.531799 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-m79vg"] Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.535446 4846 scope.go:117] "RemoveContainer" containerID="5f54c9521f3ba3a9cace7a633d97e116ed3332a7c43fda1e5db49eeb3187a18b" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.560956 4846 scope.go:117] "RemoveContainer" containerID="c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.569007 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/309e8cc0-3b70-4961-b69c-1abf373f3c61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "309e8cc0-3b70-4961-b69c-1abf373f3c61" (UID: "309e8cc0-3b70-4961-b69c-1abf373f3c61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.577348 4846 scope.go:117] "RemoveContainer" containerID="e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.591900 4846 scope.go:117] "RemoveContainer" containerID="8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.608112 4846 scope.go:117] "RemoveContainer" containerID="38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.625756 4846 scope.go:117] "RemoveContainer" containerID="96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.627328 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/309e8cc0-3b70-4961-b69c-1abf373f3c61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.643956 4846 scope.go:117] "RemoveContainer" containerID="5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.677541 4846 scope.go:117] "RemoveContainer" containerID="29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.725761 4846 scope.go:117] "RemoveContainer" containerID="80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.743754 4846 scope.go:117] "RemoveContainer" containerID="0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.761894 4846 scope.go:117] "RemoveContainer" containerID="0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.778858 4846 scope.go:117] "RemoveContainer" containerID="9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.804498 4846 scope.go:117] "RemoveContainer" containerID="387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.827754 4846 scope.go:117] "RemoveContainer" containerID="6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.846610 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.856187 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.856890 4846 scope.go:117] "RemoveContainer" containerID="28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.884698 4846 scope.go:117] "RemoveContainer" containerID="eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.907139 4846 scope.go:117] "RemoveContainer" containerID="c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.907858 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455\": container with ID starting with c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455 not found: ID does not exist" containerID="c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.907899 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455"} err="failed to get container status \"c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455\": rpc error: code = NotFound desc = could not find container \"c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455\": container with ID starting with c510cf17a016bcd93f3cd7119c4a16bcc5fffbd6f8e2a4f89f42657ca7924455 not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.907926 4846 scope.go:117] "RemoveContainer" containerID="e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.908255 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064\": container with ID starting with e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064 not found: ID does not exist" containerID="e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.908305 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064"} err="failed to get container status \"e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064\": rpc error: code = NotFound desc = could not find container \"e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064\": container with ID starting with e3dac67290e36a72f1ef431d206d979b95be5ee43e9b6dca2dfadcf80a50c064 not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.908338 4846 scope.go:117] "RemoveContainer" containerID="8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.908746 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae\": container with ID starting with 8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae not found: ID does not exist" containerID="8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.908793 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae"} err="failed to get container status \"8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae\": rpc error: code = NotFound desc = could not find container \"8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae\": container with ID starting with 8252b00aaa1450cd82eeb8793d4ae360c37bfee7a85180005f5d23afd0e9c0ae not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.908813 4846 scope.go:117] "RemoveContainer" containerID="38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.909147 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558\": container with ID starting with 38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558 not found: ID does not exist" containerID="38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.909185 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558"} err="failed to get container status \"38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558\": rpc error: code = NotFound desc = could not find container \"38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558\": container with ID starting with 38018a962df143ed39151fdd641f3bb3c36b9b50ec1583e1b6ec96e1bdff0558 not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.909204 4846 scope.go:117] "RemoveContainer" containerID="96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.909496 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891\": container with ID starting with 96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891 not found: ID does not exist" containerID="96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.909530 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891"} err="failed to get container status \"96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891\": rpc error: code = NotFound desc = could not find container \"96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891\": container with ID starting with 96dae211deb5973be0c98b758d8a415e3e072ec46c4397950c4361d5a8083891 not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.909552 4846 scope.go:117] "RemoveContainer" containerID="5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.909948 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c\": container with ID starting with 5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c not found: ID does not exist" containerID="5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.909986 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c"} err="failed to get container status \"5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c\": rpc error: code = NotFound desc = could not find container \"5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c\": container with ID starting with 5dcd0fb5e69494008813c7b7170507dd442d75cdc694ce102db08f0cdf65fc9c not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.910008 4846 scope.go:117] "RemoveContainer" containerID="29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.910262 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3\": container with ID starting with 29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3 not found: ID does not exist" containerID="29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.910286 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3"} err="failed to get container status \"29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3\": rpc error: code = NotFound desc = could not find container \"29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3\": container with ID starting with 29c5c5a70572832660493356b1f69d442ad55f5b285acba4862b0460d4d2d2d3 not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.910300 4846 scope.go:117] "RemoveContainer" containerID="80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.911012 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7\": container with ID starting with 80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7 not found: ID does not exist" containerID="80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.911033 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7"} err="failed to get container status \"80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7\": rpc error: code = NotFound desc = could not find container \"80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7\": container with ID starting with 80e7154c81938c96e444e125da18ab6fd5dfd321f09ebdf8d3fe9125822b91a7 not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.911045 4846 scope.go:117] "RemoveContainer" containerID="0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.911310 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445\": container with ID starting with 0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445 not found: ID does not exist" containerID="0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.911358 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445"} err="failed to get container status \"0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445\": rpc error: code = NotFound desc = could not find container \"0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445\": container with ID starting with 0d7710fc98963d2f8ec110a6e1ffe7c63399cc0ab3322959e7731d7a7937b445 not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.911378 4846 scope.go:117] "RemoveContainer" containerID="0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.911933 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728\": container with ID starting with 0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728 not found: ID does not exist" containerID="0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.911966 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728"} err="failed to get container status \"0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728\": rpc error: code = NotFound desc = could not find container \"0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728\": container with ID starting with 0bbe6e70b222a62ee9f1abb9b8d2ff92317ecc965d901cc03b9f241ef2d01728 not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.911982 4846 scope.go:117] "RemoveContainer" containerID="9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.912425 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d\": container with ID starting with 9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d not found: ID does not exist" containerID="9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.912462 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d"} err="failed to get container status \"9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d\": rpc error: code = NotFound desc = could not find container \"9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d\": container with ID starting with 9f60f8f6b23122ea120ebf44b4ada3597d2db896d467bd2a4c2ba767e9317e3d not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.912479 4846 scope.go:117] "RemoveContainer" containerID="387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.912769 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3\": container with ID starting with 387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3 not found: ID does not exist" containerID="387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.912791 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3"} err="failed to get container status \"387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3\": rpc error: code = NotFound desc = could not find container \"387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3\": container with ID starting with 387cf760b9eaae46c41228a2384a3ca874fdd4fd7327387efd4e69bc434c45a3 not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.912805 4846 scope.go:117] "RemoveContainer" containerID="6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.913126 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864\": container with ID starting with 6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864 not found: ID does not exist" containerID="6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.913168 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864"} err="failed to get container status \"6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864\": rpc error: code = NotFound desc = could not find container \"6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864\": container with ID starting with 6da1f8acd7438bd680d7fc325b2047c0ef75a95bdc7e94904c30d35f781e4864 not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.913197 4846 scope.go:117] "RemoveContainer" containerID="28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.913423 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f\": container with ID starting with 28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f not found: ID does not exist" containerID="28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.913451 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f"} err="failed to get container status \"28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f\": rpc error: code = NotFound desc = could not find container \"28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f\": container with ID starting with 28e5deb120abecd02e495001af96188fa84cd04bab80edddc36ce1d7a4a7b20f not found: ID does not exist" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.913465 4846 scope.go:117] "RemoveContainer" containerID="eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d" Feb 02 12:33:27 crc kubenswrapper[4846]: E0202 12:33:27.913632 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d\": container with ID starting with eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d not found: ID does not exist" containerID="eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d" Feb 02 12:33:27 crc kubenswrapper[4846]: I0202 12:33:27.913650 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d"} err="failed to get container status \"eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d\": rpc error: code = NotFound desc = could not find container \"eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d\": container with ID starting with eb292c014f15f0309a0c1ec7ff4982dda361848ded6b3b1ceddab8389dbb0a0d not found: ID does not exist" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.429587 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.436905 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom\") pod \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.436953 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts\") pod \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.436978 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-combined-ca-bundle\") pod \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.437019 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/444e13f0-bc18-4bc9-8d92-af6a379e22ab-etc-machine-id\") pod \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.437050 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9tf7\" (UniqueName: \"kubernetes.io/projected/444e13f0-bc18-4bc9-8d92-af6a379e22ab-kube-api-access-l9tf7\") pod \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.437106 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data\") pod \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\" (UID: \"444e13f0-bc18-4bc9-8d92-af6a379e22ab\") " Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.437111 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/444e13f0-bc18-4bc9-8d92-af6a379e22ab-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "444e13f0-bc18-4bc9-8d92-af6a379e22ab" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.437283 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/444e13f0-bc18-4bc9-8d92-af6a379e22ab-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.441157 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/444e13f0-bc18-4bc9-8d92-af6a379e22ab-kube-api-access-l9tf7" (OuterVolumeSpecName: "kube-api-access-l9tf7") pod "444e13f0-bc18-4bc9-8d92-af6a379e22ab" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab"). InnerVolumeSpecName "kube-api-access-l9tf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.441180 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts" (OuterVolumeSpecName: "scripts") pod "444e13f0-bc18-4bc9-8d92-af6a379e22ab" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.444957 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "444e13f0-bc18-4bc9-8d92-af6a379e22ab" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.478096 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "444e13f0-bc18-4bc9-8d92-af6a379e22ab" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.507414 4846 generic.go:334] "Generic (PLEG): container finished" podID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" containerID="7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806" exitCode=137 Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.507501 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"444e13f0-bc18-4bc9-8d92-af6a379e22ab","Type":"ContainerDied","Data":"7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806"} Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.507534 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"444e13f0-bc18-4bc9-8d92-af6a379e22ab","Type":"ContainerDied","Data":"3b49c7fefa01d749cd32612414a9a283bf51dc9e40af50dd67a7ce30bc8446f4"} Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.507553 4846 scope.go:117] "RemoveContainer" containerID="642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.507741 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.529822 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data" (OuterVolumeSpecName: "config-data") pod "444e13f0-bc18-4bc9-8d92-af6a379e22ab" (UID: "444e13f0-bc18-4bc9-8d92-af6a379e22ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.533046 4846 scope.go:117] "RemoveContainer" containerID="7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.538253 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.538305 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.538362 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.538387 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444e13f0-bc18-4bc9-8d92-af6a379e22ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.538399 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9tf7\" (UniqueName: \"kubernetes.io/projected/444e13f0-bc18-4bc9-8d92-af6a379e22ab-kube-api-access-l9tf7\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.552259 4846 scope.go:117] "RemoveContainer" containerID="642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210" Feb 02 12:33:28 crc kubenswrapper[4846]: E0202 12:33:28.553350 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210\": container with ID starting with 642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210 not found: ID does not exist" containerID="642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.553400 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210"} err="failed to get container status \"642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210\": rpc error: code = NotFound desc = could not find container \"642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210\": container with ID starting with 642f56babcfca837c4ddd24d9663e16f8175a4c5c39ba92854e63b4007c5f210 not found: ID does not exist" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.553427 4846 scope.go:117] "RemoveContainer" containerID="7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806" Feb 02 12:33:28 crc kubenswrapper[4846]: E0202 12:33:28.555082 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806\": container with ID starting with 7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806 not found: ID does not exist" containerID="7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.555132 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806"} err="failed to get container status \"7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806\": rpc error: code = NotFound desc = could not find container \"7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806\": container with ID starting with 7d3cf01a6daf5d14ff271d3853e8fa5a466627574e63763fc673348312a5e806 not found: ID does not exist" Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.844749 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:33:28 crc kubenswrapper[4846]: I0202 12:33:28.851676 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 12:33:29 crc kubenswrapper[4846]: I0202 12:33:29.438333 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" path="/var/lib/kubelet/pods/309e8cc0-3b70-4961-b69c-1abf373f3c61/volumes" Feb 02 12:33:29 crc kubenswrapper[4846]: I0202 12:33:29.441608 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" path="/var/lib/kubelet/pods/444e13f0-bc18-4bc9-8d92-af6a379e22ab/volumes" Feb 02 12:33:29 crc kubenswrapper[4846]: I0202 12:33:29.442479 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" path="/var/lib/kubelet/pods/53dd5f18-d111-4359-8170-102a8b30d0d9/volumes" Feb 02 12:33:30 crc kubenswrapper[4846]: I0202 12:33:30.479199 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:33:30 crc kubenswrapper[4846]: I0202 12:33:30.479741 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:33:33 crc kubenswrapper[4846]: I0202 12:33:33.204872 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:33 crc kubenswrapper[4846]: I0202 12:33:33.205265 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:33 crc kubenswrapper[4846]: I0202 12:33:33.261107 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:33 crc kubenswrapper[4846]: I0202 12:33:33.605958 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:33 crc kubenswrapper[4846]: I0202 12:33:33.652487 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wqsff"] Feb 02 12:33:35 crc kubenswrapper[4846]: I0202 12:33:35.576422 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wqsff" podUID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerName="registry-server" containerID="cri-o://9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6" gracePeriod=2 Feb 02 12:33:35 crc kubenswrapper[4846]: I0202 12:33:35.976570 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.140175 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-catalog-content\") pod \"775f235b-175e-4a8b-88cd-3e3f735949ec\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.140268 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9zhd\" (UniqueName: \"kubernetes.io/projected/775f235b-175e-4a8b-88cd-3e3f735949ec-kube-api-access-v9zhd\") pod \"775f235b-175e-4a8b-88cd-3e3f735949ec\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.140360 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-utilities\") pod \"775f235b-175e-4a8b-88cd-3e3f735949ec\" (UID: \"775f235b-175e-4a8b-88cd-3e3f735949ec\") " Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.141383 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-utilities" (OuterVolumeSpecName: "utilities") pod "775f235b-175e-4a8b-88cd-3e3f735949ec" (UID: "775f235b-175e-4a8b-88cd-3e3f735949ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.146706 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/775f235b-175e-4a8b-88cd-3e3f735949ec-kube-api-access-v9zhd" (OuterVolumeSpecName: "kube-api-access-v9zhd") pod "775f235b-175e-4a8b-88cd-3e3f735949ec" (UID: "775f235b-175e-4a8b-88cd-3e3f735949ec"). InnerVolumeSpecName "kube-api-access-v9zhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.242566 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9zhd\" (UniqueName: \"kubernetes.io/projected/775f235b-175e-4a8b-88cd-3e3f735949ec-kube-api-access-v9zhd\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.242602 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.284182 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "775f235b-175e-4a8b-88cd-3e3f735949ec" (UID: "775f235b-175e-4a8b-88cd-3e3f735949ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.344378 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775f235b-175e-4a8b-88cd-3e3f735949ec-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.587284 4846 generic.go:334] "Generic (PLEG): container finished" podID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerID="9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6" exitCode=0 Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.587335 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wqsff" event={"ID":"775f235b-175e-4a8b-88cd-3e3f735949ec","Type":"ContainerDied","Data":"9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6"} Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.587357 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wqsff" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.587367 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wqsff" event={"ID":"775f235b-175e-4a8b-88cd-3e3f735949ec","Type":"ContainerDied","Data":"fdf1aa4ad9ec307a910916aca0263be6f42f902d267e0d848db37cae7cc15e1d"} Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.587391 4846 scope.go:117] "RemoveContainer" containerID="9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.607862 4846 scope.go:117] "RemoveContainer" containerID="94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.629752 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wqsff"] Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.635149 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wqsff"] Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.655336 4846 scope.go:117] "RemoveContainer" containerID="7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.672323 4846 scope.go:117] "RemoveContainer" containerID="9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6" Feb 02 12:33:36 crc kubenswrapper[4846]: E0202 12:33:36.672692 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6\": container with ID starting with 9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6 not found: ID does not exist" containerID="9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.672757 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6"} err="failed to get container status \"9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6\": rpc error: code = NotFound desc = could not find container \"9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6\": container with ID starting with 9bcc8b464394e90769a065262314bd6091a9a6ab276296a9b96daf31c7b239f6 not found: ID does not exist" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.672821 4846 scope.go:117] "RemoveContainer" containerID="94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473" Feb 02 12:33:36 crc kubenswrapper[4846]: E0202 12:33:36.673186 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473\": container with ID starting with 94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473 not found: ID does not exist" containerID="94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.673222 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473"} err="failed to get container status \"94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473\": rpc error: code = NotFound desc = could not find container \"94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473\": container with ID starting with 94d76b888703f14238bb1f88432d98d32a286a248248801a74bbc2af96c64473 not found: ID does not exist" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.673243 4846 scope.go:117] "RemoveContainer" containerID="7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa" Feb 02 12:33:36 crc kubenswrapper[4846]: E0202 12:33:36.673564 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa\": container with ID starting with 7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa not found: ID does not exist" containerID="7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa" Feb 02 12:33:36 crc kubenswrapper[4846]: I0202 12:33:36.673605 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa"} err="failed to get container status \"7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa\": rpc error: code = NotFound desc = could not find container \"7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa\": container with ID starting with 7d48cfec9ccee229ecb41f8658426297d2ce932c7f2bb99b46b272c6ec9678aa not found: ID does not exist" Feb 02 12:33:37 crc kubenswrapper[4846]: I0202 12:33:37.436894 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="775f235b-175e-4a8b-88cd-3e3f735949ec" path="/var/lib/kubelet/pods/775f235b-175e-4a8b-88cd-3e3f735949ec/volumes" Feb 02 12:34:00 crc kubenswrapper[4846]: I0202 12:34:00.479096 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:34:00 crc kubenswrapper[4846]: I0202 12:34:00.479675 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:34:30 crc kubenswrapper[4846]: I0202 12:34:30.478864 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:34:30 crc kubenswrapper[4846]: I0202 12:34:30.479305 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:34:30 crc kubenswrapper[4846]: I0202 12:34:30.479347 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:34:30 crc kubenswrapper[4846]: I0202 12:34:30.480109 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:34:30 crc kubenswrapper[4846]: I0202 12:34:30.480161 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" gracePeriod=600 Feb 02 12:34:30 crc kubenswrapper[4846]: E0202 12:34:30.601030 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:34:31 crc kubenswrapper[4846]: I0202 12:34:31.130464 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" exitCode=0 Feb 02 12:34:31 crc kubenswrapper[4846]: I0202 12:34:31.130550 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b"} Feb 02 12:34:31 crc kubenswrapper[4846]: I0202 12:34:31.130863 4846 scope.go:117] "RemoveContainer" containerID="5342c3479165e81b2ca97736d850f5b8bc8a41ed04e9666168601ccf61d9277c" Feb 02 12:34:31 crc kubenswrapper[4846]: I0202 12:34:31.131494 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:34:31 crc kubenswrapper[4846]: E0202 12:34:31.131877 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:34:34 crc kubenswrapper[4846]: I0202 12:34:34.164998 4846 scope.go:117] "RemoveContainer" containerID="7e20049cbc1c96a32d8a5a8a599eef2e575b73f6df35000995ce4b76e4f2a280" Feb 02 12:34:34 crc kubenswrapper[4846]: I0202 12:34:34.194431 4846 scope.go:117] "RemoveContainer" containerID="9baac131245f62b73851f82758eac3beae324a624e97690d13a59f6f6d3e2bd8" Feb 02 12:34:34 crc kubenswrapper[4846]: I0202 12:34:34.235565 4846 scope.go:117] "RemoveContainer" containerID="6e21400ac8e3c4e79243ecc8b3b8272649285daca0a5f4b7d6f4cc9592ac49c6" Feb 02 12:34:34 crc kubenswrapper[4846]: I0202 12:34:34.273402 4846 scope.go:117] "RemoveContainer" containerID="1d67d357e4081d3bfaa9f34fd59e7809db1eb45b38eb505b82cd78ccaa91a725" Feb 02 12:34:34 crc kubenswrapper[4846]: I0202 12:34:34.295284 4846 scope.go:117] "RemoveContainer" containerID="be8dea291f883c78f6a9fe42a55b49ce1c88d18d046987607886c951725e9c18" Feb 02 12:34:44 crc kubenswrapper[4846]: I0202 12:34:44.423817 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:34:44 crc kubenswrapper[4846]: E0202 12:34:44.424870 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:34:55 crc kubenswrapper[4846]: I0202 12:34:55.424925 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:34:55 crc kubenswrapper[4846]: E0202 12:34:55.426002 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:35:09 crc kubenswrapper[4846]: I0202 12:35:09.431074 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:35:09 crc kubenswrapper[4846]: E0202 12:35:09.431739 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:35:23 crc kubenswrapper[4846]: I0202 12:35:23.423836 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:35:23 crc kubenswrapper[4846]: E0202 12:35:23.425186 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.390576 4846 scope.go:117] "RemoveContainer" containerID="6131396e4bf56b3a277f471ab103da5529eb4c3c786f308bb7fbd7abf06db02d" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.421895 4846 scope.go:117] "RemoveContainer" containerID="0c052e8fc632fa57acc71fa4082dddaf8f00051c9a1fdbd9fa203c55647113eb" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.456924 4846 scope.go:117] "RemoveContainer" containerID="c9dc6241864be92bed2674514ffa4f13c2775741f26cfa057ce276cb569c8226" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.487754 4846 scope.go:117] "RemoveContainer" containerID="49a43bd07d83a9a7f59d576c12458101178e1ae9330d85850931baa6d139ed2f" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.533194 4846 scope.go:117] "RemoveContainer" containerID="052031d7c9748afe7ec26c0238c176702ba54688df56666d646d445914895bf0" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.554986 4846 scope.go:117] "RemoveContainer" containerID="1aca36bbf97855cd72821d99a63c0d8ff444d4e356394fb4c879014ec3b75fd5" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.575148 4846 scope.go:117] "RemoveContainer" containerID="980217af7f9420706d5f9938b02bc23ed28998d36f06b2cc585c6fa0c1914b8e" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.595681 4846 scope.go:117] "RemoveContainer" containerID="0badaa9fb9c046619f3874d44014b8116430f9025733765fa1163fcae1ed3d30" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.616045 4846 scope.go:117] "RemoveContainer" containerID="82c7e76bff98e984e34c1c47e1a89029be4cfa11944cf5e9fa3ccf1450ac3a6e" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.633845 4846 scope.go:117] "RemoveContainer" containerID="8317181145f1669cb74990cc6c775fa122e6ab9aacd49fce8490f28352e2c57f" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.674712 4846 scope.go:117] "RemoveContainer" containerID="990eec9ad0d43277d54b187042320a08193aaca93bf2eb669cefbe31ab8568b3" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.710783 4846 scope.go:117] "RemoveContainer" containerID="64fad55f3bd27a34f1da83a5dc31008972283b2c847e98e400a85649b868215e" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.741130 4846 scope.go:117] "RemoveContainer" containerID="aa70f3c1cb4d9ccf0a47e1cabbf6e0b6f5b91bf416c3d8de28edf8ad6a45b351" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.760968 4846 scope.go:117] "RemoveContainer" containerID="66243d036d20a71042e4208e8b9c88b63381e5c9b16d891c7d371cdcefee7420" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.783953 4846 scope.go:117] "RemoveContainer" containerID="61857b150929bc24955da85821ee7c4e71280ebcd4d2731ae05aa0b38a82d290" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.817413 4846 scope.go:117] "RemoveContainer" containerID="6dbe9840669d94a9aa14d90cfc98a8e3f74cc3f30a489967e47e07ffa42a5e82" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.839442 4846 scope.go:117] "RemoveContainer" containerID="2acbbb7c816156b529232e467d5fe13563d39a52756e14b37ed7aa312ee11088" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.865138 4846 scope.go:117] "RemoveContainer" containerID="af62214010ef6c9ebee8a7b9733f31b2fde1d938f00d53e230f39c1f885fa12e" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.883411 4846 scope.go:117] "RemoveContainer" containerID="c62f51d03fd9f435dd3c6722b2d45f51aea94cdc0dc8bea514ed230dba6a1885" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.909061 4846 scope.go:117] "RemoveContainer" containerID="8774cd543521690df2b64182c81e025daf2b49960381c918588bd9af86e76a03" Feb 02 12:35:34 crc kubenswrapper[4846]: I0202 12:35:34.926528 4846 scope.go:117] "RemoveContainer" containerID="6ed6bdf55d9e56b0bdeb412f1c47fec912bfa311c417c14a15454e3c3c8395eb" Feb 02 12:35:35 crc kubenswrapper[4846]: I0202 12:35:35.424224 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:35:35 crc kubenswrapper[4846]: E0202 12:35:35.424558 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:35:47 crc kubenswrapper[4846]: I0202 12:35:47.423422 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:35:47 crc kubenswrapper[4846]: E0202 12:35:47.424527 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:36:00 crc kubenswrapper[4846]: I0202 12:36:00.423080 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:36:00 crc kubenswrapper[4846]: E0202 12:36:00.423806 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:36:12 crc kubenswrapper[4846]: I0202 12:36:12.423784 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:36:12 crc kubenswrapper[4846]: E0202 12:36:12.424566 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:36:23 crc kubenswrapper[4846]: I0202 12:36:23.423464 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:36:23 crc kubenswrapper[4846]: E0202 12:36:23.424189 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:36:35 crc kubenswrapper[4846]: I0202 12:36:35.233526 4846 scope.go:117] "RemoveContainer" containerID="4030d8eaddee56f1991428cfeacaa9dffb7cdec1411b2eaff3e458cb869e836c" Feb 02 12:36:35 crc kubenswrapper[4846]: I0202 12:36:35.265011 4846 scope.go:117] "RemoveContainer" containerID="a8b3459915c6a0efed9d2ec47db4ee7ad0d37fbe1e2d2df99bb10ca0e374e45f" Feb 02 12:36:35 crc kubenswrapper[4846]: I0202 12:36:35.301909 4846 scope.go:117] "RemoveContainer" containerID="cd9c89a15c32fa01cbe6922310a47bcd10ec6d5810bd28c0c659192a699b7d01" Feb 02 12:36:35 crc kubenswrapper[4846]: I0202 12:36:35.327115 4846 scope.go:117] "RemoveContainer" containerID="a7c5dc480cc9c268bfef70458e359db736dad2aa52a7110af88fdcd9bdce0c4a" Feb 02 12:36:35 crc kubenswrapper[4846]: I0202 12:36:35.349800 4846 scope.go:117] "RemoveContainer" containerID="6d07b794031104d16f41e77ce48912977df34ffc3a88d5151c8b49bbc089573c" Feb 02 12:36:35 crc kubenswrapper[4846]: I0202 12:36:35.381290 4846 scope.go:117] "RemoveContainer" containerID="6af2e184bc53c534dc08e92be427ce1c5bdc36efdacf88e01825552ab6c352fb" Feb 02 12:36:35 crc kubenswrapper[4846]: I0202 12:36:35.400431 4846 scope.go:117] "RemoveContainer" containerID="ca013b760fcb0dbd5895aba5a0850c6f7f8aa9da7493a41b415f4331df665b88" Feb 02 12:36:38 crc kubenswrapper[4846]: I0202 12:36:38.423482 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:36:38 crc kubenswrapper[4846]: E0202 12:36:38.423823 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:36:51 crc kubenswrapper[4846]: I0202 12:36:51.423526 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:36:51 crc kubenswrapper[4846]: E0202 12:36:51.424207 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:37:05 crc kubenswrapper[4846]: I0202 12:37:05.424301 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:37:05 crc kubenswrapper[4846]: E0202 12:37:05.425134 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:37:16 crc kubenswrapper[4846]: I0202 12:37:16.423037 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:37:16 crc kubenswrapper[4846]: E0202 12:37:16.423812 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:37:28 crc kubenswrapper[4846]: I0202 12:37:28.423304 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:37:28 crc kubenswrapper[4846]: E0202 12:37:28.424098 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:37:35 crc kubenswrapper[4846]: I0202 12:37:35.509396 4846 scope.go:117] "RemoveContainer" containerID="b21926f23c5b1a73584f4ed6528b202d36ba1f36ff760cbf9ed6a2eaf4aaf751" Feb 02 12:37:35 crc kubenswrapper[4846]: I0202 12:37:35.563030 4846 scope.go:117] "RemoveContainer" containerID="b89c77647ebfc096292368b4c128270a751702e9c4f3ecf26b326d816dff5fbc" Feb 02 12:37:35 crc kubenswrapper[4846]: I0202 12:37:35.588385 4846 scope.go:117] "RemoveContainer" containerID="2e665932988445cbfa0eb4e53f1f8f931bda8deb94a40abbec8844a52aed5529" Feb 02 12:37:35 crc kubenswrapper[4846]: I0202 12:37:35.613787 4846 scope.go:117] "RemoveContainer" containerID="1b9592c0a9b70ced628973c2dd7cfe317289df85e6586e3fd6bf1d7f4fd1f2d9" Feb 02 12:37:35 crc kubenswrapper[4846]: I0202 12:37:35.655235 4846 scope.go:117] "RemoveContainer" containerID="a3bf2b4a419820c2d36fe369e8dac05f3052d401b7c0bdcc51a1cdf750ef3a24" Feb 02 12:37:35 crc kubenswrapper[4846]: I0202 12:37:35.673459 4846 scope.go:117] "RemoveContainer" containerID="dbd55a0c1904df42f4b3eb5b57e2aa4fadf58f7e576128bd91743f3ff542702f" Feb 02 12:37:35 crc kubenswrapper[4846]: I0202 12:37:35.689676 4846 scope.go:117] "RemoveContainer" containerID="fe143b8c77eda416e18ab2d5f5660e969ca46e9b40cf1c43b4cfcf689c72cfa1" Feb 02 12:37:35 crc kubenswrapper[4846]: I0202 12:37:35.706007 4846 scope.go:117] "RemoveContainer" containerID="3c7bbc02055a0ecada54268b37ac8bc22b7506ea3e81485f4bcfe67d74919ccc" Feb 02 12:37:35 crc kubenswrapper[4846]: I0202 12:37:35.747225 4846 scope.go:117] "RemoveContainer" containerID="5e87284be07cce9a9217396923aec40f910b6e6f2a348d1a35df8d98c6249d80" Feb 02 12:37:39 crc kubenswrapper[4846]: I0202 12:37:39.428659 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:37:39 crc kubenswrapper[4846]: E0202 12:37:39.429220 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:37:52 crc kubenswrapper[4846]: I0202 12:37:52.423938 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:37:52 crc kubenswrapper[4846]: E0202 12:37:52.425205 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:38:04 crc kubenswrapper[4846]: I0202 12:38:04.423201 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:38:04 crc kubenswrapper[4846]: E0202 12:38:04.424801 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:38:19 crc kubenswrapper[4846]: I0202 12:38:19.428715 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:38:19 crc kubenswrapper[4846]: E0202 12:38:19.429607 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:38:30 crc kubenswrapper[4846]: I0202 12:38:30.423736 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:38:30 crc kubenswrapper[4846]: E0202 12:38:30.424462 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:38:35 crc kubenswrapper[4846]: I0202 12:38:35.869940 4846 scope.go:117] "RemoveContainer" containerID="ceb10096d86d9be3fc00e39352a6fcbaf81ecf7bd346d3e24a99c69fe60cd9f2" Feb 02 12:38:35 crc kubenswrapper[4846]: I0202 12:38:35.894833 4846 scope.go:117] "RemoveContainer" containerID="e4a716613f696100470f5f89eb0a597105b79f131836ad346870b04f41dfc79d" Feb 02 12:38:35 crc kubenswrapper[4846]: I0202 12:38:35.937269 4846 scope.go:117] "RemoveContainer" containerID="99b0a8947603d649bb9d1f3decf69b9ee0004e7f9c373151604124884a5d8445" Feb 02 12:38:42 crc kubenswrapper[4846]: I0202 12:38:42.423799 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:38:42 crc kubenswrapper[4846]: E0202 12:38:42.424544 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:38:56 crc kubenswrapper[4846]: I0202 12:38:56.423143 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:38:56 crc kubenswrapper[4846]: E0202 12:38:56.423671 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:39:10 crc kubenswrapper[4846]: I0202 12:39:10.423883 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:39:10 crc kubenswrapper[4846]: E0202 12:39:10.424580 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:39:25 crc kubenswrapper[4846]: I0202 12:39:25.423225 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:39:25 crc kubenswrapper[4846]: E0202 12:39:25.424138 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:39:36 crc kubenswrapper[4846]: I0202 12:39:36.423699 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:39:36 crc kubenswrapper[4846]: I0202 12:39:36.663556 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"410ee2677ceebfefaa85449243876d89f0c44fbef81405ad99a3ceddb58f3e7c"} Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.352096 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-22rv2"] Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.353798 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-auditor" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.353816 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-auditor" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.353827 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerName="registry-server" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.353835 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerName="registry-server" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.353851 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" containerName="probe" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.353859 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" containerName="probe" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.353869 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="rsync" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.353877 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="rsync" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.353889 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server-init" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.353897 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server-init" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.353908 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-updater" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.353916 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-updater" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.353928 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-replicator" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.353935 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-replicator" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.353949 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-replicator" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.353955 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-replicator" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.353967 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerName="extract-content" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.353975 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerName="extract-content" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.353988 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-server" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.353997 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-server" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354008 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-server" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354017 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-server" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354027 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" containerName="cinder-scheduler" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354035 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" containerName="cinder-scheduler" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354044 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-reaper" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354051 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-reaper" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354060 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-replicator" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354067 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-replicator" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354080 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-auditor" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354088 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-auditor" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354097 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="swift-recon-cron" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354104 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="swift-recon-cron" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354114 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354121 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354133 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-expirer" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354140 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-expirer" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354154 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-updater" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354161 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-updater" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354172 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerName="extract-utilities" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354179 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerName="extract-utilities" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354191 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-auditor" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354198 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-auditor" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354209 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-server" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354216 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-server" Feb 02 12:40:39 crc kubenswrapper[4846]: E0202 12:40:39.354227 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovs-vswitchd" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354234 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovs-vswitchd" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354582 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="775f235b-175e-4a8b-88cd-3e3f735949ec" containerName="registry-server" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354639 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-auditor" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354652 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovs-vswitchd" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354663 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="53dd5f18-d111-4359-8170-102a8b30d0d9" containerName="ovsdb-server" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354673 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-replicator" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354686 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-server" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354697 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-server" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354705 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-replicator" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354782 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="account-reaper" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354798 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-replicator" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354808 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-updater" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354820 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" containerName="probe" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354832 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="444e13f0-bc18-4bc9-8d92-af6a379e22ab" containerName="cinder-scheduler" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354856 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-updater" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354867 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-expirer" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354878 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="rsync" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354895 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="container-auditor" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354910 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="swift-recon-cron" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.354922 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-auditor" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.356339 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="309e8cc0-3b70-4961-b69c-1abf373f3c61" containerName="object-server" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.357640 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.369116 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-22rv2"] Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.465035 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8sxd\" (UniqueName: \"kubernetes.io/projected/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-kube-api-access-d8sxd\") pod \"redhat-marketplace-22rv2\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.465089 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-utilities\") pod \"redhat-marketplace-22rv2\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.465954 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-catalog-content\") pod \"redhat-marketplace-22rv2\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.538641 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hc8xt"] Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.540213 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.547955 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hc8xt"] Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.581269 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-catalog-content\") pod \"redhat-marketplace-22rv2\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.581342 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8sxd\" (UniqueName: \"kubernetes.io/projected/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-kube-api-access-d8sxd\") pod \"redhat-marketplace-22rv2\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.581401 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-utilities\") pod \"redhat-marketplace-22rv2\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.582053 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-utilities\") pod \"redhat-marketplace-22rv2\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.582204 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-catalog-content\") pod \"redhat-marketplace-22rv2\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.606794 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8sxd\" (UniqueName: \"kubernetes.io/projected/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-kube-api-access-d8sxd\") pod \"redhat-marketplace-22rv2\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.682867 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdgvk\" (UniqueName: \"kubernetes.io/projected/22bb6443-0edd-4312-b15f-e36efb8113d4-kube-api-access-xdgvk\") pod \"community-operators-hc8xt\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.682943 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-catalog-content\") pod \"community-operators-hc8xt\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.683001 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-utilities\") pod \"community-operators-hc8xt\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.687600 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.784110 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdgvk\" (UniqueName: \"kubernetes.io/projected/22bb6443-0edd-4312-b15f-e36efb8113d4-kube-api-access-xdgvk\") pod \"community-operators-hc8xt\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.784149 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-catalog-content\") pod \"community-operators-hc8xt\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.784187 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-utilities\") pod \"community-operators-hc8xt\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.784690 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-utilities\") pod \"community-operators-hc8xt\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.785195 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-catalog-content\") pod \"community-operators-hc8xt\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.807726 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdgvk\" (UniqueName: \"kubernetes.io/projected/22bb6443-0edd-4312-b15f-e36efb8113d4-kube-api-access-xdgvk\") pod \"community-operators-hc8xt\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:39 crc kubenswrapper[4846]: I0202 12:40:39.898913 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:40 crc kubenswrapper[4846]: I0202 12:40:40.161458 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-22rv2"] Feb 02 12:40:40 crc kubenswrapper[4846]: I0202 12:40:40.424119 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hc8xt"] Feb 02 12:40:40 crc kubenswrapper[4846]: W0202 12:40:40.426547 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22bb6443_0edd_4312_b15f_e36efb8113d4.slice/crio-0170f50aed775674a8b58b10adc301571eff0729dc58355cb6997d95b8b784be WatchSource:0}: Error finding container 0170f50aed775674a8b58b10adc301571eff0729dc58355cb6997d95b8b784be: Status 404 returned error can't find the container with id 0170f50aed775674a8b58b10adc301571eff0729dc58355cb6997d95b8b784be Feb 02 12:40:41 crc kubenswrapper[4846]: I0202 12:40:41.140104 4846 generic.go:334] "Generic (PLEG): container finished" podID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerID="db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e" exitCode=0 Feb 02 12:40:41 crc kubenswrapper[4846]: I0202 12:40:41.140217 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22rv2" event={"ID":"11828798-5abc-48cf-9f2e-ae69d2c7ffd5","Type":"ContainerDied","Data":"db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e"} Feb 02 12:40:41 crc kubenswrapper[4846]: I0202 12:40:41.140277 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22rv2" event={"ID":"11828798-5abc-48cf-9f2e-ae69d2c7ffd5","Type":"ContainerStarted","Data":"ddd77f8cd62457a9954e14e966bd7ff2592ddc73f5d3524ebe7139b7ac95c829"} Feb 02 12:40:41 crc kubenswrapper[4846]: I0202 12:40:41.141769 4846 generic.go:334] "Generic (PLEG): container finished" podID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerID="deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100" exitCode=0 Feb 02 12:40:41 crc kubenswrapper[4846]: I0202 12:40:41.141806 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc8xt" event={"ID":"22bb6443-0edd-4312-b15f-e36efb8113d4","Type":"ContainerDied","Data":"deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100"} Feb 02 12:40:41 crc kubenswrapper[4846]: I0202 12:40:41.141829 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc8xt" event={"ID":"22bb6443-0edd-4312-b15f-e36efb8113d4","Type":"ContainerStarted","Data":"0170f50aed775674a8b58b10adc301571eff0729dc58355cb6997d95b8b784be"} Feb 02 12:40:41 crc kubenswrapper[4846]: I0202 12:40:41.143399 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 12:40:42 crc kubenswrapper[4846]: I0202 12:40:42.150153 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc8xt" event={"ID":"22bb6443-0edd-4312-b15f-e36efb8113d4","Type":"ContainerStarted","Data":"eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad"} Feb 02 12:40:42 crc kubenswrapper[4846]: I0202 12:40:42.153373 4846 generic.go:334] "Generic (PLEG): container finished" podID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerID="744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4" exitCode=0 Feb 02 12:40:42 crc kubenswrapper[4846]: I0202 12:40:42.153412 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22rv2" event={"ID":"11828798-5abc-48cf-9f2e-ae69d2c7ffd5","Type":"ContainerDied","Data":"744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4"} Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.161469 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22rv2" event={"ID":"11828798-5abc-48cf-9f2e-ae69d2c7ffd5","Type":"ContainerStarted","Data":"e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7"} Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.163271 4846 generic.go:334] "Generic (PLEG): container finished" podID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerID="eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad" exitCode=0 Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.163313 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc8xt" event={"ID":"22bb6443-0edd-4312-b15f-e36efb8113d4","Type":"ContainerDied","Data":"eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad"} Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.191011 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-22rv2" podStartSLOduration=2.713337941 podStartE2EDuration="4.190989946s" podCreationTimestamp="2026-02-02 12:40:39 +0000 UTC" firstStartedPulling="2026-02-02 12:40:41.142346941 +0000 UTC m=+1872.370933804" lastFinishedPulling="2026-02-02 12:40:42.619998936 +0000 UTC m=+1873.848585809" observedRunningTime="2026-02-02 12:40:43.185212754 +0000 UTC m=+1874.413799627" watchObservedRunningTime="2026-02-02 12:40:43.190989946 +0000 UTC m=+1874.419576809" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.338396 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-knjjl"] Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.340155 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.350407 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-knjjl"] Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.434527 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-utilities\") pod \"certified-operators-knjjl\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.434795 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-catalog-content\") pod \"certified-operators-knjjl\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.434844 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v65h\" (UniqueName: \"kubernetes.io/projected/8bbcca6e-6e7c-4175-b91b-2971611ec8db-kube-api-access-6v65h\") pod \"certified-operators-knjjl\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.535960 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-catalog-content\") pod \"certified-operators-knjjl\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.536014 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v65h\" (UniqueName: \"kubernetes.io/projected/8bbcca6e-6e7c-4175-b91b-2971611ec8db-kube-api-access-6v65h\") pod \"certified-operators-knjjl\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.536074 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-utilities\") pod \"certified-operators-knjjl\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.536470 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-catalog-content\") pod \"certified-operators-knjjl\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.536514 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-utilities\") pod \"certified-operators-knjjl\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.557475 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v65h\" (UniqueName: \"kubernetes.io/projected/8bbcca6e-6e7c-4175-b91b-2971611ec8db-kube-api-access-6v65h\") pod \"certified-operators-knjjl\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:43 crc kubenswrapper[4846]: I0202 12:40:43.657077 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:44 crc kubenswrapper[4846]: I0202 12:40:44.152454 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-knjjl"] Feb 02 12:40:44 crc kubenswrapper[4846]: W0202 12:40:44.157029 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bbcca6e_6e7c_4175_b91b_2971611ec8db.slice/crio-db4562659b0161d6d5bd9930200d8a5be4ab444d7f3041463fce2bd6978e4710 WatchSource:0}: Error finding container db4562659b0161d6d5bd9930200d8a5be4ab444d7f3041463fce2bd6978e4710: Status 404 returned error can't find the container with id db4562659b0161d6d5bd9930200d8a5be4ab444d7f3041463fce2bd6978e4710 Feb 02 12:40:44 crc kubenswrapper[4846]: I0202 12:40:44.172906 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc8xt" event={"ID":"22bb6443-0edd-4312-b15f-e36efb8113d4","Type":"ContainerStarted","Data":"63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697"} Feb 02 12:40:44 crc kubenswrapper[4846]: I0202 12:40:44.174326 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-knjjl" event={"ID":"8bbcca6e-6e7c-4175-b91b-2971611ec8db","Type":"ContainerStarted","Data":"db4562659b0161d6d5bd9930200d8a5be4ab444d7f3041463fce2bd6978e4710"} Feb 02 12:40:44 crc kubenswrapper[4846]: I0202 12:40:44.207473 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hc8xt" podStartSLOduration=2.763809283 podStartE2EDuration="5.207449374s" podCreationTimestamp="2026-02-02 12:40:39 +0000 UTC" firstStartedPulling="2026-02-02 12:40:41.143194712 +0000 UTC m=+1872.371781565" lastFinishedPulling="2026-02-02 12:40:43.586834793 +0000 UTC m=+1874.815421656" observedRunningTime="2026-02-02 12:40:44.203941207 +0000 UTC m=+1875.432528080" watchObservedRunningTime="2026-02-02 12:40:44.207449374 +0000 UTC m=+1875.436036237" Feb 02 12:40:45 crc kubenswrapper[4846]: I0202 12:40:45.186612 4846 generic.go:334] "Generic (PLEG): container finished" podID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerID="ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555" exitCode=0 Feb 02 12:40:45 crc kubenswrapper[4846]: I0202 12:40:45.186718 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-knjjl" event={"ID":"8bbcca6e-6e7c-4175-b91b-2971611ec8db","Type":"ContainerDied","Data":"ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555"} Feb 02 12:40:46 crc kubenswrapper[4846]: I0202 12:40:46.195551 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-knjjl" event={"ID":"8bbcca6e-6e7c-4175-b91b-2971611ec8db","Type":"ContainerStarted","Data":"3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08"} Feb 02 12:40:47 crc kubenswrapper[4846]: I0202 12:40:47.208932 4846 generic.go:334] "Generic (PLEG): container finished" podID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerID="3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08" exitCode=0 Feb 02 12:40:47 crc kubenswrapper[4846]: I0202 12:40:47.209086 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-knjjl" event={"ID":"8bbcca6e-6e7c-4175-b91b-2971611ec8db","Type":"ContainerDied","Data":"3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08"} Feb 02 12:40:48 crc kubenswrapper[4846]: I0202 12:40:48.222106 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-knjjl" event={"ID":"8bbcca6e-6e7c-4175-b91b-2971611ec8db","Type":"ContainerStarted","Data":"da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34"} Feb 02 12:40:48 crc kubenswrapper[4846]: I0202 12:40:48.246045 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-knjjl" podStartSLOduration=2.751142557 podStartE2EDuration="5.246023809s" podCreationTimestamp="2026-02-02 12:40:43 +0000 UTC" firstStartedPulling="2026-02-02 12:40:45.190256794 +0000 UTC m=+1876.418843667" lastFinishedPulling="2026-02-02 12:40:47.685138056 +0000 UTC m=+1878.913724919" observedRunningTime="2026-02-02 12:40:48.240477916 +0000 UTC m=+1879.469064819" watchObservedRunningTime="2026-02-02 12:40:48.246023809 +0000 UTC m=+1879.474610673" Feb 02 12:40:49 crc kubenswrapper[4846]: I0202 12:40:49.688032 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:49 crc kubenswrapper[4846]: I0202 12:40:49.688421 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:49 crc kubenswrapper[4846]: I0202 12:40:49.740722 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:49 crc kubenswrapper[4846]: I0202 12:40:49.900680 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:49 crc kubenswrapper[4846]: I0202 12:40:49.900786 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:49 crc kubenswrapper[4846]: I0202 12:40:49.946060 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:50 crc kubenswrapper[4846]: I0202 12:40:50.281194 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:50 crc kubenswrapper[4846]: I0202 12:40:50.285238 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:51 crc kubenswrapper[4846]: I0202 12:40:51.733770 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-22rv2"] Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.256831 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-22rv2" podUID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerName="registry-server" containerID="cri-o://e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7" gracePeriod=2 Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.331315 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hc8xt"] Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.332116 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hc8xt" podUID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerName="registry-server" containerID="cri-o://63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697" gracePeriod=2 Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.655227 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.662701 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-utilities\") pod \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.662753 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-catalog-content\") pod \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.662776 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8sxd\" (UniqueName: \"kubernetes.io/projected/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-kube-api-access-d8sxd\") pod \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\" (UID: \"11828798-5abc-48cf-9f2e-ae69d2c7ffd5\") " Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.664018 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-utilities" (OuterVolumeSpecName: "utilities") pod "11828798-5abc-48cf-9f2e-ae69d2c7ffd5" (UID: "11828798-5abc-48cf-9f2e-ae69d2c7ffd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.673231 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-kube-api-access-d8sxd" (OuterVolumeSpecName: "kube-api-access-d8sxd") pod "11828798-5abc-48cf-9f2e-ae69d2c7ffd5" (UID: "11828798-5abc-48cf-9f2e-ae69d2c7ffd5"). InnerVolumeSpecName "kube-api-access-d8sxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.691162 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11828798-5abc-48cf-9f2e-ae69d2c7ffd5" (UID: "11828798-5abc-48cf-9f2e-ae69d2c7ffd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.706560 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.763925 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-catalog-content\") pod \"22bb6443-0edd-4312-b15f-e36efb8113d4\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.763971 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-utilities\") pod \"22bb6443-0edd-4312-b15f-e36efb8113d4\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.764024 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdgvk\" (UniqueName: \"kubernetes.io/projected/22bb6443-0edd-4312-b15f-e36efb8113d4-kube-api-access-xdgvk\") pod \"22bb6443-0edd-4312-b15f-e36efb8113d4\" (UID: \"22bb6443-0edd-4312-b15f-e36efb8113d4\") " Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.764226 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.764242 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.764252 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8sxd\" (UniqueName: \"kubernetes.io/projected/11828798-5abc-48cf-9f2e-ae69d2c7ffd5-kube-api-access-d8sxd\") on node \"crc\" DevicePath \"\"" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.764950 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-utilities" (OuterVolumeSpecName: "utilities") pod "22bb6443-0edd-4312-b15f-e36efb8113d4" (UID: "22bb6443-0edd-4312-b15f-e36efb8113d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.766556 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22bb6443-0edd-4312-b15f-e36efb8113d4-kube-api-access-xdgvk" (OuterVolumeSpecName: "kube-api-access-xdgvk") pod "22bb6443-0edd-4312-b15f-e36efb8113d4" (UID: "22bb6443-0edd-4312-b15f-e36efb8113d4"). InnerVolumeSpecName "kube-api-access-xdgvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.865649 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.865699 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdgvk\" (UniqueName: \"kubernetes.io/projected/22bb6443-0edd-4312-b15f-e36efb8113d4-kube-api-access-xdgvk\") on node \"crc\" DevicePath \"\"" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.882980 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22bb6443-0edd-4312-b15f-e36efb8113d4" (UID: "22bb6443-0edd-4312-b15f-e36efb8113d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:40:52 crc kubenswrapper[4846]: I0202 12:40:52.966887 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22bb6443-0edd-4312-b15f-e36efb8113d4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.324315 4846 generic.go:334] "Generic (PLEG): container finished" podID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerID="e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7" exitCode=0 Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.324408 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22rv2" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.324424 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22rv2" event={"ID":"11828798-5abc-48cf-9f2e-ae69d2c7ffd5","Type":"ContainerDied","Data":"e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7"} Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.324473 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22rv2" event={"ID":"11828798-5abc-48cf-9f2e-ae69d2c7ffd5","Type":"ContainerDied","Data":"ddd77f8cd62457a9954e14e966bd7ff2592ddc73f5d3524ebe7139b7ac95c829"} Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.324575 4846 scope.go:117] "RemoveContainer" containerID="e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.327780 4846 generic.go:334] "Generic (PLEG): container finished" podID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerID="63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697" exitCode=0 Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.327831 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc8xt" event={"ID":"22bb6443-0edd-4312-b15f-e36efb8113d4","Type":"ContainerDied","Data":"63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697"} Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.327851 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hc8xt" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.327872 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc8xt" event={"ID":"22bb6443-0edd-4312-b15f-e36efb8113d4","Type":"ContainerDied","Data":"0170f50aed775674a8b58b10adc301571eff0729dc58355cb6997d95b8b784be"} Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.357350 4846 scope.go:117] "RemoveContainer" containerID="744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.374791 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-22rv2"] Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.387960 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-22rv2"] Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.410612 4846 scope.go:117] "RemoveContainer" containerID="db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.411927 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hc8xt"] Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.417865 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hc8xt"] Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.431171 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" path="/var/lib/kubelet/pods/11828798-5abc-48cf-9f2e-ae69d2c7ffd5/volumes" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.431910 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22bb6443-0edd-4312-b15f-e36efb8113d4" path="/var/lib/kubelet/pods/22bb6443-0edd-4312-b15f-e36efb8113d4/volumes" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.438831 4846 scope.go:117] "RemoveContainer" containerID="e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7" Feb 02 12:40:53 crc kubenswrapper[4846]: E0202 12:40:53.439222 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7\": container with ID starting with e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7 not found: ID does not exist" containerID="e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.439312 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7"} err="failed to get container status \"e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7\": rpc error: code = NotFound desc = could not find container \"e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7\": container with ID starting with e2c81b3be6798258d1b28c4374ac0d43a55f5de2f3edf9e648b406849c4fd8d7 not found: ID does not exist" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.439392 4846 scope.go:117] "RemoveContainer" containerID="744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4" Feb 02 12:40:53 crc kubenswrapper[4846]: E0202 12:40:53.439744 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4\": container with ID starting with 744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4 not found: ID does not exist" containerID="744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.439816 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4"} err="failed to get container status \"744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4\": rpc error: code = NotFound desc = could not find container \"744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4\": container with ID starting with 744346056b25e9711363dab3d58c84294c675e9bceb8b024ff02436d04bf99d4 not found: ID does not exist" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.439887 4846 scope.go:117] "RemoveContainer" containerID="db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e" Feb 02 12:40:53 crc kubenswrapper[4846]: E0202 12:40:53.440314 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e\": container with ID starting with db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e not found: ID does not exist" containerID="db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.440427 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e"} err="failed to get container status \"db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e\": rpc error: code = NotFound desc = could not find container \"db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e\": container with ID starting with db8b0833f447f35c58d9a43968cc1de4f987ab48663e6d867bf78ebd262ab28e not found: ID does not exist" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.440516 4846 scope.go:117] "RemoveContainer" containerID="63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.454495 4846 scope.go:117] "RemoveContainer" containerID="eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.468730 4846 scope.go:117] "RemoveContainer" containerID="deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.491155 4846 scope.go:117] "RemoveContainer" containerID="63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697" Feb 02 12:40:53 crc kubenswrapper[4846]: E0202 12:40:53.491598 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697\": container with ID starting with 63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697 not found: ID does not exist" containerID="63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.491647 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697"} err="failed to get container status \"63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697\": rpc error: code = NotFound desc = could not find container \"63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697\": container with ID starting with 63810edbdec75c853c58c8022b6a7a707bf526e293bcdba05f5d853b23ac0697 not found: ID does not exist" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.491676 4846 scope.go:117] "RemoveContainer" containerID="eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad" Feb 02 12:40:53 crc kubenswrapper[4846]: E0202 12:40:53.491967 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad\": container with ID starting with eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad not found: ID does not exist" containerID="eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.492001 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad"} err="failed to get container status \"eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad\": rpc error: code = NotFound desc = could not find container \"eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad\": container with ID starting with eb5deb2874d50ff436b52c12b8bbe3783ba29c87b3188c89e441400e126fa7ad not found: ID does not exist" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.492020 4846 scope.go:117] "RemoveContainer" containerID="deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100" Feb 02 12:40:53 crc kubenswrapper[4846]: E0202 12:40:53.492247 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100\": container with ID starting with deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100 not found: ID does not exist" containerID="deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.492265 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100"} err="failed to get container status \"deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100\": rpc error: code = NotFound desc = could not find container \"deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100\": container with ID starting with deea8fc22f93986db187f05f0ae512b676ec8bb79477bdcdac81a0a578d48100 not found: ID does not exist" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.657185 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.657234 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:53 crc kubenswrapper[4846]: I0202 12:40:53.730271 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:54 crc kubenswrapper[4846]: I0202 12:40:54.411357 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:56 crc kubenswrapper[4846]: I0202 12:40:56.737109 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-knjjl"] Feb 02 12:40:56 crc kubenswrapper[4846]: I0202 12:40:56.737458 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-knjjl" podUID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerName="registry-server" containerID="cri-o://da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34" gracePeriod=2 Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.187998 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.230936 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-catalog-content\") pod \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.231020 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-utilities\") pod \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.231056 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v65h\" (UniqueName: \"kubernetes.io/projected/8bbcca6e-6e7c-4175-b91b-2971611ec8db-kube-api-access-6v65h\") pod \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\" (UID: \"8bbcca6e-6e7c-4175-b91b-2971611ec8db\") " Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.232268 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-utilities" (OuterVolumeSpecName: "utilities") pod "8bbcca6e-6e7c-4175-b91b-2971611ec8db" (UID: "8bbcca6e-6e7c-4175-b91b-2971611ec8db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.237813 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bbcca6e-6e7c-4175-b91b-2971611ec8db-kube-api-access-6v65h" (OuterVolumeSpecName: "kube-api-access-6v65h") pod "8bbcca6e-6e7c-4175-b91b-2971611ec8db" (UID: "8bbcca6e-6e7c-4175-b91b-2971611ec8db"). InnerVolumeSpecName "kube-api-access-6v65h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.291775 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bbcca6e-6e7c-4175-b91b-2971611ec8db" (UID: "8bbcca6e-6e7c-4175-b91b-2971611ec8db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.332378 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.332413 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bbcca6e-6e7c-4175-b91b-2971611ec8db-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.332426 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v65h\" (UniqueName: \"kubernetes.io/projected/8bbcca6e-6e7c-4175-b91b-2971611ec8db-kube-api-access-6v65h\") on node \"crc\" DevicePath \"\"" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.366893 4846 generic.go:334] "Generic (PLEG): container finished" podID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerID="da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34" exitCode=0 Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.366959 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-knjjl" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.366949 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-knjjl" event={"ID":"8bbcca6e-6e7c-4175-b91b-2971611ec8db","Type":"ContainerDied","Data":"da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34"} Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.367428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-knjjl" event={"ID":"8bbcca6e-6e7c-4175-b91b-2971611ec8db","Type":"ContainerDied","Data":"db4562659b0161d6d5bd9930200d8a5be4ab444d7f3041463fce2bd6978e4710"} Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.367487 4846 scope.go:117] "RemoveContainer" containerID="da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.392175 4846 scope.go:117] "RemoveContainer" containerID="3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.407360 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-knjjl"] Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.412513 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-knjjl"] Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.431441 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" path="/var/lib/kubelet/pods/8bbcca6e-6e7c-4175-b91b-2971611ec8db/volumes" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.440532 4846 scope.go:117] "RemoveContainer" containerID="ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.457306 4846 scope.go:117] "RemoveContainer" containerID="da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34" Feb 02 12:40:57 crc kubenswrapper[4846]: E0202 12:40:57.457825 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34\": container with ID starting with da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34 not found: ID does not exist" containerID="da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.457881 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34"} err="failed to get container status \"da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34\": rpc error: code = NotFound desc = could not find container \"da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34\": container with ID starting with da2fd00772b2fe9b9bc920d9f95ba6ecc753b5b5243873ce5b40946d3a253b34 not found: ID does not exist" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.457914 4846 scope.go:117] "RemoveContainer" containerID="3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08" Feb 02 12:40:57 crc kubenswrapper[4846]: E0202 12:40:57.458506 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08\": container with ID starting with 3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08 not found: ID does not exist" containerID="3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.458555 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08"} err="failed to get container status \"3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08\": rpc error: code = NotFound desc = could not find container \"3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08\": container with ID starting with 3a7e3c626ba74610f2d986ff04da6e8f803b31ad3e309a40413f79a667330a08 not found: ID does not exist" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.458592 4846 scope.go:117] "RemoveContainer" containerID="ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555" Feb 02 12:40:57 crc kubenswrapper[4846]: E0202 12:40:57.459238 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555\": container with ID starting with ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555 not found: ID does not exist" containerID="ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555" Feb 02 12:40:57 crc kubenswrapper[4846]: I0202 12:40:57.459339 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555"} err="failed to get container status \"ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555\": rpc error: code = NotFound desc = could not find container \"ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555\": container with ID starting with ebcb560f3720e1bad859c2e79dcda47168148f6f70e26e42891a3645c441a555 not found: ID does not exist" Feb 02 12:42:00 crc kubenswrapper[4846]: I0202 12:42:00.479170 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:42:00 crc kubenswrapper[4846]: I0202 12:42:00.479971 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:42:30 crc kubenswrapper[4846]: I0202 12:42:30.479396 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:42:30 crc kubenswrapper[4846]: I0202 12:42:30.480077 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:43:00 crc kubenswrapper[4846]: I0202 12:43:00.479548 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:43:00 crc kubenswrapper[4846]: I0202 12:43:00.480147 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:43:00 crc kubenswrapper[4846]: I0202 12:43:00.480230 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:43:00 crc kubenswrapper[4846]: I0202 12:43:00.481356 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"410ee2677ceebfefaa85449243876d89f0c44fbef81405ad99a3ceddb58f3e7c"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:43:00 crc kubenswrapper[4846]: I0202 12:43:00.481483 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://410ee2677ceebfefaa85449243876d89f0c44fbef81405ad99a3ceddb58f3e7c" gracePeriod=600 Feb 02 12:43:01 crc kubenswrapper[4846]: I0202 12:43:01.358345 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="410ee2677ceebfefaa85449243876d89f0c44fbef81405ad99a3ceddb58f3e7c" exitCode=0 Feb 02 12:43:01 crc kubenswrapper[4846]: I0202 12:43:01.358396 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"410ee2677ceebfefaa85449243876d89f0c44fbef81405ad99a3ceddb58f3e7c"} Feb 02 12:43:01 crc kubenswrapper[4846]: I0202 12:43:01.358926 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9"} Feb 02 12:43:01 crc kubenswrapper[4846]: I0202 12:43:01.358949 4846 scope.go:117] "RemoveContainer" containerID="0928dbda829dd2200cc9eb62f2cd65301e7b2ea5abee4d66341d59c4394cee8b" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.612044 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cf9h9"] Feb 02 12:44:20 crc kubenswrapper[4846]: E0202 12:44:20.613592 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerName="extract-content" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.613614 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerName="extract-content" Feb 02 12:44:20 crc kubenswrapper[4846]: E0202 12:44:20.613661 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerName="extract-utilities" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.613669 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerName="extract-utilities" Feb 02 12:44:20 crc kubenswrapper[4846]: E0202 12:44:20.613690 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerName="extract-content" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.613706 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerName="extract-content" Feb 02 12:44:20 crc kubenswrapper[4846]: E0202 12:44:20.613728 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerName="extract-utilities" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.613739 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerName="extract-utilities" Feb 02 12:44:20 crc kubenswrapper[4846]: E0202 12:44:20.613753 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerName="extract-content" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.613761 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerName="extract-content" Feb 02 12:44:20 crc kubenswrapper[4846]: E0202 12:44:20.613774 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerName="registry-server" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.613783 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerName="registry-server" Feb 02 12:44:20 crc kubenswrapper[4846]: E0202 12:44:20.613800 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerName="registry-server" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.613809 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerName="registry-server" Feb 02 12:44:20 crc kubenswrapper[4846]: E0202 12:44:20.613823 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerName="registry-server" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.613829 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerName="registry-server" Feb 02 12:44:20 crc kubenswrapper[4846]: E0202 12:44:20.613836 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerName="extract-utilities" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.613843 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerName="extract-utilities" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.614041 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bbcca6e-6e7c-4175-b91b-2971611ec8db" containerName="registry-server" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.614066 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="22bb6443-0edd-4312-b15f-e36efb8113d4" containerName="registry-server" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.614077 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="11828798-5abc-48cf-9f2e-ae69d2c7ffd5" containerName="registry-server" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.615420 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.623577 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cf9h9"] Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.703348 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-catalog-content\") pod \"redhat-operators-cf9h9\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.703961 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-utilities\") pod \"redhat-operators-cf9h9\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.704134 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69g29\" (UniqueName: \"kubernetes.io/projected/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-kube-api-access-69g29\") pod \"redhat-operators-cf9h9\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.805706 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-utilities\") pod \"redhat-operators-cf9h9\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.805791 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69g29\" (UniqueName: \"kubernetes.io/projected/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-kube-api-access-69g29\") pod \"redhat-operators-cf9h9\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.805841 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-catalog-content\") pod \"redhat-operators-cf9h9\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.806451 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-catalog-content\") pod \"redhat-operators-cf9h9\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.806573 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-utilities\") pod \"redhat-operators-cf9h9\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.825531 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69g29\" (UniqueName: \"kubernetes.io/projected/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-kube-api-access-69g29\") pod \"redhat-operators-cf9h9\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:20 crc kubenswrapper[4846]: I0202 12:44:20.933290 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:21 crc kubenswrapper[4846]: I0202 12:44:21.392915 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cf9h9"] Feb 02 12:44:21 crc kubenswrapper[4846]: I0202 12:44:21.945228 4846 generic.go:334] "Generic (PLEG): container finished" podID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerID="32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8" exitCode=0 Feb 02 12:44:21 crc kubenswrapper[4846]: I0202 12:44:21.945271 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf9h9" event={"ID":"e6c95b81-5f7f-4244-bde1-33606fe2a3b2","Type":"ContainerDied","Data":"32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8"} Feb 02 12:44:21 crc kubenswrapper[4846]: I0202 12:44:21.945292 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf9h9" event={"ID":"e6c95b81-5f7f-4244-bde1-33606fe2a3b2","Type":"ContainerStarted","Data":"17d24ec040a36bc26974a99bc51760836c93ef098a11a27d4b6eb6174b0691b1"} Feb 02 12:44:23 crc kubenswrapper[4846]: I0202 12:44:23.965428 4846 generic.go:334] "Generic (PLEG): container finished" podID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerID="dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee" exitCode=0 Feb 02 12:44:23 crc kubenswrapper[4846]: I0202 12:44:23.965523 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf9h9" event={"ID":"e6c95b81-5f7f-4244-bde1-33606fe2a3b2","Type":"ContainerDied","Data":"dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee"} Feb 02 12:44:24 crc kubenswrapper[4846]: I0202 12:44:24.975435 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf9h9" event={"ID":"e6c95b81-5f7f-4244-bde1-33606fe2a3b2","Type":"ContainerStarted","Data":"9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f"} Feb 02 12:44:24 crc kubenswrapper[4846]: I0202 12:44:24.999580 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cf9h9" podStartSLOduration=2.581721218 podStartE2EDuration="4.999562505s" podCreationTimestamp="2026-02-02 12:44:20 +0000 UTC" firstStartedPulling="2026-02-02 12:44:21.946676786 +0000 UTC m=+2093.175263649" lastFinishedPulling="2026-02-02 12:44:24.364518073 +0000 UTC m=+2095.593104936" observedRunningTime="2026-02-02 12:44:24.997147506 +0000 UTC m=+2096.225734379" watchObservedRunningTime="2026-02-02 12:44:24.999562505 +0000 UTC m=+2096.228149368" Feb 02 12:44:30 crc kubenswrapper[4846]: I0202 12:44:30.934296 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:30 crc kubenswrapper[4846]: I0202 12:44:30.934841 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:30 crc kubenswrapper[4846]: I0202 12:44:30.996497 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:31 crc kubenswrapper[4846]: I0202 12:44:31.064820 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:31 crc kubenswrapper[4846]: I0202 12:44:31.241664 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cf9h9"] Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.026907 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cf9h9" podUID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerName="registry-server" containerID="cri-o://9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f" gracePeriod=2 Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.372296 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.485063 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69g29\" (UniqueName: \"kubernetes.io/projected/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-kube-api-access-69g29\") pod \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.485244 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-utilities\") pod \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.485327 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-catalog-content\") pod \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\" (UID: \"e6c95b81-5f7f-4244-bde1-33606fe2a3b2\") " Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.486427 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-utilities" (OuterVolumeSpecName: "utilities") pod "e6c95b81-5f7f-4244-bde1-33606fe2a3b2" (UID: "e6c95b81-5f7f-4244-bde1-33606fe2a3b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.491295 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-kube-api-access-69g29" (OuterVolumeSpecName: "kube-api-access-69g29") pod "e6c95b81-5f7f-4244-bde1-33606fe2a3b2" (UID: "e6c95b81-5f7f-4244-bde1-33606fe2a3b2"). InnerVolumeSpecName "kube-api-access-69g29". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.587180 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69g29\" (UniqueName: \"kubernetes.io/projected/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-kube-api-access-69g29\") on node \"crc\" DevicePath \"\"" Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.587215 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.628911 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6c95b81-5f7f-4244-bde1-33606fe2a3b2" (UID: "e6c95b81-5f7f-4244-bde1-33606fe2a3b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:44:33 crc kubenswrapper[4846]: I0202 12:44:33.689316 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6c95b81-5f7f-4244-bde1-33606fe2a3b2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.037159 4846 generic.go:334] "Generic (PLEG): container finished" podID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerID="9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f" exitCode=0 Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.037223 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf9h9" event={"ID":"e6c95b81-5f7f-4244-bde1-33606fe2a3b2","Type":"ContainerDied","Data":"9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f"} Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.037263 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf9h9" event={"ID":"e6c95b81-5f7f-4244-bde1-33606fe2a3b2","Type":"ContainerDied","Data":"17d24ec040a36bc26974a99bc51760836c93ef098a11a27d4b6eb6174b0691b1"} Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.037292 4846 scope.go:117] "RemoveContainer" containerID="9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f" Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.037333 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf9h9" Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.085066 4846 scope.go:117] "RemoveContainer" containerID="dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee" Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.093212 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cf9h9"] Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.099222 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cf9h9"] Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.116509 4846 scope.go:117] "RemoveContainer" containerID="32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8" Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.134380 4846 scope.go:117] "RemoveContainer" containerID="9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f" Feb 02 12:44:34 crc kubenswrapper[4846]: E0202 12:44:34.134866 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f\": container with ID starting with 9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f not found: ID does not exist" containerID="9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f" Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.134897 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f"} err="failed to get container status \"9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f\": rpc error: code = NotFound desc = could not find container \"9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f\": container with ID starting with 9ddcdb0f0375c330d3373448000e077d46c8b6ae6ed831461854e8ef0ccab02f not found: ID does not exist" Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.134919 4846 scope.go:117] "RemoveContainer" containerID="dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee" Feb 02 12:44:34 crc kubenswrapper[4846]: E0202 12:44:34.135196 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee\": container with ID starting with dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee not found: ID does not exist" containerID="dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee" Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.135215 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee"} err="failed to get container status \"dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee\": rpc error: code = NotFound desc = could not find container \"dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee\": container with ID starting with dbaf4d7d905153bd6ffc55351eec0e197092feaf1f88061e04c4728972c8f5ee not found: ID does not exist" Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.135226 4846 scope.go:117] "RemoveContainer" containerID="32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8" Feb 02 12:44:34 crc kubenswrapper[4846]: E0202 12:44:34.135487 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8\": container with ID starting with 32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8 not found: ID does not exist" containerID="32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8" Feb 02 12:44:34 crc kubenswrapper[4846]: I0202 12:44:34.135509 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8"} err="failed to get container status \"32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8\": rpc error: code = NotFound desc = could not find container \"32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8\": container with ID starting with 32275b95963a43a3f9848857b2ab9ce6cb4f7e980061908bc82417d28c0124d8 not found: ID does not exist" Feb 02 12:44:35 crc kubenswrapper[4846]: I0202 12:44:35.439189 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" path="/var/lib/kubelet/pods/e6c95b81-5f7f-4244-bde1-33606fe2a3b2/volumes" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.154833 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn"] Feb 02 12:45:00 crc kubenswrapper[4846]: E0202 12:45:00.155753 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerName="registry-server" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.155770 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerName="registry-server" Feb 02 12:45:00 crc kubenswrapper[4846]: E0202 12:45:00.155786 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerName="extract-content" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.155794 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerName="extract-content" Feb 02 12:45:00 crc kubenswrapper[4846]: E0202 12:45:00.155822 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerName="extract-utilities" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.155831 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerName="extract-utilities" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.155981 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6c95b81-5f7f-4244-bde1-33606fe2a3b2" containerName="registry-server" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.156531 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.158961 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.159787 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.168834 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn"] Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.281553 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x492c\" (UniqueName: \"kubernetes.io/projected/0f814e99-ae56-4c39-a852-5efa6e466b81-kube-api-access-x492c\") pod \"collect-profiles-29500605-swkkn\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.281917 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f814e99-ae56-4c39-a852-5efa6e466b81-config-volume\") pod \"collect-profiles-29500605-swkkn\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.282037 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f814e99-ae56-4c39-a852-5efa6e466b81-secret-volume\") pod \"collect-profiles-29500605-swkkn\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.383296 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x492c\" (UniqueName: \"kubernetes.io/projected/0f814e99-ae56-4c39-a852-5efa6e466b81-kube-api-access-x492c\") pod \"collect-profiles-29500605-swkkn\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.383380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f814e99-ae56-4c39-a852-5efa6e466b81-config-volume\") pod \"collect-profiles-29500605-swkkn\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.383409 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f814e99-ae56-4c39-a852-5efa6e466b81-secret-volume\") pod \"collect-profiles-29500605-swkkn\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.384683 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f814e99-ae56-4c39-a852-5efa6e466b81-config-volume\") pod \"collect-profiles-29500605-swkkn\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.394172 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f814e99-ae56-4c39-a852-5efa6e466b81-secret-volume\") pod \"collect-profiles-29500605-swkkn\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.400312 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x492c\" (UniqueName: \"kubernetes.io/projected/0f814e99-ae56-4c39-a852-5efa6e466b81-kube-api-access-x492c\") pod \"collect-profiles-29500605-swkkn\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.478784 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.479144 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.484243 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:00 crc kubenswrapper[4846]: I0202 12:45:00.889586 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn"] Feb 02 12:45:01 crc kubenswrapper[4846]: I0202 12:45:01.279860 4846 generic.go:334] "Generic (PLEG): container finished" podID="0f814e99-ae56-4c39-a852-5efa6e466b81" containerID="8b6165518b94d6d62387f0620967b1c1b5c7f81696b97b4f2f79733d4fce3ebb" exitCode=0 Feb 02 12:45:01 crc kubenswrapper[4846]: I0202 12:45:01.280034 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" event={"ID":"0f814e99-ae56-4c39-a852-5efa6e466b81","Type":"ContainerDied","Data":"8b6165518b94d6d62387f0620967b1c1b5c7f81696b97b4f2f79733d4fce3ebb"} Feb 02 12:45:01 crc kubenswrapper[4846]: I0202 12:45:01.280083 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" event={"ID":"0f814e99-ae56-4c39-a852-5efa6e466b81","Type":"ContainerStarted","Data":"51abc737747b9bf7ba71dd29ac61b3ebec4619d70e24e05558ab44ab778720d9"} Feb 02 12:45:02 crc kubenswrapper[4846]: I0202 12:45:02.588406 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:02 crc kubenswrapper[4846]: I0202 12:45:02.715326 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f814e99-ae56-4c39-a852-5efa6e466b81-config-volume\") pod \"0f814e99-ae56-4c39-a852-5efa6e466b81\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " Feb 02 12:45:02 crc kubenswrapper[4846]: I0202 12:45:02.715386 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x492c\" (UniqueName: \"kubernetes.io/projected/0f814e99-ae56-4c39-a852-5efa6e466b81-kube-api-access-x492c\") pod \"0f814e99-ae56-4c39-a852-5efa6e466b81\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " Feb 02 12:45:02 crc kubenswrapper[4846]: I0202 12:45:02.715544 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f814e99-ae56-4c39-a852-5efa6e466b81-secret-volume\") pod \"0f814e99-ae56-4c39-a852-5efa6e466b81\" (UID: \"0f814e99-ae56-4c39-a852-5efa6e466b81\") " Feb 02 12:45:02 crc kubenswrapper[4846]: I0202 12:45:02.715927 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f814e99-ae56-4c39-a852-5efa6e466b81-config-volume" (OuterVolumeSpecName: "config-volume") pod "0f814e99-ae56-4c39-a852-5efa6e466b81" (UID: "0f814e99-ae56-4c39-a852-5efa6e466b81"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 12:45:02 crc kubenswrapper[4846]: I0202 12:45:02.720581 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f814e99-ae56-4c39-a852-5efa6e466b81-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0f814e99-ae56-4c39-a852-5efa6e466b81" (UID: "0f814e99-ae56-4c39-a852-5efa6e466b81"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 12:45:02 crc kubenswrapper[4846]: I0202 12:45:02.720881 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f814e99-ae56-4c39-a852-5efa6e466b81-kube-api-access-x492c" (OuterVolumeSpecName: "kube-api-access-x492c") pod "0f814e99-ae56-4c39-a852-5efa6e466b81" (UID: "0f814e99-ae56-4c39-a852-5efa6e466b81"). InnerVolumeSpecName "kube-api-access-x492c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:45:02 crc kubenswrapper[4846]: I0202 12:45:02.816893 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0f814e99-ae56-4c39-a852-5efa6e466b81-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 12:45:02 crc kubenswrapper[4846]: I0202 12:45:02.816947 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0f814e99-ae56-4c39-a852-5efa6e466b81-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 12:45:02 crc kubenswrapper[4846]: I0202 12:45:02.816969 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x492c\" (UniqueName: \"kubernetes.io/projected/0f814e99-ae56-4c39-a852-5efa6e466b81-kube-api-access-x492c\") on node \"crc\" DevicePath \"\"" Feb 02 12:45:03 crc kubenswrapper[4846]: I0202 12:45:03.294292 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" event={"ID":"0f814e99-ae56-4c39-a852-5efa6e466b81","Type":"ContainerDied","Data":"51abc737747b9bf7ba71dd29ac61b3ebec4619d70e24e05558ab44ab778720d9"} Feb 02 12:45:03 crc kubenswrapper[4846]: I0202 12:45:03.294614 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51abc737747b9bf7ba71dd29ac61b3ebec4619d70e24e05558ab44ab778720d9" Feb 02 12:45:03 crc kubenswrapper[4846]: I0202 12:45:03.294352 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn" Feb 02 12:45:03 crc kubenswrapper[4846]: I0202 12:45:03.673862 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk"] Feb 02 12:45:03 crc kubenswrapper[4846]: I0202 12:45:03.680084 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500560-wb4lk"] Feb 02 12:45:05 crc kubenswrapper[4846]: I0202 12:45:05.439328 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d559728-5463-4a87-b4e7-b767dfe50bd7" path="/var/lib/kubelet/pods/7d559728-5463-4a87-b4e7-b767dfe50bd7/volumes" Feb 02 12:45:30 crc kubenswrapper[4846]: I0202 12:45:30.479187 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:45:30 crc kubenswrapper[4846]: I0202 12:45:30.479756 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:45:36 crc kubenswrapper[4846]: I0202 12:45:36.415304 4846 scope.go:117] "RemoveContainer" containerID="b77000692f2fa0089b2dc53b65e836c076ed0cf6d08245baa4599490ad60038d" Feb 02 12:46:00 crc kubenswrapper[4846]: I0202 12:46:00.479253 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:46:00 crc kubenswrapper[4846]: I0202 12:46:00.479888 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:46:00 crc kubenswrapper[4846]: I0202 12:46:00.479942 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:46:00 crc kubenswrapper[4846]: I0202 12:46:00.480609 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:46:00 crc kubenswrapper[4846]: I0202 12:46:00.480681 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" gracePeriod=600 Feb 02 12:46:00 crc kubenswrapper[4846]: E0202 12:46:00.605114 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:46:00 crc kubenswrapper[4846]: I0202 12:46:00.724686 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" exitCode=0 Feb 02 12:46:00 crc kubenswrapper[4846]: I0202 12:46:00.724738 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9"} Feb 02 12:46:00 crc kubenswrapper[4846]: I0202 12:46:00.724783 4846 scope.go:117] "RemoveContainer" containerID="410ee2677ceebfefaa85449243876d89f0c44fbef81405ad99a3ceddb58f3e7c" Feb 02 12:46:00 crc kubenswrapper[4846]: I0202 12:46:00.725440 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:46:00 crc kubenswrapper[4846]: E0202 12:46:00.725705 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:46:11 crc kubenswrapper[4846]: I0202 12:46:11.423880 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:46:11 crc kubenswrapper[4846]: E0202 12:46:11.424971 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:46:23 crc kubenswrapper[4846]: I0202 12:46:23.423584 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:46:23 crc kubenswrapper[4846]: E0202 12:46:23.424495 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:46:37 crc kubenswrapper[4846]: I0202 12:46:37.423295 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:46:37 crc kubenswrapper[4846]: E0202 12:46:37.424061 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:46:52 crc kubenswrapper[4846]: I0202 12:46:52.423705 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:46:52 crc kubenswrapper[4846]: E0202 12:46:52.424538 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:47:05 crc kubenswrapper[4846]: I0202 12:47:05.424406 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:47:05 crc kubenswrapper[4846]: E0202 12:47:05.425848 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:47:17 crc kubenswrapper[4846]: I0202 12:47:17.424063 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:47:17 crc kubenswrapper[4846]: E0202 12:47:17.425265 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:47:32 crc kubenswrapper[4846]: I0202 12:47:32.423808 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:47:32 crc kubenswrapper[4846]: E0202 12:47:32.424572 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:47:44 crc kubenswrapper[4846]: I0202 12:47:44.423790 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:47:44 crc kubenswrapper[4846]: E0202 12:47:44.424775 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:47:58 crc kubenswrapper[4846]: I0202 12:47:58.424153 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:47:58 crc kubenswrapper[4846]: E0202 12:47:58.424846 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:48:09 crc kubenswrapper[4846]: I0202 12:48:09.430057 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:48:09 crc kubenswrapper[4846]: E0202 12:48:09.430791 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:48:21 crc kubenswrapper[4846]: I0202 12:48:21.424347 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:48:21 crc kubenswrapper[4846]: E0202 12:48:21.425399 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:48:36 crc kubenswrapper[4846]: I0202 12:48:36.424675 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:48:36 crc kubenswrapper[4846]: E0202 12:48:36.425841 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:48:48 crc kubenswrapper[4846]: I0202 12:48:48.423274 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:48:48 crc kubenswrapper[4846]: E0202 12:48:48.424083 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:49:01 crc kubenswrapper[4846]: I0202 12:49:01.425100 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:49:01 crc kubenswrapper[4846]: E0202 12:49:01.426323 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:49:12 crc kubenswrapper[4846]: I0202 12:49:12.423772 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:49:12 crc kubenswrapper[4846]: E0202 12:49:12.424894 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:49:27 crc kubenswrapper[4846]: I0202 12:49:27.423753 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:49:27 crc kubenswrapper[4846]: E0202 12:49:27.424256 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:49:37 crc kubenswrapper[4846]: I0202 12:49:37.873214 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:49:37 crc kubenswrapper[4846]: E0202 12:49:37.874029 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:49:51 crc kubenswrapper[4846]: I0202 12:49:51.423697 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:49:51 crc kubenswrapper[4846]: E0202 12:49:51.425100 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:50:02 crc kubenswrapper[4846]: I0202 12:50:02.423016 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:50:02 crc kubenswrapper[4846]: E0202 12:50:02.423720 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:50:13 crc kubenswrapper[4846]: I0202 12:50:13.424687 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:50:13 crc kubenswrapper[4846]: E0202 12:50:13.425515 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:50:24 crc kubenswrapper[4846]: I0202 12:50:24.423385 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:50:24 crc kubenswrapper[4846]: E0202 12:50:24.425033 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:50:35 crc kubenswrapper[4846]: I0202 12:50:35.424095 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:50:35 crc kubenswrapper[4846]: E0202 12:50:35.425114 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:50:46 crc kubenswrapper[4846]: I0202 12:50:46.423553 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:50:46 crc kubenswrapper[4846]: E0202 12:50:46.424768 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.763618 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2js6p"] Feb 02 12:50:54 crc kubenswrapper[4846]: E0202 12:50:54.764513 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f814e99-ae56-4c39-a852-5efa6e466b81" containerName="collect-profiles" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.764532 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f814e99-ae56-4c39-a852-5efa6e466b81" containerName="collect-profiles" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.764749 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f814e99-ae56-4c39-a852-5efa6e466b81" containerName="collect-profiles" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.766221 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.782379 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2js6p"] Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.819439 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9cl4\" (UniqueName: \"kubernetes.io/projected/e7869f6a-afa5-47db-80c0-e4ae54e939bc-kube-api-access-l9cl4\") pod \"community-operators-2js6p\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.819544 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-catalog-content\") pod \"community-operators-2js6p\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.819576 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-utilities\") pod \"community-operators-2js6p\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.921320 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9cl4\" (UniqueName: \"kubernetes.io/projected/e7869f6a-afa5-47db-80c0-e4ae54e939bc-kube-api-access-l9cl4\") pod \"community-operators-2js6p\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.921409 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-catalog-content\") pod \"community-operators-2js6p\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.921435 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-utilities\") pod \"community-operators-2js6p\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.921927 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-utilities\") pod \"community-operators-2js6p\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.922071 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-catalog-content\") pod \"community-operators-2js6p\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:54 crc kubenswrapper[4846]: I0202 12:50:54.949774 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9cl4\" (UniqueName: \"kubernetes.io/projected/e7869f6a-afa5-47db-80c0-e4ae54e939bc-kube-api-access-l9cl4\") pod \"community-operators-2js6p\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:55 crc kubenswrapper[4846]: I0202 12:50:55.092041 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:50:55 crc kubenswrapper[4846]: I0202 12:50:55.432104 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2js6p"] Feb 02 12:50:55 crc kubenswrapper[4846]: I0202 12:50:55.460607 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2js6p" event={"ID":"e7869f6a-afa5-47db-80c0-e4ae54e939bc","Type":"ContainerStarted","Data":"941acc7d2c1ecb2c61ff60ddf95cf1549b197f99078988c2e99af113e6fef435"} Feb 02 12:50:56 crc kubenswrapper[4846]: I0202 12:50:56.472514 4846 generic.go:334] "Generic (PLEG): container finished" podID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerID="7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec" exitCode=0 Feb 02 12:50:56 crc kubenswrapper[4846]: I0202 12:50:56.472693 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2js6p" event={"ID":"e7869f6a-afa5-47db-80c0-e4ae54e939bc","Type":"ContainerDied","Data":"7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec"} Feb 02 12:50:56 crc kubenswrapper[4846]: I0202 12:50:56.475045 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 12:50:58 crc kubenswrapper[4846]: I0202 12:50:58.491340 4846 generic.go:334] "Generic (PLEG): container finished" podID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerID="faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a" exitCode=0 Feb 02 12:50:58 crc kubenswrapper[4846]: I0202 12:50:58.491425 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2js6p" event={"ID":"e7869f6a-afa5-47db-80c0-e4ae54e939bc","Type":"ContainerDied","Data":"faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a"} Feb 02 12:50:59 crc kubenswrapper[4846]: I0202 12:50:59.426940 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:50:59 crc kubenswrapper[4846]: E0202 12:50:59.427262 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:50:59 crc kubenswrapper[4846]: I0202 12:50:59.516465 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2js6p" event={"ID":"e7869f6a-afa5-47db-80c0-e4ae54e939bc","Type":"ContainerStarted","Data":"80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c"} Feb 02 12:50:59 crc kubenswrapper[4846]: I0202 12:50:59.542035 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2js6p" podStartSLOduration=2.836250769 podStartE2EDuration="5.542010335s" podCreationTimestamp="2026-02-02 12:50:54 +0000 UTC" firstStartedPulling="2026-02-02 12:50:56.474525969 +0000 UTC m=+2487.703112842" lastFinishedPulling="2026-02-02 12:50:59.180285545 +0000 UTC m=+2490.408872408" observedRunningTime="2026-02-02 12:50:59.537707271 +0000 UTC m=+2490.766294164" watchObservedRunningTime="2026-02-02 12:50:59.542010335 +0000 UTC m=+2490.770597198" Feb 02 12:51:05 crc kubenswrapper[4846]: I0202 12:51:05.093717 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:51:05 crc kubenswrapper[4846]: I0202 12:51:05.094150 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:51:05 crc kubenswrapper[4846]: I0202 12:51:05.148437 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:51:05 crc kubenswrapper[4846]: I0202 12:51:05.616415 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:51:05 crc kubenswrapper[4846]: I0202 12:51:05.668806 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2js6p"] Feb 02 12:51:07 crc kubenswrapper[4846]: I0202 12:51:07.585572 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2js6p" podUID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerName="registry-server" containerID="cri-o://80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c" gracePeriod=2 Feb 02 12:51:07 crc kubenswrapper[4846]: I0202 12:51:07.964834 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.104534 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-catalog-content\") pod \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.104613 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9cl4\" (UniqueName: \"kubernetes.io/projected/e7869f6a-afa5-47db-80c0-e4ae54e939bc-kube-api-access-l9cl4\") pod \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.104662 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-utilities\") pod \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\" (UID: \"e7869f6a-afa5-47db-80c0-e4ae54e939bc\") " Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.105717 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-utilities" (OuterVolumeSpecName: "utilities") pod "e7869f6a-afa5-47db-80c0-e4ae54e939bc" (UID: "e7869f6a-afa5-47db-80c0-e4ae54e939bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.109592 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7869f6a-afa5-47db-80c0-e4ae54e939bc-kube-api-access-l9cl4" (OuterVolumeSpecName: "kube-api-access-l9cl4") pod "e7869f6a-afa5-47db-80c0-e4ae54e939bc" (UID: "e7869f6a-afa5-47db-80c0-e4ae54e939bc"). InnerVolumeSpecName "kube-api-access-l9cl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.206085 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9cl4\" (UniqueName: \"kubernetes.io/projected/e7869f6a-afa5-47db-80c0-e4ae54e939bc-kube-api-access-l9cl4\") on node \"crc\" DevicePath \"\"" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.206137 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.599194 4846 generic.go:334] "Generic (PLEG): container finished" podID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerID="80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c" exitCode=0 Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.599248 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2js6p" event={"ID":"e7869f6a-afa5-47db-80c0-e4ae54e939bc","Type":"ContainerDied","Data":"80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c"} Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.599305 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2js6p" event={"ID":"e7869f6a-afa5-47db-80c0-e4ae54e939bc","Type":"ContainerDied","Data":"941acc7d2c1ecb2c61ff60ddf95cf1549b197f99078988c2e99af113e6fef435"} Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.599311 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2js6p" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.599331 4846 scope.go:117] "RemoveContainer" containerID="80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.621180 4846 scope.go:117] "RemoveContainer" containerID="faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.639956 4846 scope.go:117] "RemoveContainer" containerID="7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.675357 4846 scope.go:117] "RemoveContainer" containerID="80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c" Feb 02 12:51:08 crc kubenswrapper[4846]: E0202 12:51:08.676039 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c\": container with ID starting with 80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c not found: ID does not exist" containerID="80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.676104 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c"} err="failed to get container status \"80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c\": rpc error: code = NotFound desc = could not find container \"80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c\": container with ID starting with 80db812a4dc264b42579dd82a1387286af1ff18f55a3cfd03aa893953dfdea8c not found: ID does not exist" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.676131 4846 scope.go:117] "RemoveContainer" containerID="faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a" Feb 02 12:51:08 crc kubenswrapper[4846]: E0202 12:51:08.676696 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a\": container with ID starting with faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a not found: ID does not exist" containerID="faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.676730 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a"} err="failed to get container status \"faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a\": rpc error: code = NotFound desc = could not find container \"faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a\": container with ID starting with faa726d073e4116dc5fa4f48353a1b1538b5bb0a3221299b2225956116a3743a not found: ID does not exist" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.676748 4846 scope.go:117] "RemoveContainer" containerID="7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec" Feb 02 12:51:08 crc kubenswrapper[4846]: E0202 12:51:08.677117 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec\": container with ID starting with 7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec not found: ID does not exist" containerID="7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.677184 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec"} err="failed to get container status \"7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec\": rpc error: code = NotFound desc = could not find container \"7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec\": container with ID starting with 7f6ca2bb24ce30d715a17982c8a260862aa60c7ab456f546c15a68c1fbcee6ec not found: ID does not exist" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.755020 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7869f6a-afa5-47db-80c0-e4ae54e939bc" (UID: "e7869f6a-afa5-47db-80c0-e4ae54e939bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.816722 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7869f6a-afa5-47db-80c0-e4ae54e939bc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.931361 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2js6p"] Feb 02 12:51:08 crc kubenswrapper[4846]: I0202 12:51:08.937824 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2js6p"] Feb 02 12:51:09 crc kubenswrapper[4846]: I0202 12:51:09.433922 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" path="/var/lib/kubelet/pods/e7869f6a-afa5-47db-80c0-e4ae54e939bc/volumes" Feb 02 12:51:14 crc kubenswrapper[4846]: I0202 12:51:14.423735 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:51:14 crc kubenswrapper[4846]: I0202 12:51:14.646593 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"9c8fe16fc8317c37aa6ce7e6311396c32b17f6b7c919ae6bb531928492cf9949"} Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.592709 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-brchh"] Feb 02 12:51:46 crc kubenswrapper[4846]: E0202 12:51:46.593421 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerName="registry-server" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.593433 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerName="registry-server" Feb 02 12:51:46 crc kubenswrapper[4846]: E0202 12:51:46.593445 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerName="extract-utilities" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.593451 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerName="extract-utilities" Feb 02 12:51:46 crc kubenswrapper[4846]: E0202 12:51:46.593462 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerName="extract-content" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.593469 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerName="extract-content" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.593604 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7869f6a-afa5-47db-80c0-e4ae54e939bc" containerName="registry-server" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.594697 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.618532 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-brchh"] Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.664392 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-utilities\") pod \"redhat-marketplace-brchh\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.664694 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-catalog-content\") pod \"redhat-marketplace-brchh\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.664817 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjdqd\" (UniqueName: \"kubernetes.io/projected/2803123c-d7da-48c8-8f33-d850ecd9e870-kube-api-access-fjdqd\") pod \"redhat-marketplace-brchh\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.766664 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjdqd\" (UniqueName: \"kubernetes.io/projected/2803123c-d7da-48c8-8f33-d850ecd9e870-kube-api-access-fjdqd\") pod \"redhat-marketplace-brchh\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.767036 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-utilities\") pod \"redhat-marketplace-brchh\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.767161 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-catalog-content\") pod \"redhat-marketplace-brchh\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.767680 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-utilities\") pod \"redhat-marketplace-brchh\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.767766 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-catalog-content\") pod \"redhat-marketplace-brchh\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.792869 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjdqd\" (UniqueName: \"kubernetes.io/projected/2803123c-d7da-48c8-8f33-d850ecd9e870-kube-api-access-fjdqd\") pod \"redhat-marketplace-brchh\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:46 crc kubenswrapper[4846]: I0202 12:51:46.931152 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:47 crc kubenswrapper[4846]: I0202 12:51:47.385744 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-brchh"] Feb 02 12:51:47 crc kubenswrapper[4846]: I0202 12:51:47.918259 4846 generic.go:334] "Generic (PLEG): container finished" podID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerID="3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f" exitCode=0 Feb 02 12:51:47 crc kubenswrapper[4846]: I0202 12:51:47.918381 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brchh" event={"ID":"2803123c-d7da-48c8-8f33-d850ecd9e870","Type":"ContainerDied","Data":"3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f"} Feb 02 12:51:47 crc kubenswrapper[4846]: I0202 12:51:47.918807 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brchh" event={"ID":"2803123c-d7da-48c8-8f33-d850ecd9e870","Type":"ContainerStarted","Data":"e69505abad97a8956bd8b0088a5af5fb8f8704120989c9f209060d282a04a540"} Feb 02 12:51:49 crc kubenswrapper[4846]: I0202 12:51:49.937550 4846 generic.go:334] "Generic (PLEG): container finished" podID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerID="c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256" exitCode=0 Feb 02 12:51:49 crc kubenswrapper[4846]: I0202 12:51:49.937637 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brchh" event={"ID":"2803123c-d7da-48c8-8f33-d850ecd9e870","Type":"ContainerDied","Data":"c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256"} Feb 02 12:51:50 crc kubenswrapper[4846]: I0202 12:51:50.949385 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brchh" event={"ID":"2803123c-d7da-48c8-8f33-d850ecd9e870","Type":"ContainerStarted","Data":"f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa"} Feb 02 12:51:50 crc kubenswrapper[4846]: I0202 12:51:50.979075 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-brchh" podStartSLOduration=2.496272329 podStartE2EDuration="4.979055047s" podCreationTimestamp="2026-02-02 12:51:46 +0000 UTC" firstStartedPulling="2026-02-02 12:51:47.920302891 +0000 UTC m=+2539.148889754" lastFinishedPulling="2026-02-02 12:51:50.403085589 +0000 UTC m=+2541.631672472" observedRunningTime="2026-02-02 12:51:50.971265529 +0000 UTC m=+2542.199852392" watchObservedRunningTime="2026-02-02 12:51:50.979055047 +0000 UTC m=+2542.207641910" Feb 02 12:51:56 crc kubenswrapper[4846]: I0202 12:51:56.931949 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:56 crc kubenswrapper[4846]: I0202 12:51:56.932249 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:57 crc kubenswrapper[4846]: I0202 12:51:57.001938 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:57 crc kubenswrapper[4846]: I0202 12:51:57.049496 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:57 crc kubenswrapper[4846]: I0202 12:51:57.235119 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-brchh"] Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.012504 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-brchh" podUID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerName="registry-server" containerID="cri-o://f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa" gracePeriod=2 Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.388812 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.457144 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjdqd\" (UniqueName: \"kubernetes.io/projected/2803123c-d7da-48c8-8f33-d850ecd9e870-kube-api-access-fjdqd\") pod \"2803123c-d7da-48c8-8f33-d850ecd9e870\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.457240 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-catalog-content\") pod \"2803123c-d7da-48c8-8f33-d850ecd9e870\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.457276 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-utilities\") pod \"2803123c-d7da-48c8-8f33-d850ecd9e870\" (UID: \"2803123c-d7da-48c8-8f33-d850ecd9e870\") " Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.458612 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-utilities" (OuterVolumeSpecName: "utilities") pod "2803123c-d7da-48c8-8f33-d850ecd9e870" (UID: "2803123c-d7da-48c8-8f33-d850ecd9e870"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.465943 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2803123c-d7da-48c8-8f33-d850ecd9e870-kube-api-access-fjdqd" (OuterVolumeSpecName: "kube-api-access-fjdqd") pod "2803123c-d7da-48c8-8f33-d850ecd9e870" (UID: "2803123c-d7da-48c8-8f33-d850ecd9e870"). InnerVolumeSpecName "kube-api-access-fjdqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.481455 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2803123c-d7da-48c8-8f33-d850ecd9e870" (UID: "2803123c-d7da-48c8-8f33-d850ecd9e870"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.559926 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjdqd\" (UniqueName: \"kubernetes.io/projected/2803123c-d7da-48c8-8f33-d850ecd9e870-kube-api-access-fjdqd\") on node \"crc\" DevicePath \"\"" Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.560079 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:51:59 crc kubenswrapper[4846]: I0202 12:51:59.560699 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2803123c-d7da-48c8-8f33-d850ecd9e870-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.023564 4846 generic.go:334] "Generic (PLEG): container finished" podID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerID="f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa" exitCode=0 Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.023703 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brchh" Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.025054 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brchh" event={"ID":"2803123c-d7da-48c8-8f33-d850ecd9e870","Type":"ContainerDied","Data":"f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa"} Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.025599 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brchh" event={"ID":"2803123c-d7da-48c8-8f33-d850ecd9e870","Type":"ContainerDied","Data":"e69505abad97a8956bd8b0088a5af5fb8f8704120989c9f209060d282a04a540"} Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.025662 4846 scope.go:117] "RemoveContainer" containerID="f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa" Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.064945 4846 scope.go:117] "RemoveContainer" containerID="c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256" Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.069976 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-brchh"] Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.076509 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-brchh"] Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.086211 4846 scope.go:117] "RemoveContainer" containerID="3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f" Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.102785 4846 scope.go:117] "RemoveContainer" containerID="f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa" Feb 02 12:52:00 crc kubenswrapper[4846]: E0202 12:52:00.103219 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa\": container with ID starting with f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa not found: ID does not exist" containerID="f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa" Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.103348 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa"} err="failed to get container status \"f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa\": rpc error: code = NotFound desc = could not find container \"f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa\": container with ID starting with f7222097e6ab860f3da3d7a1741ad3561956bce96550d521acd5965a41054afa not found: ID does not exist" Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.103431 4846 scope.go:117] "RemoveContainer" containerID="c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256" Feb 02 12:52:00 crc kubenswrapper[4846]: E0202 12:52:00.103740 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256\": container with ID starting with c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256 not found: ID does not exist" containerID="c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256" Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.103828 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256"} err="failed to get container status \"c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256\": rpc error: code = NotFound desc = could not find container \"c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256\": container with ID starting with c6988183d527a46a58f8fcdac703eb31d93d841ec5fa18d40873d9d30100a256 not found: ID does not exist" Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.103940 4846 scope.go:117] "RemoveContainer" containerID="3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f" Feb 02 12:52:00 crc kubenswrapper[4846]: E0202 12:52:00.104217 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f\": container with ID starting with 3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f not found: ID does not exist" containerID="3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f" Feb 02 12:52:00 crc kubenswrapper[4846]: I0202 12:52:00.104347 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f"} err="failed to get container status \"3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f\": rpc error: code = NotFound desc = could not find container \"3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f\": container with ID starting with 3486d114b3e0acde2b3d2cd6e147fd253c38491b1f9324fd230bd83897ba449f not found: ID does not exist" Feb 02 12:52:01 crc kubenswrapper[4846]: I0202 12:52:01.439194 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2803123c-d7da-48c8-8f33-d850ecd9e870" path="/var/lib/kubelet/pods/2803123c-d7da-48c8-8f33-d850ecd9e870/volumes" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.029585 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p2w7h"] Feb 02 12:53:08 crc kubenswrapper[4846]: E0202 12:53:08.030478 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerName="extract-content" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.030493 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerName="extract-content" Feb 02 12:53:08 crc kubenswrapper[4846]: E0202 12:53:08.030524 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerName="registry-server" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.030533 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerName="registry-server" Feb 02 12:53:08 crc kubenswrapper[4846]: E0202 12:53:08.030554 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerName="extract-utilities" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.030561 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerName="extract-utilities" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.030765 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2803123c-d7da-48c8-8f33-d850ecd9e870" containerName="registry-server" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.031966 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.040418 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p2w7h"] Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.218408 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvzlq\" (UniqueName: \"kubernetes.io/projected/f7ba1cf4-978f-460e-93ac-3063b141cec0-kube-api-access-lvzlq\") pod \"certified-operators-p2w7h\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.218723 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-catalog-content\") pod \"certified-operators-p2w7h\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.218886 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-utilities\") pod \"certified-operators-p2w7h\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.319719 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-utilities\") pod \"certified-operators-p2w7h\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.320060 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvzlq\" (UniqueName: \"kubernetes.io/projected/f7ba1cf4-978f-460e-93ac-3063b141cec0-kube-api-access-lvzlq\") pod \"certified-operators-p2w7h\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.320147 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-catalog-content\") pod \"certified-operators-p2w7h\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.320157 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-utilities\") pod \"certified-operators-p2w7h\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.320461 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-catalog-content\") pod \"certified-operators-p2w7h\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.338677 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvzlq\" (UniqueName: \"kubernetes.io/projected/f7ba1cf4-978f-460e-93ac-3063b141cec0-kube-api-access-lvzlq\") pod \"certified-operators-p2w7h\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.360244 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:08 crc kubenswrapper[4846]: I0202 12:53:08.782290 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p2w7h"] Feb 02 12:53:09 crc kubenswrapper[4846]: I0202 12:53:09.533266 4846 generic.go:334] "Generic (PLEG): container finished" podID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerID="98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee" exitCode=0 Feb 02 12:53:09 crc kubenswrapper[4846]: I0202 12:53:09.533371 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2w7h" event={"ID":"f7ba1cf4-978f-460e-93ac-3063b141cec0","Type":"ContainerDied","Data":"98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee"} Feb 02 12:53:09 crc kubenswrapper[4846]: I0202 12:53:09.533433 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2w7h" event={"ID":"f7ba1cf4-978f-460e-93ac-3063b141cec0","Type":"ContainerStarted","Data":"c461c829591deb3320d063621344bcfd8f209f7f0ba9bccd42077e18586133b7"} Feb 02 12:53:10 crc kubenswrapper[4846]: I0202 12:53:10.540165 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2w7h" event={"ID":"f7ba1cf4-978f-460e-93ac-3063b141cec0","Type":"ContainerStarted","Data":"47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac"} Feb 02 12:53:11 crc kubenswrapper[4846]: I0202 12:53:11.549829 4846 generic.go:334] "Generic (PLEG): container finished" podID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerID="47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac" exitCode=0 Feb 02 12:53:11 crc kubenswrapper[4846]: I0202 12:53:11.549884 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2w7h" event={"ID":"f7ba1cf4-978f-460e-93ac-3063b141cec0","Type":"ContainerDied","Data":"47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac"} Feb 02 12:53:12 crc kubenswrapper[4846]: I0202 12:53:12.561285 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2w7h" event={"ID":"f7ba1cf4-978f-460e-93ac-3063b141cec0","Type":"ContainerStarted","Data":"8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6"} Feb 02 12:53:12 crc kubenswrapper[4846]: I0202 12:53:12.592847 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p2w7h" podStartSLOduration=2.11951596 podStartE2EDuration="4.59282182s" podCreationTimestamp="2026-02-02 12:53:08 +0000 UTC" firstStartedPulling="2026-02-02 12:53:09.536073824 +0000 UTC m=+2620.764660687" lastFinishedPulling="2026-02-02 12:53:12.009379684 +0000 UTC m=+2623.237966547" observedRunningTime="2026-02-02 12:53:12.580211397 +0000 UTC m=+2623.808798280" watchObservedRunningTime="2026-02-02 12:53:12.59282182 +0000 UTC m=+2623.821408693" Feb 02 12:53:18 crc kubenswrapper[4846]: I0202 12:53:18.361123 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:18 crc kubenswrapper[4846]: I0202 12:53:18.361578 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:18 crc kubenswrapper[4846]: I0202 12:53:18.427271 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:18 crc kubenswrapper[4846]: I0202 12:53:18.678727 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:18 crc kubenswrapper[4846]: I0202 12:53:18.722360 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p2w7h"] Feb 02 12:53:20 crc kubenswrapper[4846]: I0202 12:53:20.630437 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p2w7h" podUID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerName="registry-server" containerID="cri-o://8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6" gracePeriod=2 Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.530673 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.622481 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-catalog-content\") pod \"f7ba1cf4-978f-460e-93ac-3063b141cec0\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.622745 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-utilities\") pod \"f7ba1cf4-978f-460e-93ac-3063b141cec0\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.622810 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvzlq\" (UniqueName: \"kubernetes.io/projected/f7ba1cf4-978f-460e-93ac-3063b141cec0-kube-api-access-lvzlq\") pod \"f7ba1cf4-978f-460e-93ac-3063b141cec0\" (UID: \"f7ba1cf4-978f-460e-93ac-3063b141cec0\") " Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.623637 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-utilities" (OuterVolumeSpecName: "utilities") pod "f7ba1cf4-978f-460e-93ac-3063b141cec0" (UID: "f7ba1cf4-978f-460e-93ac-3063b141cec0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.627462 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7ba1cf4-978f-460e-93ac-3063b141cec0-kube-api-access-lvzlq" (OuterVolumeSpecName: "kube-api-access-lvzlq") pod "f7ba1cf4-978f-460e-93ac-3063b141cec0" (UID: "f7ba1cf4-978f-460e-93ac-3063b141cec0"). InnerVolumeSpecName "kube-api-access-lvzlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.638024 4846 generic.go:334] "Generic (PLEG): container finished" podID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerID="8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6" exitCode=0 Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.638067 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2w7h" event={"ID":"f7ba1cf4-978f-460e-93ac-3063b141cec0","Type":"ContainerDied","Data":"8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6"} Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.638096 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p2w7h" event={"ID":"f7ba1cf4-978f-460e-93ac-3063b141cec0","Type":"ContainerDied","Data":"c461c829591deb3320d063621344bcfd8f209f7f0ba9bccd42077e18586133b7"} Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.638117 4846 scope.go:117] "RemoveContainer" containerID="8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.638240 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p2w7h" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.670267 4846 scope.go:117] "RemoveContainer" containerID="47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.685299 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7ba1cf4-978f-460e-93ac-3063b141cec0" (UID: "f7ba1cf4-978f-460e-93ac-3063b141cec0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.687397 4846 scope.go:117] "RemoveContainer" containerID="98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.711480 4846 scope.go:117] "RemoveContainer" containerID="8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6" Feb 02 12:53:21 crc kubenswrapper[4846]: E0202 12:53:21.712093 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6\": container with ID starting with 8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6 not found: ID does not exist" containerID="8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.712206 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6"} err="failed to get container status \"8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6\": rpc error: code = NotFound desc = could not find container \"8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6\": container with ID starting with 8c7e20ca427f1de0d40f627ae8899dfbd55674f8e4be7235a6ce72cc38246ac6 not found: ID does not exist" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.712238 4846 scope.go:117] "RemoveContainer" containerID="47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac" Feb 02 12:53:21 crc kubenswrapper[4846]: E0202 12:53:21.712706 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac\": container with ID starting with 47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac not found: ID does not exist" containerID="47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.712749 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac"} err="failed to get container status \"47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac\": rpc error: code = NotFound desc = could not find container \"47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac\": container with ID starting with 47577a8d16224ec5e32d554798f7aa67ea8123a8d8f722278716e3ae920992ac not found: ID does not exist" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.712790 4846 scope.go:117] "RemoveContainer" containerID="98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee" Feb 02 12:53:21 crc kubenswrapper[4846]: E0202 12:53:21.713199 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee\": container with ID starting with 98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee not found: ID does not exist" containerID="98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.713227 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee"} err="failed to get container status \"98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee\": rpc error: code = NotFound desc = could not find container \"98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee\": container with ID starting with 98499d87afaef3a59ae7f5e6c31fe8318f04e77239dfe06b82e81160c77dc2ee not found: ID does not exist" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.724130 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.724150 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7ba1cf4-978f-460e-93ac-3063b141cec0-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.724160 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvzlq\" (UniqueName: \"kubernetes.io/projected/f7ba1cf4-978f-460e-93ac-3063b141cec0-kube-api-access-lvzlq\") on node \"crc\" DevicePath \"\"" Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.972829 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p2w7h"] Feb 02 12:53:21 crc kubenswrapper[4846]: I0202 12:53:21.979489 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p2w7h"] Feb 02 12:53:23 crc kubenswrapper[4846]: I0202 12:53:23.431746 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7ba1cf4-978f-460e-93ac-3063b141cec0" path="/var/lib/kubelet/pods/f7ba1cf4-978f-460e-93ac-3063b141cec0/volumes" Feb 02 12:53:30 crc kubenswrapper[4846]: I0202 12:53:30.479667 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:53:30 crc kubenswrapper[4846]: I0202 12:53:30.480402 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:54:00 crc kubenswrapper[4846]: I0202 12:54:00.478849 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:54:00 crc kubenswrapper[4846]: I0202 12:54:00.479317 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:54:30 crc kubenswrapper[4846]: I0202 12:54:30.479371 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:54:30 crc kubenswrapper[4846]: I0202 12:54:30.480043 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:54:30 crc kubenswrapper[4846]: I0202 12:54:30.480103 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:54:30 crc kubenswrapper[4846]: I0202 12:54:30.480933 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9c8fe16fc8317c37aa6ce7e6311396c32b17f6b7c919ae6bb531928492cf9949"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:54:30 crc kubenswrapper[4846]: I0202 12:54:30.481019 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://9c8fe16fc8317c37aa6ce7e6311396c32b17f6b7c919ae6bb531928492cf9949" gracePeriod=600 Feb 02 12:54:31 crc kubenswrapper[4846]: I0202 12:54:31.348072 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="9c8fe16fc8317c37aa6ce7e6311396c32b17f6b7c919ae6bb531928492cf9949" exitCode=0 Feb 02 12:54:31 crc kubenswrapper[4846]: I0202 12:54:31.348140 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"9c8fe16fc8317c37aa6ce7e6311396c32b17f6b7c919ae6bb531928492cf9949"} Feb 02 12:54:31 crc kubenswrapper[4846]: I0202 12:54:31.348412 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c"} Feb 02 12:54:31 crc kubenswrapper[4846]: I0202 12:54:31.348435 4846 scope.go:117] "RemoveContainer" containerID="e460fd2c65d10ee0d75608fa4e3ca61ba677a4afa8eee11acb26d815ebcec7d9" Feb 02 12:56:30 crc kubenswrapper[4846]: I0202 12:56:30.479077 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:56:30 crc kubenswrapper[4846]: I0202 12:56:30.479905 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:57:00 crc kubenswrapper[4846]: I0202 12:57:00.479182 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:57:00 crc kubenswrapper[4846]: I0202 12:57:00.479901 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:57:30 crc kubenswrapper[4846]: I0202 12:57:30.479737 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 12:57:30 crc kubenswrapper[4846]: I0202 12:57:30.480396 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 12:57:30 crc kubenswrapper[4846]: I0202 12:57:30.480453 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 12:57:30 crc kubenswrapper[4846]: I0202 12:57:30.481406 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 12:57:30 crc kubenswrapper[4846]: I0202 12:57:30.481502 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" gracePeriod=600 Feb 02 12:57:30 crc kubenswrapper[4846]: E0202 12:57:30.609609 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:57:30 crc kubenswrapper[4846]: I0202 12:57:30.848075 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" exitCode=0 Feb 02 12:57:30 crc kubenswrapper[4846]: I0202 12:57:30.848221 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c"} Feb 02 12:57:30 crc kubenswrapper[4846]: I0202 12:57:30.848792 4846 scope.go:117] "RemoveContainer" containerID="9c8fe16fc8317c37aa6ce7e6311396c32b17f6b7c919ae6bb531928492cf9949" Feb 02 12:57:30 crc kubenswrapper[4846]: I0202 12:57:30.849591 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:57:30 crc kubenswrapper[4846]: E0202 12:57:30.850094 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:57:43 crc kubenswrapper[4846]: I0202 12:57:43.423959 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:57:43 crc kubenswrapper[4846]: E0202 12:57:43.424981 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:57:55 crc kubenswrapper[4846]: I0202 12:57:55.424346 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:57:55 crc kubenswrapper[4846]: E0202 12:57:55.425408 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:58:06 crc kubenswrapper[4846]: I0202 12:58:06.424994 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:58:06 crc kubenswrapper[4846]: E0202 12:58:06.426659 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:58:17 crc kubenswrapper[4846]: I0202 12:58:17.422915 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:58:17 crc kubenswrapper[4846]: E0202 12:58:17.423768 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:58:28 crc kubenswrapper[4846]: I0202 12:58:28.424308 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:58:28 crc kubenswrapper[4846]: E0202 12:58:28.425157 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:58:42 crc kubenswrapper[4846]: I0202 12:58:42.423897 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:58:42 crc kubenswrapper[4846]: E0202 12:58:42.425131 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:58:54 crc kubenswrapper[4846]: I0202 12:58:54.423832 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:58:54 crc kubenswrapper[4846]: E0202 12:58:54.424536 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.579522 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rxv8w"] Feb 02 12:59:02 crc kubenswrapper[4846]: E0202 12:59:02.581173 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerName="registry-server" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.581282 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerName="registry-server" Feb 02 12:59:02 crc kubenswrapper[4846]: E0202 12:59:02.581325 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerName="extract-utilities" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.581336 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerName="extract-utilities" Feb 02 12:59:02 crc kubenswrapper[4846]: E0202 12:59:02.581365 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerName="extract-content" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.581373 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerName="extract-content" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.581610 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ba1cf4-978f-460e-93ac-3063b141cec0" containerName="registry-server" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.583394 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.597658 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rxv8w"] Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.768848 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-utilities\") pod \"redhat-operators-rxv8w\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.768913 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-catalog-content\") pod \"redhat-operators-rxv8w\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.768975 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65k77\" (UniqueName: \"kubernetes.io/projected/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-kube-api-access-65k77\") pod \"redhat-operators-rxv8w\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.870179 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65k77\" (UniqueName: \"kubernetes.io/projected/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-kube-api-access-65k77\") pod \"redhat-operators-rxv8w\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.870286 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-utilities\") pod \"redhat-operators-rxv8w\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.870357 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-catalog-content\") pod \"redhat-operators-rxv8w\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.870944 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-utilities\") pod \"redhat-operators-rxv8w\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.871209 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-catalog-content\") pod \"redhat-operators-rxv8w\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.902398 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65k77\" (UniqueName: \"kubernetes.io/projected/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-kube-api-access-65k77\") pod \"redhat-operators-rxv8w\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:02 crc kubenswrapper[4846]: I0202 12:59:02.907917 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:03 crc kubenswrapper[4846]: I0202 12:59:03.329379 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rxv8w"] Feb 02 12:59:03 crc kubenswrapper[4846]: I0202 12:59:03.633709 4846 generic.go:334] "Generic (PLEG): container finished" podID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerID="f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00" exitCode=0 Feb 02 12:59:03 crc kubenswrapper[4846]: I0202 12:59:03.633752 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxv8w" event={"ID":"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1","Type":"ContainerDied","Data":"f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00"} Feb 02 12:59:03 crc kubenswrapper[4846]: I0202 12:59:03.633780 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxv8w" event={"ID":"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1","Type":"ContainerStarted","Data":"c943e60cc8cfe13478875f24283093bc85d8c683d32b548ed08ac266160f431b"} Feb 02 12:59:03 crc kubenswrapper[4846]: I0202 12:59:03.635964 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 12:59:05 crc kubenswrapper[4846]: I0202 12:59:05.649912 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxv8w" event={"ID":"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1","Type":"ContainerStarted","Data":"44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf"} Feb 02 12:59:06 crc kubenswrapper[4846]: I0202 12:59:06.423205 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:59:06 crc kubenswrapper[4846]: E0202 12:59:06.423512 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:59:06 crc kubenswrapper[4846]: I0202 12:59:06.664344 4846 generic.go:334] "Generic (PLEG): container finished" podID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerID="44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf" exitCode=0 Feb 02 12:59:06 crc kubenswrapper[4846]: I0202 12:59:06.664402 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxv8w" event={"ID":"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1","Type":"ContainerDied","Data":"44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf"} Feb 02 12:59:07 crc kubenswrapper[4846]: I0202 12:59:07.680244 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxv8w" event={"ID":"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1","Type":"ContainerStarted","Data":"4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685"} Feb 02 12:59:07 crc kubenswrapper[4846]: I0202 12:59:07.700857 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rxv8w" podStartSLOduration=1.989696058 podStartE2EDuration="5.70083833s" podCreationTimestamp="2026-02-02 12:59:02 +0000 UTC" firstStartedPulling="2026-02-02 12:59:03.635400194 +0000 UTC m=+2974.863987067" lastFinishedPulling="2026-02-02 12:59:07.346542456 +0000 UTC m=+2978.575129339" observedRunningTime="2026-02-02 12:59:07.695191613 +0000 UTC m=+2978.923778476" watchObservedRunningTime="2026-02-02 12:59:07.70083833 +0000 UTC m=+2978.929425193" Feb 02 12:59:12 crc kubenswrapper[4846]: I0202 12:59:12.908133 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:12 crc kubenswrapper[4846]: I0202 12:59:12.909836 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:13 crc kubenswrapper[4846]: I0202 12:59:13.946703 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rxv8w" podUID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerName="registry-server" probeResult="failure" output=< Feb 02 12:59:13 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 12:59:13 crc kubenswrapper[4846]: > Feb 02 12:59:18 crc kubenswrapper[4846]: I0202 12:59:18.424399 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:59:18 crc kubenswrapper[4846]: E0202 12:59:18.424956 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:59:22 crc kubenswrapper[4846]: I0202 12:59:22.959776 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:23 crc kubenswrapper[4846]: I0202 12:59:23.014385 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:23 crc kubenswrapper[4846]: I0202 12:59:23.194944 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rxv8w"] Feb 02 12:59:24 crc kubenswrapper[4846]: I0202 12:59:24.809981 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rxv8w" podUID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerName="registry-server" containerID="cri-o://4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685" gracePeriod=2 Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.287939 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.488422 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-catalog-content\") pod \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.488951 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-utilities\") pod \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.489111 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65k77\" (UniqueName: \"kubernetes.io/projected/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-kube-api-access-65k77\") pod \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\" (UID: \"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1\") " Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.489854 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-utilities" (OuterVolumeSpecName: "utilities") pod "6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" (UID: "6398bd73-38e8-46cb-a5c6-f5c8fe546eb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.489957 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.500840 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-kube-api-access-65k77" (OuterVolumeSpecName: "kube-api-access-65k77") pod "6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" (UID: "6398bd73-38e8-46cb-a5c6-f5c8fe546eb1"). InnerVolumeSpecName "kube-api-access-65k77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.591443 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65k77\" (UniqueName: \"kubernetes.io/projected/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-kube-api-access-65k77\") on node \"crc\" DevicePath \"\"" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.647495 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" (UID: "6398bd73-38e8-46cb-a5c6-f5c8fe546eb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.692296 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.829794 4846 generic.go:334] "Generic (PLEG): container finished" podID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerID="4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685" exitCode=0 Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.829855 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxv8w" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.829854 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxv8w" event={"ID":"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1","Type":"ContainerDied","Data":"4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685"} Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.829997 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxv8w" event={"ID":"6398bd73-38e8-46cb-a5c6-f5c8fe546eb1","Type":"ContainerDied","Data":"c943e60cc8cfe13478875f24283093bc85d8c683d32b548ed08ac266160f431b"} Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.830023 4846 scope.go:117] "RemoveContainer" containerID="4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.853729 4846 scope.go:117] "RemoveContainer" containerID="44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.878742 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rxv8w"] Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.882361 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rxv8w"] Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.883843 4846 scope.go:117] "RemoveContainer" containerID="f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.903472 4846 scope.go:117] "RemoveContainer" containerID="4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685" Feb 02 12:59:25 crc kubenswrapper[4846]: E0202 12:59:25.903921 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685\": container with ID starting with 4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685 not found: ID does not exist" containerID="4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.903953 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685"} err="failed to get container status \"4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685\": rpc error: code = NotFound desc = could not find container \"4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685\": container with ID starting with 4650309625cdb3a3c44e95c22ea93d3e7c7e754c9bdde06720bf9d96916fa685 not found: ID does not exist" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.903976 4846 scope.go:117] "RemoveContainer" containerID="44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf" Feb 02 12:59:25 crc kubenswrapper[4846]: E0202 12:59:25.904268 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf\": container with ID starting with 44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf not found: ID does not exist" containerID="44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.904289 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf"} err="failed to get container status \"44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf\": rpc error: code = NotFound desc = could not find container \"44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf\": container with ID starting with 44f2165026b8bc2b0e87ab05cd7a630163cb11d201b419849f24541f7a848caf not found: ID does not exist" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.904303 4846 scope.go:117] "RemoveContainer" containerID="f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00" Feb 02 12:59:25 crc kubenswrapper[4846]: E0202 12:59:25.904567 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00\": container with ID starting with f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00 not found: ID does not exist" containerID="f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00" Feb 02 12:59:25 crc kubenswrapper[4846]: I0202 12:59:25.904589 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00"} err="failed to get container status \"f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00\": rpc error: code = NotFound desc = could not find container \"f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00\": container with ID starting with f22e460772eb521e52ce29200957554d6b30339e24c23779d327b5067c54df00 not found: ID does not exist" Feb 02 12:59:27 crc kubenswrapper[4846]: I0202 12:59:27.431756 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" path="/var/lib/kubelet/pods/6398bd73-38e8-46cb-a5c6-f5c8fe546eb1/volumes" Feb 02 12:59:32 crc kubenswrapper[4846]: I0202 12:59:32.424268 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:59:32 crc kubenswrapper[4846]: E0202 12:59:32.425001 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 12:59:46 crc kubenswrapper[4846]: I0202 12:59:46.423837 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 12:59:46 crc kubenswrapper[4846]: E0202 12:59:46.425009 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.165891 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f"] Feb 02 13:00:00 crc kubenswrapper[4846]: E0202 13:00:00.166866 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerName="extract-content" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.166884 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerName="extract-content" Feb 02 13:00:00 crc kubenswrapper[4846]: E0202 13:00:00.166906 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerName="registry-server" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.166914 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerName="registry-server" Feb 02 13:00:00 crc kubenswrapper[4846]: E0202 13:00:00.166936 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerName="extract-utilities" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.166943 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerName="extract-utilities" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.167130 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6398bd73-38e8-46cb-a5c6-f5c8fe546eb1" containerName="registry-server" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.167797 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.169880 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.169965 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.182487 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f"] Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.267691 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8vzb\" (UniqueName: \"kubernetes.io/projected/d826ccc7-488f-4c36-9427-015882a3cbcd-kube-api-access-f8vzb\") pod \"collect-profiles-29500620-4xq7f\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.267764 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d826ccc7-488f-4c36-9427-015882a3cbcd-secret-volume\") pod \"collect-profiles-29500620-4xq7f\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.268119 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d826ccc7-488f-4c36-9427-015882a3cbcd-config-volume\") pod \"collect-profiles-29500620-4xq7f\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.369609 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d826ccc7-488f-4c36-9427-015882a3cbcd-config-volume\") pod \"collect-profiles-29500620-4xq7f\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.369724 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8vzb\" (UniqueName: \"kubernetes.io/projected/d826ccc7-488f-4c36-9427-015882a3cbcd-kube-api-access-f8vzb\") pod \"collect-profiles-29500620-4xq7f\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.369755 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d826ccc7-488f-4c36-9427-015882a3cbcd-secret-volume\") pod \"collect-profiles-29500620-4xq7f\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.370789 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d826ccc7-488f-4c36-9427-015882a3cbcd-config-volume\") pod \"collect-profiles-29500620-4xq7f\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.378751 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d826ccc7-488f-4c36-9427-015882a3cbcd-secret-volume\") pod \"collect-profiles-29500620-4xq7f\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.391237 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8vzb\" (UniqueName: \"kubernetes.io/projected/d826ccc7-488f-4c36-9427-015882a3cbcd-kube-api-access-f8vzb\") pod \"collect-profiles-29500620-4xq7f\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.526763 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:00 crc kubenswrapper[4846]: I0202 13:00:00.985006 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f"] Feb 02 13:00:01 crc kubenswrapper[4846]: I0202 13:00:01.176137 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" event={"ID":"d826ccc7-488f-4c36-9427-015882a3cbcd","Type":"ContainerStarted","Data":"2fb34f76229970f84c5cf931187b40d3199c06f257995dcf7afcafa2919be0c2"} Feb 02 13:00:01 crc kubenswrapper[4846]: I0202 13:00:01.176440 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" event={"ID":"d826ccc7-488f-4c36-9427-015882a3cbcd","Type":"ContainerStarted","Data":"58a4bd95df4ebdfac73f6f4eb7d9f9ad443260f1128c8cffe071d9f600144b1c"} Feb 02 13:00:01 crc kubenswrapper[4846]: I0202 13:00:01.193303 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" podStartSLOduration=1.193286114 podStartE2EDuration="1.193286114s" podCreationTimestamp="2026-02-02 13:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:00:01.190401324 +0000 UTC m=+3032.418988187" watchObservedRunningTime="2026-02-02 13:00:01.193286114 +0000 UTC m=+3032.421872977" Feb 02 13:00:01 crc kubenswrapper[4846]: I0202 13:00:01.423373 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:00:01 crc kubenswrapper[4846]: E0202 13:00:01.423586 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:00:02 crc kubenswrapper[4846]: I0202 13:00:02.184295 4846 generic.go:334] "Generic (PLEG): container finished" podID="d826ccc7-488f-4c36-9427-015882a3cbcd" containerID="2fb34f76229970f84c5cf931187b40d3199c06f257995dcf7afcafa2919be0c2" exitCode=0 Feb 02 13:00:02 crc kubenswrapper[4846]: I0202 13:00:02.184336 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" event={"ID":"d826ccc7-488f-4c36-9427-015882a3cbcd","Type":"ContainerDied","Data":"2fb34f76229970f84c5cf931187b40d3199c06f257995dcf7afcafa2919be0c2"} Feb 02 13:00:03 crc kubenswrapper[4846]: I0202 13:00:03.491766 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:03 crc kubenswrapper[4846]: I0202 13:00:03.620732 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d826ccc7-488f-4c36-9427-015882a3cbcd-secret-volume\") pod \"d826ccc7-488f-4c36-9427-015882a3cbcd\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " Feb 02 13:00:03 crc kubenswrapper[4846]: I0202 13:00:03.620807 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d826ccc7-488f-4c36-9427-015882a3cbcd-config-volume\") pod \"d826ccc7-488f-4c36-9427-015882a3cbcd\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " Feb 02 13:00:03 crc kubenswrapper[4846]: I0202 13:00:03.620833 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8vzb\" (UniqueName: \"kubernetes.io/projected/d826ccc7-488f-4c36-9427-015882a3cbcd-kube-api-access-f8vzb\") pod \"d826ccc7-488f-4c36-9427-015882a3cbcd\" (UID: \"d826ccc7-488f-4c36-9427-015882a3cbcd\") " Feb 02 13:00:03 crc kubenswrapper[4846]: I0202 13:00:03.621912 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d826ccc7-488f-4c36-9427-015882a3cbcd-config-volume" (OuterVolumeSpecName: "config-volume") pod "d826ccc7-488f-4c36-9427-015882a3cbcd" (UID: "d826ccc7-488f-4c36-9427-015882a3cbcd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:00:03 crc kubenswrapper[4846]: I0202 13:00:03.627857 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d826ccc7-488f-4c36-9427-015882a3cbcd-kube-api-access-f8vzb" (OuterVolumeSpecName: "kube-api-access-f8vzb") pod "d826ccc7-488f-4c36-9427-015882a3cbcd" (UID: "d826ccc7-488f-4c36-9427-015882a3cbcd"). InnerVolumeSpecName "kube-api-access-f8vzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:00:03 crc kubenswrapper[4846]: I0202 13:00:03.628641 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d826ccc7-488f-4c36-9427-015882a3cbcd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d826ccc7-488f-4c36-9427-015882a3cbcd" (UID: "d826ccc7-488f-4c36-9427-015882a3cbcd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:00:03 crc kubenswrapper[4846]: I0202 13:00:03.722152 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d826ccc7-488f-4c36-9427-015882a3cbcd-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 13:00:03 crc kubenswrapper[4846]: I0202 13:00:03.722188 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d826ccc7-488f-4c36-9427-015882a3cbcd-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 13:00:03 crc kubenswrapper[4846]: I0202 13:00:03.722198 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8vzb\" (UniqueName: \"kubernetes.io/projected/d826ccc7-488f-4c36-9427-015882a3cbcd-kube-api-access-f8vzb\") on node \"crc\" DevicePath \"\"" Feb 02 13:00:04 crc kubenswrapper[4846]: I0202 13:00:04.201666 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" event={"ID":"d826ccc7-488f-4c36-9427-015882a3cbcd","Type":"ContainerDied","Data":"58a4bd95df4ebdfac73f6f4eb7d9f9ad443260f1128c8cffe071d9f600144b1c"} Feb 02 13:00:04 crc kubenswrapper[4846]: I0202 13:00:04.201912 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58a4bd95df4ebdfac73f6f4eb7d9f9ad443260f1128c8cffe071d9f600144b1c" Feb 02 13:00:04 crc kubenswrapper[4846]: I0202 13:00:04.201720 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f" Feb 02 13:00:04 crc kubenswrapper[4846]: I0202 13:00:04.264727 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb"] Feb 02 13:00:04 crc kubenswrapper[4846]: I0202 13:00:04.269054 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500575-zrjwb"] Feb 02 13:00:05 crc kubenswrapper[4846]: I0202 13:00:05.437002 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65fb3012-a93f-44f3-b880-06d0c0bdc839" path="/var/lib/kubelet/pods/65fb3012-a93f-44f3-b880-06d0c0bdc839/volumes" Feb 02 13:00:14 crc kubenswrapper[4846]: I0202 13:00:14.424190 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:00:14 crc kubenswrapper[4846]: E0202 13:00:14.425300 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:00:28 crc kubenswrapper[4846]: I0202 13:00:28.424142 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:00:28 crc kubenswrapper[4846]: E0202 13:00:28.424789 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:00:36 crc kubenswrapper[4846]: I0202 13:00:36.693480 4846 scope.go:117] "RemoveContainer" containerID="b7907a8b432b7c421a1995997c6296b92a4d5ecb0be3e314160efb0b1dd0ff0b" Feb 02 13:00:42 crc kubenswrapper[4846]: I0202 13:00:42.424592 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:00:42 crc kubenswrapper[4846]: E0202 13:00:42.426278 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:00:54 crc kubenswrapper[4846]: I0202 13:00:54.423067 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:00:54 crc kubenswrapper[4846]: E0202 13:00:54.423924 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:01:06 crc kubenswrapper[4846]: I0202 13:01:06.423908 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:01:06 crc kubenswrapper[4846]: E0202 13:01:06.424819 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:01:18 crc kubenswrapper[4846]: I0202 13:01:18.427292 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:01:18 crc kubenswrapper[4846]: E0202 13:01:18.428607 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:01:33 crc kubenswrapper[4846]: I0202 13:01:33.424537 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:01:33 crc kubenswrapper[4846]: E0202 13:01:33.426008 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:01:47 crc kubenswrapper[4846]: I0202 13:01:47.423149 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:01:47 crc kubenswrapper[4846]: E0202 13:01:47.423853 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.839945 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bfw4d"] Feb 02 13:01:55 crc kubenswrapper[4846]: E0202 13:01:55.840891 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d826ccc7-488f-4c36-9427-015882a3cbcd" containerName="collect-profiles" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.840912 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d826ccc7-488f-4c36-9427-015882a3cbcd" containerName="collect-profiles" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.841113 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d826ccc7-488f-4c36-9427-015882a3cbcd" containerName="collect-profiles" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.842277 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.856188 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfw4d"] Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.874528 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-catalog-content\") pod \"redhat-marketplace-bfw4d\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.874596 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6vrg\" (UniqueName: \"kubernetes.io/projected/4c9f83cf-8125-4993-b28c-1255e5e46912-kube-api-access-c6vrg\") pod \"redhat-marketplace-bfw4d\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.874678 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-utilities\") pod \"redhat-marketplace-bfw4d\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.976231 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-catalog-content\") pod \"redhat-marketplace-bfw4d\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.976544 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6vrg\" (UniqueName: \"kubernetes.io/projected/4c9f83cf-8125-4993-b28c-1255e5e46912-kube-api-access-c6vrg\") pod \"redhat-marketplace-bfw4d\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.976700 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-utilities\") pod \"redhat-marketplace-bfw4d\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.976770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-catalog-content\") pod \"redhat-marketplace-bfw4d\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:55 crc kubenswrapper[4846]: I0202 13:01:55.977253 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-utilities\") pod \"redhat-marketplace-bfw4d\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:56 crc kubenswrapper[4846]: I0202 13:01:56.002540 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6vrg\" (UniqueName: \"kubernetes.io/projected/4c9f83cf-8125-4993-b28c-1255e5e46912-kube-api-access-c6vrg\") pod \"redhat-marketplace-bfw4d\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:56 crc kubenswrapper[4846]: I0202 13:01:56.165987 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:01:56 crc kubenswrapper[4846]: I0202 13:01:56.399004 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfw4d"] Feb 02 13:01:57 crc kubenswrapper[4846]: I0202 13:01:57.247587 4846 generic.go:334] "Generic (PLEG): container finished" podID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerID="27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3" exitCode=0 Feb 02 13:01:57 crc kubenswrapper[4846]: I0202 13:01:57.247680 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfw4d" event={"ID":"4c9f83cf-8125-4993-b28c-1255e5e46912","Type":"ContainerDied","Data":"27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3"} Feb 02 13:01:57 crc kubenswrapper[4846]: I0202 13:01:57.248062 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfw4d" event={"ID":"4c9f83cf-8125-4993-b28c-1255e5e46912","Type":"ContainerStarted","Data":"bce62b635344769724dbb2e203e2a473a540370f345fdd2aadcfa1f5942d20d4"} Feb 02 13:01:58 crc kubenswrapper[4846]: I0202 13:01:58.255830 4846 generic.go:334] "Generic (PLEG): container finished" podID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerID="327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066" exitCode=0 Feb 02 13:01:58 crc kubenswrapper[4846]: I0202 13:01:58.255902 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfw4d" event={"ID":"4c9f83cf-8125-4993-b28c-1255e5e46912","Type":"ContainerDied","Data":"327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066"} Feb 02 13:01:59 crc kubenswrapper[4846]: I0202 13:01:59.266731 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfw4d" event={"ID":"4c9f83cf-8125-4993-b28c-1255e5e46912","Type":"ContainerStarted","Data":"8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27"} Feb 02 13:01:59 crc kubenswrapper[4846]: I0202 13:01:59.288363 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bfw4d" podStartSLOduration=2.899069502 podStartE2EDuration="4.288342053s" podCreationTimestamp="2026-02-02 13:01:55 +0000 UTC" firstStartedPulling="2026-02-02 13:01:57.249734058 +0000 UTC m=+3148.478320921" lastFinishedPulling="2026-02-02 13:01:58.639006569 +0000 UTC m=+3149.867593472" observedRunningTime="2026-02-02 13:01:59.288185089 +0000 UTC m=+3150.516772052" watchObservedRunningTime="2026-02-02 13:01:59.288342053 +0000 UTC m=+3150.516928926" Feb 02 13:02:01 crc kubenswrapper[4846]: I0202 13:02:01.423868 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:02:01 crc kubenswrapper[4846]: E0202 13:02:01.424336 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:02:06 crc kubenswrapper[4846]: I0202 13:02:06.166502 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:02:06 crc kubenswrapper[4846]: I0202 13:02:06.167075 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:02:06 crc kubenswrapper[4846]: I0202 13:02:06.211292 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:02:06 crc kubenswrapper[4846]: I0202 13:02:06.416711 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:02:06 crc kubenswrapper[4846]: I0202 13:02:06.477970 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfw4d"] Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.360217 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bfw4d" podUID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerName="registry-server" containerID="cri-o://8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27" gracePeriod=2 Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.775895 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.875819 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-utilities\") pod \"4c9f83cf-8125-4993-b28c-1255e5e46912\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.875917 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6vrg\" (UniqueName: \"kubernetes.io/projected/4c9f83cf-8125-4993-b28c-1255e5e46912-kube-api-access-c6vrg\") pod \"4c9f83cf-8125-4993-b28c-1255e5e46912\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.876178 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-catalog-content\") pod \"4c9f83cf-8125-4993-b28c-1255e5e46912\" (UID: \"4c9f83cf-8125-4993-b28c-1255e5e46912\") " Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.879673 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-utilities" (OuterVolumeSpecName: "utilities") pod "4c9f83cf-8125-4993-b28c-1255e5e46912" (UID: "4c9f83cf-8125-4993-b28c-1255e5e46912"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.886135 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c9f83cf-8125-4993-b28c-1255e5e46912-kube-api-access-c6vrg" (OuterVolumeSpecName: "kube-api-access-c6vrg") pod "4c9f83cf-8125-4993-b28c-1255e5e46912" (UID: "4c9f83cf-8125-4993-b28c-1255e5e46912"). InnerVolumeSpecName "kube-api-access-c6vrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.906336 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c9f83cf-8125-4993-b28c-1255e5e46912" (UID: "4c9f83cf-8125-4993-b28c-1255e5e46912"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.979041 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.979091 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6vrg\" (UniqueName: \"kubernetes.io/projected/4c9f83cf-8125-4993-b28c-1255e5e46912-kube-api-access-c6vrg\") on node \"crc\" DevicePath \"\"" Feb 02 13:02:08 crc kubenswrapper[4846]: I0202 13:02:08.979104 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c9f83cf-8125-4993-b28c-1255e5e46912-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.389386 4846 generic.go:334] "Generic (PLEG): container finished" podID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerID="8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27" exitCode=0 Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.389424 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfw4d" event={"ID":"4c9f83cf-8125-4993-b28c-1255e5e46912","Type":"ContainerDied","Data":"8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27"} Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.389448 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bfw4d" event={"ID":"4c9f83cf-8125-4993-b28c-1255e5e46912","Type":"ContainerDied","Data":"bce62b635344769724dbb2e203e2a473a540370f345fdd2aadcfa1f5942d20d4"} Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.389467 4846 scope.go:117] "RemoveContainer" containerID="8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27" Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.389498 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bfw4d" Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.426571 4846 scope.go:117] "RemoveContainer" containerID="327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066" Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.442749 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfw4d"] Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.449071 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bfw4d"] Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.466230 4846 scope.go:117] "RemoveContainer" containerID="27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3" Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.487115 4846 scope.go:117] "RemoveContainer" containerID="8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27" Feb 02 13:02:09 crc kubenswrapper[4846]: E0202 13:02:09.487601 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27\": container with ID starting with 8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27 not found: ID does not exist" containerID="8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27" Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.487659 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27"} err="failed to get container status \"8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27\": rpc error: code = NotFound desc = could not find container \"8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27\": container with ID starting with 8f71e421db5da32697cad2a98672c369f946cf279bbf6f2ef83d4fb25cf1da27 not found: ID does not exist" Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.487689 4846 scope.go:117] "RemoveContainer" containerID="327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066" Feb 02 13:02:09 crc kubenswrapper[4846]: E0202 13:02:09.488105 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066\": container with ID starting with 327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066 not found: ID does not exist" containerID="327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066" Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.488140 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066"} err="failed to get container status \"327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066\": rpc error: code = NotFound desc = could not find container \"327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066\": container with ID starting with 327d83abcf68dbaa46806dd0adbf32b073fc06709aaab32c8c2adc87864e1066 not found: ID does not exist" Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.488162 4846 scope.go:117] "RemoveContainer" containerID="27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3" Feb 02 13:02:09 crc kubenswrapper[4846]: E0202 13:02:09.488416 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3\": container with ID starting with 27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3 not found: ID does not exist" containerID="27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3" Feb 02 13:02:09 crc kubenswrapper[4846]: I0202 13:02:09.488448 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3"} err="failed to get container status \"27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3\": rpc error: code = NotFound desc = could not find container \"27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3\": container with ID starting with 27cd66d9cb4d09c51dd6ab3c8ea4d44cc191c2c6036f67b1484d3d78e2e18de3 not found: ID does not exist" Feb 02 13:02:11 crc kubenswrapper[4846]: I0202 13:02:11.436116 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c9f83cf-8125-4993-b28c-1255e5e46912" path="/var/lib/kubelet/pods/4c9f83cf-8125-4993-b28c-1255e5e46912/volumes" Feb 02 13:02:16 crc kubenswrapper[4846]: I0202 13:02:16.423774 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:02:16 crc kubenswrapper[4846]: E0202 13:02:16.424300 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:02:27 crc kubenswrapper[4846]: I0202 13:02:27.423168 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:02:27 crc kubenswrapper[4846]: E0202 13:02:27.424029 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:02:39 crc kubenswrapper[4846]: I0202 13:02:39.428346 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:02:40 crc kubenswrapper[4846]: I0202 13:02:40.686866 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"c737d3e4b437d80a9e19942f95596d5cf9e628590f73d6944387b8cbc8f76885"} Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.420009 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xz68p"] Feb 02 13:03:59 crc kubenswrapper[4846]: E0202 13:03:59.421666 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerName="extract-utilities" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.421695 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerName="extract-utilities" Feb 02 13:03:59 crc kubenswrapper[4846]: E0202 13:03:59.421741 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerName="registry-server" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.421754 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerName="registry-server" Feb 02 13:03:59 crc kubenswrapper[4846]: E0202 13:03:59.421779 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerName="extract-content" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.421792 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerName="extract-content" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.422797 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c9f83cf-8125-4993-b28c-1255e5e46912" containerName="registry-server" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.424884 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.441924 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xz68p"] Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.470555 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-utilities\") pod \"certified-operators-xz68p\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.471032 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fskd8\" (UniqueName: \"kubernetes.io/projected/e40c003c-0dc6-46a9-b872-fa575598f3a5-kube-api-access-fskd8\") pod \"certified-operators-xz68p\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.471081 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-catalog-content\") pod \"certified-operators-xz68p\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.573117 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-utilities\") pod \"certified-operators-xz68p\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.573233 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fskd8\" (UniqueName: \"kubernetes.io/projected/e40c003c-0dc6-46a9-b872-fa575598f3a5-kube-api-access-fskd8\") pod \"certified-operators-xz68p\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.573284 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-catalog-content\") pod \"certified-operators-xz68p\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.573723 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-utilities\") pod \"certified-operators-xz68p\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.573987 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-catalog-content\") pod \"certified-operators-xz68p\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.622253 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fskd8\" (UniqueName: \"kubernetes.io/projected/e40c003c-0dc6-46a9-b872-fa575598f3a5-kube-api-access-fskd8\") pod \"certified-operators-xz68p\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:03:59 crc kubenswrapper[4846]: I0202 13:03:59.767329 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:04:00 crc kubenswrapper[4846]: I0202 13:04:00.223180 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xz68p"] Feb 02 13:04:00 crc kubenswrapper[4846]: I0202 13:04:00.413522 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xz68p" event={"ID":"e40c003c-0dc6-46a9-b872-fa575598f3a5","Type":"ContainerStarted","Data":"0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9"} Feb 02 13:04:00 crc kubenswrapper[4846]: I0202 13:04:00.413875 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xz68p" event={"ID":"e40c003c-0dc6-46a9-b872-fa575598f3a5","Type":"ContainerStarted","Data":"1b628e8a9f899d6ac55e96b3601ee84282bff12760960a2f3175b062de73e00a"} Feb 02 13:04:01 crc kubenswrapper[4846]: I0202 13:04:01.423474 4846 generic.go:334] "Generic (PLEG): container finished" podID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerID="0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9" exitCode=0 Feb 02 13:04:01 crc kubenswrapper[4846]: I0202 13:04:01.437689 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xz68p" event={"ID":"e40c003c-0dc6-46a9-b872-fa575598f3a5","Type":"ContainerDied","Data":"0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9"} Feb 02 13:04:02 crc kubenswrapper[4846]: I0202 13:04:02.431967 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xz68p" event={"ID":"e40c003c-0dc6-46a9-b872-fa575598f3a5","Type":"ContainerStarted","Data":"1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5"} Feb 02 13:04:03 crc kubenswrapper[4846]: I0202 13:04:03.439818 4846 generic.go:334] "Generic (PLEG): container finished" podID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerID="1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5" exitCode=0 Feb 02 13:04:03 crc kubenswrapper[4846]: I0202 13:04:03.440309 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xz68p" event={"ID":"e40c003c-0dc6-46a9-b872-fa575598f3a5","Type":"ContainerDied","Data":"1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5"} Feb 02 13:04:04 crc kubenswrapper[4846]: I0202 13:04:04.457364 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xz68p" event={"ID":"e40c003c-0dc6-46a9-b872-fa575598f3a5","Type":"ContainerStarted","Data":"79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1"} Feb 02 13:04:04 crc kubenswrapper[4846]: I0202 13:04:04.492107 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xz68p" podStartSLOduration=2.826610283 podStartE2EDuration="5.492087958s" podCreationTimestamp="2026-02-02 13:03:59 +0000 UTC" firstStartedPulling="2026-02-02 13:04:01.42614113 +0000 UTC m=+3272.654728243" lastFinishedPulling="2026-02-02 13:04:04.091619015 +0000 UTC m=+3275.320205918" observedRunningTime="2026-02-02 13:04:04.489340721 +0000 UTC m=+3275.717927634" watchObservedRunningTime="2026-02-02 13:04:04.492087958 +0000 UTC m=+3275.720674831" Feb 02 13:04:09 crc kubenswrapper[4846]: I0202 13:04:09.768295 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:04:09 crc kubenswrapper[4846]: I0202 13:04:09.769159 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:04:09 crc kubenswrapper[4846]: I0202 13:04:09.849482 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:04:10 crc kubenswrapper[4846]: I0202 13:04:10.569079 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:04:10 crc kubenswrapper[4846]: I0202 13:04:10.639235 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xz68p"] Feb 02 13:04:12 crc kubenswrapper[4846]: I0202 13:04:12.523934 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xz68p" podUID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerName="registry-server" containerID="cri-o://79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1" gracePeriod=2 Feb 02 13:04:12 crc kubenswrapper[4846]: I0202 13:04:12.966149 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.095828 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fskd8\" (UniqueName: \"kubernetes.io/projected/e40c003c-0dc6-46a9-b872-fa575598f3a5-kube-api-access-fskd8\") pod \"e40c003c-0dc6-46a9-b872-fa575598f3a5\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.095972 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-utilities\") pod \"e40c003c-0dc6-46a9-b872-fa575598f3a5\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.096063 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-catalog-content\") pod \"e40c003c-0dc6-46a9-b872-fa575598f3a5\" (UID: \"e40c003c-0dc6-46a9-b872-fa575598f3a5\") " Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.097359 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-utilities" (OuterVolumeSpecName: "utilities") pod "e40c003c-0dc6-46a9-b872-fa575598f3a5" (UID: "e40c003c-0dc6-46a9-b872-fa575598f3a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.102512 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e40c003c-0dc6-46a9-b872-fa575598f3a5-kube-api-access-fskd8" (OuterVolumeSpecName: "kube-api-access-fskd8") pod "e40c003c-0dc6-46a9-b872-fa575598f3a5" (UID: "e40c003c-0dc6-46a9-b872-fa575598f3a5"). InnerVolumeSpecName "kube-api-access-fskd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.141586 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e40c003c-0dc6-46a9-b872-fa575598f3a5" (UID: "e40c003c-0dc6-46a9-b872-fa575598f3a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.197867 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fskd8\" (UniqueName: \"kubernetes.io/projected/e40c003c-0dc6-46a9-b872-fa575598f3a5-kube-api-access-fskd8\") on node \"crc\" DevicePath \"\"" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.197925 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.197948 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e40c003c-0dc6-46a9-b872-fa575598f3a5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.533651 4846 generic.go:334] "Generic (PLEG): container finished" podID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerID="79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1" exitCode=0 Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.533728 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xz68p" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.533725 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xz68p" event={"ID":"e40c003c-0dc6-46a9-b872-fa575598f3a5","Type":"ContainerDied","Data":"79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1"} Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.533820 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xz68p" event={"ID":"e40c003c-0dc6-46a9-b872-fa575598f3a5","Type":"ContainerDied","Data":"1b628e8a9f899d6ac55e96b3601ee84282bff12760960a2f3175b062de73e00a"} Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.533861 4846 scope.go:117] "RemoveContainer" containerID="79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.555151 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xz68p"] Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.566267 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xz68p"] Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.586885 4846 scope.go:117] "RemoveContainer" containerID="1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.612306 4846 scope.go:117] "RemoveContainer" containerID="0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.646790 4846 scope.go:117] "RemoveContainer" containerID="79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1" Feb 02 13:04:13 crc kubenswrapper[4846]: E0202 13:04:13.647469 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1\": container with ID starting with 79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1 not found: ID does not exist" containerID="79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.647521 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1"} err="failed to get container status \"79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1\": rpc error: code = NotFound desc = could not find container \"79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1\": container with ID starting with 79bcbee92075ac6dd2ee5389171c3a245823f4ac840900c7566c9478621e9ae1 not found: ID does not exist" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.647555 4846 scope.go:117] "RemoveContainer" containerID="1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5" Feb 02 13:04:13 crc kubenswrapper[4846]: E0202 13:04:13.648023 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5\": container with ID starting with 1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5 not found: ID does not exist" containerID="1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.648054 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5"} err="failed to get container status \"1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5\": rpc error: code = NotFound desc = could not find container \"1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5\": container with ID starting with 1928b99df41dbbb62ecf258b5efc6877a8cb1144f934cd555762ea92c794fad5 not found: ID does not exist" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.648073 4846 scope.go:117] "RemoveContainer" containerID="0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9" Feb 02 13:04:13 crc kubenswrapper[4846]: E0202 13:04:13.648510 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9\": container with ID starting with 0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9 not found: ID does not exist" containerID="0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9" Feb 02 13:04:13 crc kubenswrapper[4846]: I0202 13:04:13.648539 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9"} err="failed to get container status \"0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9\": rpc error: code = NotFound desc = could not find container \"0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9\": container with ID starting with 0d4a936ad4dc60a9e3ce7c080fd60d418ae1fdd9c94f50cb9af3a2ee9e21efc9 not found: ID does not exist" Feb 02 13:04:15 crc kubenswrapper[4846]: I0202 13:04:15.437879 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e40c003c-0dc6-46a9-b872-fa575598f3a5" path="/var/lib/kubelet/pods/e40c003c-0dc6-46a9-b872-fa575598f3a5/volumes" Feb 02 13:05:00 crc kubenswrapper[4846]: I0202 13:05:00.479534 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:05:00 crc kubenswrapper[4846]: I0202 13:05:00.480221 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.478911 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.479549 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.721764 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hgffc"] Feb 02 13:05:30 crc kubenswrapper[4846]: E0202 13:05:30.722651 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerName="extract-content" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.722673 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerName="extract-content" Feb 02 13:05:30 crc kubenswrapper[4846]: E0202 13:05:30.722711 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerName="registry-server" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.723058 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerName="registry-server" Feb 02 13:05:30 crc kubenswrapper[4846]: E0202 13:05:30.723149 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerName="extract-utilities" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.723158 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerName="extract-utilities" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.723317 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e40c003c-0dc6-46a9-b872-fa575598f3a5" containerName="registry-server" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.724337 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.750951 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hgffc"] Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.803941 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-utilities\") pod \"community-operators-hgffc\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.804030 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-catalog-content\") pod \"community-operators-hgffc\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.804100 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgwgd\" (UniqueName: \"kubernetes.io/projected/cde895c2-ca80-4674-b917-a6e38307cf54-kube-api-access-jgwgd\") pod \"community-operators-hgffc\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.905963 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgwgd\" (UniqueName: \"kubernetes.io/projected/cde895c2-ca80-4674-b917-a6e38307cf54-kube-api-access-jgwgd\") pod \"community-operators-hgffc\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.906042 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-utilities\") pod \"community-operators-hgffc\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.906085 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-catalog-content\") pod \"community-operators-hgffc\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.906537 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-catalog-content\") pod \"community-operators-hgffc\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.906741 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-utilities\") pod \"community-operators-hgffc\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:30 crc kubenswrapper[4846]: I0202 13:05:30.928481 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgwgd\" (UniqueName: \"kubernetes.io/projected/cde895c2-ca80-4674-b917-a6e38307cf54-kube-api-access-jgwgd\") pod \"community-operators-hgffc\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:31 crc kubenswrapper[4846]: I0202 13:05:31.042003 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:31 crc kubenswrapper[4846]: I0202 13:05:31.624190 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hgffc"] Feb 02 13:05:32 crc kubenswrapper[4846]: I0202 13:05:32.140255 4846 generic.go:334] "Generic (PLEG): container finished" podID="cde895c2-ca80-4674-b917-a6e38307cf54" containerID="2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496" exitCode=0 Feb 02 13:05:32 crc kubenswrapper[4846]: I0202 13:05:32.140311 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgffc" event={"ID":"cde895c2-ca80-4674-b917-a6e38307cf54","Type":"ContainerDied","Data":"2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496"} Feb 02 13:05:32 crc kubenswrapper[4846]: I0202 13:05:32.140352 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgffc" event={"ID":"cde895c2-ca80-4674-b917-a6e38307cf54","Type":"ContainerStarted","Data":"b770a4404f6289a6a7fe9f27752f7361928bd303369c49601d0388f72042ec9a"} Feb 02 13:05:32 crc kubenswrapper[4846]: I0202 13:05:32.142978 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 13:05:34 crc kubenswrapper[4846]: I0202 13:05:34.159298 4846 generic.go:334] "Generic (PLEG): container finished" podID="cde895c2-ca80-4674-b917-a6e38307cf54" containerID="1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755" exitCode=0 Feb 02 13:05:34 crc kubenswrapper[4846]: I0202 13:05:34.159373 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgffc" event={"ID":"cde895c2-ca80-4674-b917-a6e38307cf54","Type":"ContainerDied","Data":"1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755"} Feb 02 13:05:35 crc kubenswrapper[4846]: I0202 13:05:35.168533 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgffc" event={"ID":"cde895c2-ca80-4674-b917-a6e38307cf54","Type":"ContainerStarted","Data":"6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8"} Feb 02 13:05:35 crc kubenswrapper[4846]: I0202 13:05:35.190294 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hgffc" podStartSLOduration=2.751755299 podStartE2EDuration="5.190277284s" podCreationTimestamp="2026-02-02 13:05:30 +0000 UTC" firstStartedPulling="2026-02-02 13:05:32.142385834 +0000 UTC m=+3363.370972737" lastFinishedPulling="2026-02-02 13:05:34.580907839 +0000 UTC m=+3365.809494722" observedRunningTime="2026-02-02 13:05:35.189342691 +0000 UTC m=+3366.417929564" watchObservedRunningTime="2026-02-02 13:05:35.190277284 +0000 UTC m=+3366.418864157" Feb 02 13:05:41 crc kubenswrapper[4846]: I0202 13:05:41.043502 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:41 crc kubenswrapper[4846]: I0202 13:05:41.044095 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:41 crc kubenswrapper[4846]: I0202 13:05:41.093786 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:41 crc kubenswrapper[4846]: I0202 13:05:41.261233 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:41 crc kubenswrapper[4846]: I0202 13:05:41.326611 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hgffc"] Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.218479 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hgffc" podUID="cde895c2-ca80-4674-b917-a6e38307cf54" containerName="registry-server" containerID="cri-o://6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8" gracePeriod=2 Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.635508 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.765533 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-catalog-content\") pod \"cde895c2-ca80-4674-b917-a6e38307cf54\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.765704 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgwgd\" (UniqueName: \"kubernetes.io/projected/cde895c2-ca80-4674-b917-a6e38307cf54-kube-api-access-jgwgd\") pod \"cde895c2-ca80-4674-b917-a6e38307cf54\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.765740 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-utilities\") pod \"cde895c2-ca80-4674-b917-a6e38307cf54\" (UID: \"cde895c2-ca80-4674-b917-a6e38307cf54\") " Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.766762 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-utilities" (OuterVolumeSpecName: "utilities") pod "cde895c2-ca80-4674-b917-a6e38307cf54" (UID: "cde895c2-ca80-4674-b917-a6e38307cf54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.770895 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cde895c2-ca80-4674-b917-a6e38307cf54-kube-api-access-jgwgd" (OuterVolumeSpecName: "kube-api-access-jgwgd") pod "cde895c2-ca80-4674-b917-a6e38307cf54" (UID: "cde895c2-ca80-4674-b917-a6e38307cf54"). InnerVolumeSpecName "kube-api-access-jgwgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.832522 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cde895c2-ca80-4674-b917-a6e38307cf54" (UID: "cde895c2-ca80-4674-b917-a6e38307cf54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.867432 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgwgd\" (UniqueName: \"kubernetes.io/projected/cde895c2-ca80-4674-b917-a6e38307cf54-kube-api-access-jgwgd\") on node \"crc\" DevicePath \"\"" Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.867477 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:05:43 crc kubenswrapper[4846]: I0202 13:05:43.867488 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde895c2-ca80-4674-b917-a6e38307cf54-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.227666 4846 generic.go:334] "Generic (PLEG): container finished" podID="cde895c2-ca80-4674-b917-a6e38307cf54" containerID="6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8" exitCode=0 Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.227720 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgffc" event={"ID":"cde895c2-ca80-4674-b917-a6e38307cf54","Type":"ContainerDied","Data":"6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8"} Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.227760 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgffc" event={"ID":"cde895c2-ca80-4674-b917-a6e38307cf54","Type":"ContainerDied","Data":"b770a4404f6289a6a7fe9f27752f7361928bd303369c49601d0388f72042ec9a"} Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.227780 4846 scope.go:117] "RemoveContainer" containerID="6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8" Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.227761 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hgffc" Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.249920 4846 scope.go:117] "RemoveContainer" containerID="1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755" Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.265268 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hgffc"] Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.277237 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hgffc"] Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.289295 4846 scope.go:117] "RemoveContainer" containerID="2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496" Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.315927 4846 scope.go:117] "RemoveContainer" containerID="6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8" Feb 02 13:05:44 crc kubenswrapper[4846]: E0202 13:05:44.316500 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8\": container with ID starting with 6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8 not found: ID does not exist" containerID="6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8" Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.316542 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8"} err="failed to get container status \"6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8\": rpc error: code = NotFound desc = could not find container \"6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8\": container with ID starting with 6e8fb6298fd02fc2a7f358ecda8d7784f54859c702b3cbbe8f0dc200e93a05b8 not found: ID does not exist" Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.316569 4846 scope.go:117] "RemoveContainer" containerID="1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755" Feb 02 13:05:44 crc kubenswrapper[4846]: E0202 13:05:44.317114 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755\": container with ID starting with 1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755 not found: ID does not exist" containerID="1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755" Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.317149 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755"} err="failed to get container status \"1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755\": rpc error: code = NotFound desc = could not find container \"1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755\": container with ID starting with 1c400b8636ffae4d93272750dc5f87dead720d8fbc3ea79a73b14e7a34ee6755 not found: ID does not exist" Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.317174 4846 scope.go:117] "RemoveContainer" containerID="2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496" Feb 02 13:05:44 crc kubenswrapper[4846]: E0202 13:05:44.317548 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496\": container with ID starting with 2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496 not found: ID does not exist" containerID="2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496" Feb 02 13:05:44 crc kubenswrapper[4846]: I0202 13:05:44.317577 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496"} err="failed to get container status \"2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496\": rpc error: code = NotFound desc = could not find container \"2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496\": container with ID starting with 2bca4fb0b8282210aac1bdd81132cf4f9946c4036f2efd614ebfde96fe8b1496 not found: ID does not exist" Feb 02 13:05:45 crc kubenswrapper[4846]: I0202 13:05:45.435362 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cde895c2-ca80-4674-b917-a6e38307cf54" path="/var/lib/kubelet/pods/cde895c2-ca80-4674-b917-a6e38307cf54/volumes" Feb 02 13:06:00 crc kubenswrapper[4846]: I0202 13:06:00.479598 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:06:00 crc kubenswrapper[4846]: I0202 13:06:00.480324 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:06:00 crc kubenswrapper[4846]: I0202 13:06:00.480388 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 13:06:00 crc kubenswrapper[4846]: I0202 13:06:00.481278 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c737d3e4b437d80a9e19942f95596d5cf9e628590f73d6944387b8cbc8f76885"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 13:06:00 crc kubenswrapper[4846]: I0202 13:06:00.481382 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://c737d3e4b437d80a9e19942f95596d5cf9e628590f73d6944387b8cbc8f76885" gracePeriod=600 Feb 02 13:06:01 crc kubenswrapper[4846]: I0202 13:06:01.367551 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="c737d3e4b437d80a9e19942f95596d5cf9e628590f73d6944387b8cbc8f76885" exitCode=0 Feb 02 13:06:01 crc kubenswrapper[4846]: I0202 13:06:01.367607 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"c737d3e4b437d80a9e19942f95596d5cf9e628590f73d6944387b8cbc8f76885"} Feb 02 13:06:01 crc kubenswrapper[4846]: I0202 13:06:01.368181 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c"} Feb 02 13:06:01 crc kubenswrapper[4846]: I0202 13:06:01.368224 4846 scope.go:117] "RemoveContainer" containerID="cfc59838a32af507de646c21cd4acd032a6f8882ca0366afae4cf0e71c9a1b7c" Feb 02 13:08:00 crc kubenswrapper[4846]: I0202 13:08:00.479256 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:08:00 crc kubenswrapper[4846]: I0202 13:08:00.480907 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:08:30 crc kubenswrapper[4846]: I0202 13:08:30.478896 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:08:30 crc kubenswrapper[4846]: I0202 13:08:30.479455 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:09:00 crc kubenswrapper[4846]: I0202 13:09:00.479867 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:09:00 crc kubenswrapper[4846]: I0202 13:09:00.480352 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:09:00 crc kubenswrapper[4846]: I0202 13:09:00.480400 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 13:09:00 crc kubenswrapper[4846]: I0202 13:09:00.481201 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 13:09:00 crc kubenswrapper[4846]: I0202 13:09:00.481271 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" gracePeriod=600 Feb 02 13:09:00 crc kubenswrapper[4846]: E0202 13:09:00.605582 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:09:01 crc kubenswrapper[4846]: I0202 13:09:01.249454 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" exitCode=0 Feb 02 13:09:01 crc kubenswrapper[4846]: I0202 13:09:01.249533 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c"} Feb 02 13:09:01 crc kubenswrapper[4846]: I0202 13:09:01.249824 4846 scope.go:117] "RemoveContainer" containerID="c737d3e4b437d80a9e19942f95596d5cf9e628590f73d6944387b8cbc8f76885" Feb 02 13:09:01 crc kubenswrapper[4846]: I0202 13:09:01.250296 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:09:01 crc kubenswrapper[4846]: E0202 13:09:01.250500 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:09:15 crc kubenswrapper[4846]: I0202 13:09:15.424293 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:09:15 crc kubenswrapper[4846]: E0202 13:09:15.425402 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:09:27 crc kubenswrapper[4846]: I0202 13:09:27.423296 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:09:27 crc kubenswrapper[4846]: E0202 13:09:27.424545 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:09:39 crc kubenswrapper[4846]: I0202 13:09:39.433715 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:09:39 crc kubenswrapper[4846]: E0202 13:09:39.435861 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:09:54 crc kubenswrapper[4846]: I0202 13:09:54.423685 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:09:54 crc kubenswrapper[4846]: E0202 13:09:54.424841 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:10:07 crc kubenswrapper[4846]: I0202 13:10:07.423275 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:10:07 crc kubenswrapper[4846]: E0202 13:10:07.423765 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:10:19 crc kubenswrapper[4846]: I0202 13:10:19.428620 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:10:19 crc kubenswrapper[4846]: E0202 13:10:19.429527 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.577769 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-27r5f"] Feb 02 13:10:29 crc kubenswrapper[4846]: E0202 13:10:29.580162 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde895c2-ca80-4674-b917-a6e38307cf54" containerName="registry-server" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.580265 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde895c2-ca80-4674-b917-a6e38307cf54" containerName="registry-server" Feb 02 13:10:29 crc kubenswrapper[4846]: E0202 13:10:29.580385 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde895c2-ca80-4674-b917-a6e38307cf54" containerName="extract-content" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.580469 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde895c2-ca80-4674-b917-a6e38307cf54" containerName="extract-content" Feb 02 13:10:29 crc kubenswrapper[4846]: E0202 13:10:29.580556 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde895c2-ca80-4674-b917-a6e38307cf54" containerName="extract-utilities" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.580648 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde895c2-ca80-4674-b917-a6e38307cf54" containerName="extract-utilities" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.580909 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cde895c2-ca80-4674-b917-a6e38307cf54" containerName="registry-server" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.582350 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.589067 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-27r5f"] Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.685975 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-utilities\") pod \"redhat-operators-27r5f\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.686357 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24cn6\" (UniqueName: \"kubernetes.io/projected/a173d010-b54c-450d-aef1-c5f75c40b786-kube-api-access-24cn6\") pod \"redhat-operators-27r5f\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.686534 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-catalog-content\") pod \"redhat-operators-27r5f\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.788436 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-catalog-content\") pod \"redhat-operators-27r5f\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.788501 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-utilities\") pod \"redhat-operators-27r5f\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.788525 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24cn6\" (UniqueName: \"kubernetes.io/projected/a173d010-b54c-450d-aef1-c5f75c40b786-kube-api-access-24cn6\") pod \"redhat-operators-27r5f\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.789138 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-catalog-content\") pod \"redhat-operators-27r5f\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.789464 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-utilities\") pod \"redhat-operators-27r5f\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.807883 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24cn6\" (UniqueName: \"kubernetes.io/projected/a173d010-b54c-450d-aef1-c5f75c40b786-kube-api-access-24cn6\") pod \"redhat-operators-27r5f\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:29 crc kubenswrapper[4846]: I0202 13:10:29.913766 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:30 crc kubenswrapper[4846]: I0202 13:10:30.162074 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-27r5f"] Feb 02 13:10:30 crc kubenswrapper[4846]: I0202 13:10:30.979153 4846 generic.go:334] "Generic (PLEG): container finished" podID="a173d010-b54c-450d-aef1-c5f75c40b786" containerID="d0b7881bc40046cdd5a969601374e2f0db13534518252b5ac1eaa5729eccb653" exitCode=0 Feb 02 13:10:30 crc kubenswrapper[4846]: I0202 13:10:30.979199 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27r5f" event={"ID":"a173d010-b54c-450d-aef1-c5f75c40b786","Type":"ContainerDied","Data":"d0b7881bc40046cdd5a969601374e2f0db13534518252b5ac1eaa5729eccb653"} Feb 02 13:10:30 crc kubenswrapper[4846]: I0202 13:10:30.979240 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27r5f" event={"ID":"a173d010-b54c-450d-aef1-c5f75c40b786","Type":"ContainerStarted","Data":"f819d565e34781372be86173303a611ebbd505e4e99f76ca9bfbc54e3c41c5a5"} Feb 02 13:10:31 crc kubenswrapper[4846]: I0202 13:10:31.423906 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:10:31 crc kubenswrapper[4846]: E0202 13:10:31.424328 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:10:31 crc kubenswrapper[4846]: I0202 13:10:31.989394 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27r5f" event={"ID":"a173d010-b54c-450d-aef1-c5f75c40b786","Type":"ContainerStarted","Data":"32658203c5278790badc358d5a390b1505d7432e842c53a6e692a9202e0a499e"} Feb 02 13:10:33 crc kubenswrapper[4846]: I0202 13:10:33.000100 4846 generic.go:334] "Generic (PLEG): container finished" podID="a173d010-b54c-450d-aef1-c5f75c40b786" containerID="32658203c5278790badc358d5a390b1505d7432e842c53a6e692a9202e0a499e" exitCode=0 Feb 02 13:10:33 crc kubenswrapper[4846]: I0202 13:10:33.000163 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27r5f" event={"ID":"a173d010-b54c-450d-aef1-c5f75c40b786","Type":"ContainerDied","Data":"32658203c5278790badc358d5a390b1505d7432e842c53a6e692a9202e0a499e"} Feb 02 13:10:33 crc kubenswrapper[4846]: I0202 13:10:33.001816 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 13:10:34 crc kubenswrapper[4846]: I0202 13:10:34.018638 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27r5f" event={"ID":"a173d010-b54c-450d-aef1-c5f75c40b786","Type":"ContainerStarted","Data":"13734beecdf9d8d93f590fd71eaf9de1a7032dadc59796013de7980335daf7be"} Feb 02 13:10:34 crc kubenswrapper[4846]: I0202 13:10:34.047347 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-27r5f" podStartSLOduration=2.592102321 podStartE2EDuration="5.047324568s" podCreationTimestamp="2026-02-02 13:10:29 +0000 UTC" firstStartedPulling="2026-02-02 13:10:30.981595837 +0000 UTC m=+3662.210182700" lastFinishedPulling="2026-02-02 13:10:33.436818084 +0000 UTC m=+3664.665404947" observedRunningTime="2026-02-02 13:10:34.042677015 +0000 UTC m=+3665.271263948" watchObservedRunningTime="2026-02-02 13:10:34.047324568 +0000 UTC m=+3665.275911431" Feb 02 13:10:39 crc kubenswrapper[4846]: I0202 13:10:39.914819 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:39 crc kubenswrapper[4846]: I0202 13:10:39.915183 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:39 crc kubenswrapper[4846]: I0202 13:10:39.952685 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:40 crc kubenswrapper[4846]: I0202 13:10:40.105062 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:40 crc kubenswrapper[4846]: I0202 13:10:40.190830 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-27r5f"] Feb 02 13:10:42 crc kubenswrapper[4846]: I0202 13:10:42.076015 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-27r5f" podUID="a173d010-b54c-450d-aef1-c5f75c40b786" containerName="registry-server" containerID="cri-o://13734beecdf9d8d93f590fd71eaf9de1a7032dadc59796013de7980335daf7be" gracePeriod=2 Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.098850 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27r5f" event={"ID":"a173d010-b54c-450d-aef1-c5f75c40b786","Type":"ContainerDied","Data":"13734beecdf9d8d93f590fd71eaf9de1a7032dadc59796013de7980335daf7be"} Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.098851 4846 generic.go:334] "Generic (PLEG): container finished" podID="a173d010-b54c-450d-aef1-c5f75c40b786" containerID="13734beecdf9d8d93f590fd71eaf9de1a7032dadc59796013de7980335daf7be" exitCode=0 Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.316199 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.402298 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-catalog-content\") pod \"a173d010-b54c-450d-aef1-c5f75c40b786\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.402357 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24cn6\" (UniqueName: \"kubernetes.io/projected/a173d010-b54c-450d-aef1-c5f75c40b786-kube-api-access-24cn6\") pod \"a173d010-b54c-450d-aef1-c5f75c40b786\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.402396 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-utilities\") pod \"a173d010-b54c-450d-aef1-c5f75c40b786\" (UID: \"a173d010-b54c-450d-aef1-c5f75c40b786\") " Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.403312 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-utilities" (OuterVolumeSpecName: "utilities") pod "a173d010-b54c-450d-aef1-c5f75c40b786" (UID: "a173d010-b54c-450d-aef1-c5f75c40b786"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.409917 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a173d010-b54c-450d-aef1-c5f75c40b786-kube-api-access-24cn6" (OuterVolumeSpecName: "kube-api-access-24cn6") pod "a173d010-b54c-450d-aef1-c5f75c40b786" (UID: "a173d010-b54c-450d-aef1-c5f75c40b786"). InnerVolumeSpecName "kube-api-access-24cn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.423083 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:10:44 crc kubenswrapper[4846]: E0202 13:10:44.423347 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.503856 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24cn6\" (UniqueName: \"kubernetes.io/projected/a173d010-b54c-450d-aef1-c5f75c40b786-kube-api-access-24cn6\") on node \"crc\" DevicePath \"\"" Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.503897 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.547960 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a173d010-b54c-450d-aef1-c5f75c40b786" (UID: "a173d010-b54c-450d-aef1-c5f75c40b786"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:10:44 crc kubenswrapper[4846]: I0202 13:10:44.604809 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a173d010-b54c-450d-aef1-c5f75c40b786-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:10:45 crc kubenswrapper[4846]: I0202 13:10:45.111224 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-27r5f" event={"ID":"a173d010-b54c-450d-aef1-c5f75c40b786","Type":"ContainerDied","Data":"f819d565e34781372be86173303a611ebbd505e4e99f76ca9bfbc54e3c41c5a5"} Feb 02 13:10:45 crc kubenswrapper[4846]: I0202 13:10:45.111816 4846 scope.go:117] "RemoveContainer" containerID="13734beecdf9d8d93f590fd71eaf9de1a7032dadc59796013de7980335daf7be" Feb 02 13:10:45 crc kubenswrapper[4846]: I0202 13:10:45.111398 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-27r5f" Feb 02 13:10:45 crc kubenswrapper[4846]: I0202 13:10:45.141846 4846 scope.go:117] "RemoveContainer" containerID="32658203c5278790badc358d5a390b1505d7432e842c53a6e692a9202e0a499e" Feb 02 13:10:45 crc kubenswrapper[4846]: I0202 13:10:45.174747 4846 scope.go:117] "RemoveContainer" containerID="d0b7881bc40046cdd5a969601374e2f0db13534518252b5ac1eaa5729eccb653" Feb 02 13:10:45 crc kubenswrapper[4846]: I0202 13:10:45.178189 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-27r5f"] Feb 02 13:10:45 crc kubenswrapper[4846]: I0202 13:10:45.187232 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-27r5f"] Feb 02 13:10:45 crc kubenswrapper[4846]: I0202 13:10:45.452662 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a173d010-b54c-450d-aef1-c5f75c40b786" path="/var/lib/kubelet/pods/a173d010-b54c-450d-aef1-c5f75c40b786/volumes" Feb 02 13:10:55 crc kubenswrapper[4846]: I0202 13:10:55.423893 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:10:55 crc kubenswrapper[4846]: E0202 13:10:55.425079 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:11:07 crc kubenswrapper[4846]: I0202 13:11:07.424106 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:11:07 crc kubenswrapper[4846]: E0202 13:11:07.425326 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:11:21 crc kubenswrapper[4846]: I0202 13:11:21.423697 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:11:21 crc kubenswrapper[4846]: E0202 13:11:21.424753 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:11:33 crc kubenswrapper[4846]: I0202 13:11:33.423884 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:11:33 crc kubenswrapper[4846]: E0202 13:11:33.425026 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:11:47 crc kubenswrapper[4846]: I0202 13:11:47.423916 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:11:47 crc kubenswrapper[4846]: E0202 13:11:47.424776 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:11:58 crc kubenswrapper[4846]: I0202 13:11:58.423358 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:11:58 crc kubenswrapper[4846]: E0202 13:11:58.425567 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:12:11 crc kubenswrapper[4846]: I0202 13:12:11.423794 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:12:11 crc kubenswrapper[4846]: E0202 13:12:11.424535 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:12:23 crc kubenswrapper[4846]: I0202 13:12:23.423773 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:12:23 crc kubenswrapper[4846]: E0202 13:12:23.424374 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:12:36 crc kubenswrapper[4846]: I0202 13:12:36.424168 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:12:36 crc kubenswrapper[4846]: E0202 13:12:36.425433 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:12:49 crc kubenswrapper[4846]: I0202 13:12:49.427380 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:12:49 crc kubenswrapper[4846]: E0202 13:12:49.428407 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.178736 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fwmb6"] Feb 02 13:12:53 crc kubenswrapper[4846]: E0202 13:12:53.179451 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a173d010-b54c-450d-aef1-c5f75c40b786" containerName="extract-utilities" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.179465 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a173d010-b54c-450d-aef1-c5f75c40b786" containerName="extract-utilities" Feb 02 13:12:53 crc kubenswrapper[4846]: E0202 13:12:53.179486 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a173d010-b54c-450d-aef1-c5f75c40b786" containerName="extract-content" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.179493 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a173d010-b54c-450d-aef1-c5f75c40b786" containerName="extract-content" Feb 02 13:12:53 crc kubenswrapper[4846]: E0202 13:12:53.179503 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a173d010-b54c-450d-aef1-c5f75c40b786" containerName="registry-server" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.179510 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a173d010-b54c-450d-aef1-c5f75c40b786" containerName="registry-server" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.179684 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a173d010-b54c-450d-aef1-c5f75c40b786" containerName="registry-server" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.180871 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.197262 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwmb6"] Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.264420 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-utilities\") pod \"redhat-marketplace-fwmb6\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.264547 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-catalog-content\") pod \"redhat-marketplace-fwmb6\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.264577 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xthbx\" (UniqueName: \"kubernetes.io/projected/71493ec2-886c-46a6-b59a-73230d4493fa-kube-api-access-xthbx\") pod \"redhat-marketplace-fwmb6\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.366544 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-catalog-content\") pod \"redhat-marketplace-fwmb6\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.366588 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xthbx\" (UniqueName: \"kubernetes.io/projected/71493ec2-886c-46a6-b59a-73230d4493fa-kube-api-access-xthbx\") pod \"redhat-marketplace-fwmb6\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.366663 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-utilities\") pod \"redhat-marketplace-fwmb6\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.367026 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-catalog-content\") pod \"redhat-marketplace-fwmb6\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.367074 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-utilities\") pod \"redhat-marketplace-fwmb6\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.387120 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xthbx\" (UniqueName: \"kubernetes.io/projected/71493ec2-886c-46a6-b59a-73230d4493fa-kube-api-access-xthbx\") pod \"redhat-marketplace-fwmb6\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.521524 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:12:53 crc kubenswrapper[4846]: I0202 13:12:53.954720 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwmb6"] Feb 02 13:12:54 crc kubenswrapper[4846]: I0202 13:12:54.047038 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwmb6" event={"ID":"71493ec2-886c-46a6-b59a-73230d4493fa","Type":"ContainerStarted","Data":"0e2d42a91aa378099879d4034108badd506a4da7b800b579661b1db493e92dbf"} Feb 02 13:12:55 crc kubenswrapper[4846]: I0202 13:12:55.054719 4846 generic.go:334] "Generic (PLEG): container finished" podID="71493ec2-886c-46a6-b59a-73230d4493fa" containerID="2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82" exitCode=0 Feb 02 13:12:55 crc kubenswrapper[4846]: I0202 13:12:55.054784 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwmb6" event={"ID":"71493ec2-886c-46a6-b59a-73230d4493fa","Type":"ContainerDied","Data":"2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82"} Feb 02 13:12:57 crc kubenswrapper[4846]: I0202 13:12:57.070608 4846 generic.go:334] "Generic (PLEG): container finished" podID="71493ec2-886c-46a6-b59a-73230d4493fa" containerID="e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517" exitCode=0 Feb 02 13:12:57 crc kubenswrapper[4846]: I0202 13:12:57.070665 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwmb6" event={"ID":"71493ec2-886c-46a6-b59a-73230d4493fa","Type":"ContainerDied","Data":"e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517"} Feb 02 13:12:58 crc kubenswrapper[4846]: I0202 13:12:58.080863 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwmb6" event={"ID":"71493ec2-886c-46a6-b59a-73230d4493fa","Type":"ContainerStarted","Data":"e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2"} Feb 02 13:12:58 crc kubenswrapper[4846]: I0202 13:12:58.103150 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fwmb6" podStartSLOduration=2.6742289980000002 podStartE2EDuration="5.103128226s" podCreationTimestamp="2026-02-02 13:12:53 +0000 UTC" firstStartedPulling="2026-02-02 13:12:55.056193732 +0000 UTC m=+3806.284780595" lastFinishedPulling="2026-02-02 13:12:57.48509296 +0000 UTC m=+3808.713679823" observedRunningTime="2026-02-02 13:12:58.099954669 +0000 UTC m=+3809.328541552" watchObservedRunningTime="2026-02-02 13:12:58.103128226 +0000 UTC m=+3809.331715089" Feb 02 13:13:00 crc kubenswrapper[4846]: I0202 13:13:00.424120 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:13:00 crc kubenswrapper[4846]: E0202 13:13:00.424773 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:13:03 crc kubenswrapper[4846]: I0202 13:13:03.522529 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:13:03 crc kubenswrapper[4846]: I0202 13:13:03.522944 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:13:03 crc kubenswrapper[4846]: I0202 13:13:03.573653 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:13:04 crc kubenswrapper[4846]: I0202 13:13:04.173039 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:13:04 crc kubenswrapper[4846]: I0202 13:13:04.218203 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwmb6"] Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.135813 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fwmb6" podUID="71493ec2-886c-46a6-b59a-73230d4493fa" containerName="registry-server" containerID="cri-o://e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2" gracePeriod=2 Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.524827 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.661968 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xthbx\" (UniqueName: \"kubernetes.io/projected/71493ec2-886c-46a6-b59a-73230d4493fa-kube-api-access-xthbx\") pod \"71493ec2-886c-46a6-b59a-73230d4493fa\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.662126 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-utilities\") pod \"71493ec2-886c-46a6-b59a-73230d4493fa\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.662253 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-catalog-content\") pod \"71493ec2-886c-46a6-b59a-73230d4493fa\" (UID: \"71493ec2-886c-46a6-b59a-73230d4493fa\") " Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.662907 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-utilities" (OuterVolumeSpecName: "utilities") pod "71493ec2-886c-46a6-b59a-73230d4493fa" (UID: "71493ec2-886c-46a6-b59a-73230d4493fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.670795 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71493ec2-886c-46a6-b59a-73230d4493fa-kube-api-access-xthbx" (OuterVolumeSpecName: "kube-api-access-xthbx") pod "71493ec2-886c-46a6-b59a-73230d4493fa" (UID: "71493ec2-886c-46a6-b59a-73230d4493fa"). InnerVolumeSpecName "kube-api-access-xthbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.696831 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71493ec2-886c-46a6-b59a-73230d4493fa" (UID: "71493ec2-886c-46a6-b59a-73230d4493fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.763356 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.763391 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xthbx\" (UniqueName: \"kubernetes.io/projected/71493ec2-886c-46a6-b59a-73230d4493fa-kube-api-access-xthbx\") on node \"crc\" DevicePath \"\"" Feb 02 13:13:06 crc kubenswrapper[4846]: I0202 13:13:06.763402 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71493ec2-886c-46a6-b59a-73230d4493fa-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.146302 4846 generic.go:334] "Generic (PLEG): container finished" podID="71493ec2-886c-46a6-b59a-73230d4493fa" containerID="e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2" exitCode=0 Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.146379 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fwmb6" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.146428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwmb6" event={"ID":"71493ec2-886c-46a6-b59a-73230d4493fa","Type":"ContainerDied","Data":"e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2"} Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.146914 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fwmb6" event={"ID":"71493ec2-886c-46a6-b59a-73230d4493fa","Type":"ContainerDied","Data":"0e2d42a91aa378099879d4034108badd506a4da7b800b579661b1db493e92dbf"} Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.147019 4846 scope.go:117] "RemoveContainer" containerID="e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.194836 4846 scope.go:117] "RemoveContainer" containerID="e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.196088 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwmb6"] Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.210610 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fwmb6"] Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.219035 4846 scope.go:117] "RemoveContainer" containerID="2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.253514 4846 scope.go:117] "RemoveContainer" containerID="e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2" Feb 02 13:13:07 crc kubenswrapper[4846]: E0202 13:13:07.254040 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2\": container with ID starting with e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2 not found: ID does not exist" containerID="e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.254083 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2"} err="failed to get container status \"e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2\": rpc error: code = NotFound desc = could not find container \"e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2\": container with ID starting with e46aa3005a170cc17e1372dc45a29490687048a7ac367801edfd5482ff2355f2 not found: ID does not exist" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.254110 4846 scope.go:117] "RemoveContainer" containerID="e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517" Feb 02 13:13:07 crc kubenswrapper[4846]: E0202 13:13:07.254416 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517\": container with ID starting with e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517 not found: ID does not exist" containerID="e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.254448 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517"} err="failed to get container status \"e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517\": rpc error: code = NotFound desc = could not find container \"e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517\": container with ID starting with e3e2960647cb7eb5cd8e2679e3b81cf117e92ab05a2a5bf0f7d0cacd8179d517 not found: ID does not exist" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.254470 4846 scope.go:117] "RemoveContainer" containerID="2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82" Feb 02 13:13:07 crc kubenswrapper[4846]: E0202 13:13:07.254837 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82\": container with ID starting with 2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82 not found: ID does not exist" containerID="2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.254905 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82"} err="failed to get container status \"2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82\": rpc error: code = NotFound desc = could not find container \"2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82\": container with ID starting with 2257a09252619c0ce24ea4ff33cee5778fb945b8b835d1749f7cbdd0c2695a82 not found: ID does not exist" Feb 02 13:13:07 crc kubenswrapper[4846]: I0202 13:13:07.433900 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71493ec2-886c-46a6-b59a-73230d4493fa" path="/var/lib/kubelet/pods/71493ec2-886c-46a6-b59a-73230d4493fa/volumes" Feb 02 13:13:14 crc kubenswrapper[4846]: I0202 13:13:14.424020 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:13:14 crc kubenswrapper[4846]: E0202 13:13:14.425091 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:13:25 crc kubenswrapper[4846]: I0202 13:13:25.423688 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:13:25 crc kubenswrapper[4846]: E0202 13:13:25.424773 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:13:39 crc kubenswrapper[4846]: I0202 13:13:39.427444 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:13:39 crc kubenswrapper[4846]: E0202 13:13:39.428352 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:13:50 crc kubenswrapper[4846]: I0202 13:13:50.423007 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:13:50 crc kubenswrapper[4846]: E0202 13:13:50.423755 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.540378 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ks6t8"] Feb 02 13:14:02 crc kubenswrapper[4846]: E0202 13:14:02.541429 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71493ec2-886c-46a6-b59a-73230d4493fa" containerName="registry-server" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.541454 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="71493ec2-886c-46a6-b59a-73230d4493fa" containerName="registry-server" Feb 02 13:14:02 crc kubenswrapper[4846]: E0202 13:14:02.541486 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71493ec2-886c-46a6-b59a-73230d4493fa" containerName="extract-utilities" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.541534 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="71493ec2-886c-46a6-b59a-73230d4493fa" containerName="extract-utilities" Feb 02 13:14:02 crc kubenswrapper[4846]: E0202 13:14:02.541557 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71493ec2-886c-46a6-b59a-73230d4493fa" containerName="extract-content" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.541568 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="71493ec2-886c-46a6-b59a-73230d4493fa" containerName="extract-content" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.541834 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="71493ec2-886c-46a6-b59a-73230d4493fa" containerName="registry-server" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.543486 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.553333 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ks6t8"] Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.721955 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqjgv\" (UniqueName: \"kubernetes.io/projected/2f68e1c1-ce23-4532-9dd5-5a195a959e84-kube-api-access-vqjgv\") pod \"certified-operators-ks6t8\" (UID: \"2f68e1c1-ce23-4532-9dd5-5a195a959e84\") " pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.722135 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f68e1c1-ce23-4532-9dd5-5a195a959e84-catalog-content\") pod \"certified-operators-ks6t8\" (UID: \"2f68e1c1-ce23-4532-9dd5-5a195a959e84\") " pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.722182 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f68e1c1-ce23-4532-9dd5-5a195a959e84-utilities\") pod \"certified-operators-ks6t8\" (UID: \"2f68e1c1-ce23-4532-9dd5-5a195a959e84\") " pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.823698 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f68e1c1-ce23-4532-9dd5-5a195a959e84-catalog-content\") pod \"certified-operators-ks6t8\" (UID: \"2f68e1c1-ce23-4532-9dd5-5a195a959e84\") " pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.823753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f68e1c1-ce23-4532-9dd5-5a195a959e84-utilities\") pod \"certified-operators-ks6t8\" (UID: \"2f68e1c1-ce23-4532-9dd5-5a195a959e84\") " pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.823857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqjgv\" (UniqueName: \"kubernetes.io/projected/2f68e1c1-ce23-4532-9dd5-5a195a959e84-kube-api-access-vqjgv\") pod \"certified-operators-ks6t8\" (UID: \"2f68e1c1-ce23-4532-9dd5-5a195a959e84\") " pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.824319 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f68e1c1-ce23-4532-9dd5-5a195a959e84-catalog-content\") pod \"certified-operators-ks6t8\" (UID: \"2f68e1c1-ce23-4532-9dd5-5a195a959e84\") " pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.824381 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f68e1c1-ce23-4532-9dd5-5a195a959e84-utilities\") pod \"certified-operators-ks6t8\" (UID: \"2f68e1c1-ce23-4532-9dd5-5a195a959e84\") " pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:02 crc kubenswrapper[4846]: I0202 13:14:02.901999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqjgv\" (UniqueName: \"kubernetes.io/projected/2f68e1c1-ce23-4532-9dd5-5a195a959e84-kube-api-access-vqjgv\") pod \"certified-operators-ks6t8\" (UID: \"2f68e1c1-ce23-4532-9dd5-5a195a959e84\") " pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:03 crc kubenswrapper[4846]: I0202 13:14:03.175155 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:03 crc kubenswrapper[4846]: I0202 13:14:03.681697 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ks6t8"] Feb 02 13:14:04 crc kubenswrapper[4846]: I0202 13:14:04.607323 4846 generic.go:334] "Generic (PLEG): container finished" podID="2f68e1c1-ce23-4532-9dd5-5a195a959e84" containerID="ff65043794cd9e4428f15d67ef916657a937726fab58952a1a047191106e41aa" exitCode=0 Feb 02 13:14:04 crc kubenswrapper[4846]: I0202 13:14:04.607396 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ks6t8" event={"ID":"2f68e1c1-ce23-4532-9dd5-5a195a959e84","Type":"ContainerDied","Data":"ff65043794cd9e4428f15d67ef916657a937726fab58952a1a047191106e41aa"} Feb 02 13:14:04 crc kubenswrapper[4846]: I0202 13:14:04.607684 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ks6t8" event={"ID":"2f68e1c1-ce23-4532-9dd5-5a195a959e84","Type":"ContainerStarted","Data":"fe40ad76cb6dec7cdac09f3740d5fbd80a52b287e2f29944afecc40627b8a7dd"} Feb 02 13:14:05 crc kubenswrapper[4846]: I0202 13:14:05.422870 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:14:05 crc kubenswrapper[4846]: I0202 13:14:05.619086 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"b74edd9a518dbcb657b4541bd37c1507460aaa8aebee2269d785534a7f1cfd6b"} Feb 02 13:14:08 crc kubenswrapper[4846]: I0202 13:14:08.637828 4846 generic.go:334] "Generic (PLEG): container finished" podID="2f68e1c1-ce23-4532-9dd5-5a195a959e84" containerID="f1cf4e199f61de34e1fbef9d67b00edcca02dc1e242d321fc7b39576021901cc" exitCode=0 Feb 02 13:14:08 crc kubenswrapper[4846]: I0202 13:14:08.637935 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ks6t8" event={"ID":"2f68e1c1-ce23-4532-9dd5-5a195a959e84","Type":"ContainerDied","Data":"f1cf4e199f61de34e1fbef9d67b00edcca02dc1e242d321fc7b39576021901cc"} Feb 02 13:14:09 crc kubenswrapper[4846]: I0202 13:14:09.648773 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ks6t8" event={"ID":"2f68e1c1-ce23-4532-9dd5-5a195a959e84","Type":"ContainerStarted","Data":"2cfbb42616fa70a5cefa16278fe43b39e569a9dc92d364c91e3d998dacc605c2"} Feb 02 13:14:09 crc kubenswrapper[4846]: I0202 13:14:09.674118 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ks6t8" podStartSLOduration=3.24227945 podStartE2EDuration="7.674096078s" podCreationTimestamp="2026-02-02 13:14:02 +0000 UTC" firstStartedPulling="2026-02-02 13:14:04.610751426 +0000 UTC m=+3875.839338289" lastFinishedPulling="2026-02-02 13:14:09.042568014 +0000 UTC m=+3880.271154917" observedRunningTime="2026-02-02 13:14:09.665594982 +0000 UTC m=+3880.894181875" watchObservedRunningTime="2026-02-02 13:14:09.674096078 +0000 UTC m=+3880.902682951" Feb 02 13:14:13 crc kubenswrapper[4846]: I0202 13:14:13.176136 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:13 crc kubenswrapper[4846]: I0202 13:14:13.176720 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:13 crc kubenswrapper[4846]: I0202 13:14:13.340115 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:23 crc kubenswrapper[4846]: I0202 13:14:23.450204 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ks6t8" Feb 02 13:14:23 crc kubenswrapper[4846]: I0202 13:14:23.526185 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ks6t8"] Feb 02 13:14:23 crc kubenswrapper[4846]: I0202 13:14:23.564560 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vjszt"] Feb 02 13:14:23 crc kubenswrapper[4846]: I0202 13:14:23.564972 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vjszt" podUID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerName="registry-server" containerID="cri-o://3f97a13e10729c233dc9f25bffab25cd76b892ca2736cf199e8382bae879cb0e" gracePeriod=2 Feb 02 13:14:23 crc kubenswrapper[4846]: I0202 13:14:23.754995 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerID="3f97a13e10729c233dc9f25bffab25cd76b892ca2736cf199e8382bae879cb0e" exitCode=0 Feb 02 13:14:23 crc kubenswrapper[4846]: I0202 13:14:23.755061 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vjszt" event={"ID":"5e7333ab-9306-45aa-a3aa-2e637120e4f9","Type":"ContainerDied","Data":"3f97a13e10729c233dc9f25bffab25cd76b892ca2736cf199e8382bae879cb0e"} Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.035758 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vjszt" Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.144676 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-catalog-content\") pod \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.144758 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-utilities\") pod \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.144877 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfx7p\" (UniqueName: \"kubernetes.io/projected/5e7333ab-9306-45aa-a3aa-2e637120e4f9-kube-api-access-nfx7p\") pod \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\" (UID: \"5e7333ab-9306-45aa-a3aa-2e637120e4f9\") " Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.146302 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-utilities" (OuterVolumeSpecName: "utilities") pod "5e7333ab-9306-45aa-a3aa-2e637120e4f9" (UID: "5e7333ab-9306-45aa-a3aa-2e637120e4f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.155906 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e7333ab-9306-45aa-a3aa-2e637120e4f9-kube-api-access-nfx7p" (OuterVolumeSpecName: "kube-api-access-nfx7p") pod "5e7333ab-9306-45aa-a3aa-2e637120e4f9" (UID: "5e7333ab-9306-45aa-a3aa-2e637120e4f9"). InnerVolumeSpecName "kube-api-access-nfx7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.205286 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e7333ab-9306-45aa-a3aa-2e637120e4f9" (UID: "5e7333ab-9306-45aa-a3aa-2e637120e4f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.246495 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.246523 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e7333ab-9306-45aa-a3aa-2e637120e4f9-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.246534 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfx7p\" (UniqueName: \"kubernetes.io/projected/5e7333ab-9306-45aa-a3aa-2e637120e4f9-kube-api-access-nfx7p\") on node \"crc\" DevicePath \"\"" Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.765785 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vjszt" event={"ID":"5e7333ab-9306-45aa-a3aa-2e637120e4f9","Type":"ContainerDied","Data":"3e698eda7cf2d210e118f4a7d054964de326997059739704953aa14b8e4e46ca"} Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.765831 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vjszt" Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.766175 4846 scope.go:117] "RemoveContainer" containerID="3f97a13e10729c233dc9f25bffab25cd76b892ca2736cf199e8382bae879cb0e" Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.799704 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vjszt"] Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.804553 4846 scope.go:117] "RemoveContainer" containerID="caa9fff96d26773cfc441e827143ea5faa16713881df451cd94a4badac58653a" Feb 02 13:14:24 crc kubenswrapper[4846]: I0202 13:14:24.806077 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vjszt"] Feb 02 13:14:25 crc kubenswrapper[4846]: I0202 13:14:25.009934 4846 scope.go:117] "RemoveContainer" containerID="2366afb59437b127afdd2301b249d90f1061f32bcee73e6893dfddfb5ff3e5d3" Feb 02 13:14:25 crc kubenswrapper[4846]: I0202 13:14:25.439785 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" path="/var/lib/kubelet/pods/5e7333ab-9306-45aa-a3aa-2e637120e4f9/volumes" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.192544 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm"] Feb 02 13:15:00 crc kubenswrapper[4846]: E0202 13:15:00.193482 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerName="extract-utilities" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.193497 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerName="extract-utilities" Feb 02 13:15:00 crc kubenswrapper[4846]: E0202 13:15:00.193514 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerName="registry-server" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.193521 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerName="registry-server" Feb 02 13:15:00 crc kubenswrapper[4846]: E0202 13:15:00.193536 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerName="extract-content" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.193543 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerName="extract-content" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.193741 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7333ab-9306-45aa-a3aa-2e637120e4f9" containerName="registry-server" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.194376 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.196336 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.196540 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.205301 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm"] Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.216854 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqn7d\" (UniqueName: \"kubernetes.io/projected/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-kube-api-access-mqn7d\") pod \"collect-profiles-29500635-89vhm\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.217224 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-config-volume\") pod \"collect-profiles-29500635-89vhm\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.217491 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-secret-volume\") pod \"collect-profiles-29500635-89vhm\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.320243 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-secret-volume\") pod \"collect-profiles-29500635-89vhm\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.320490 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqn7d\" (UniqueName: \"kubernetes.io/projected/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-kube-api-access-mqn7d\") pod \"collect-profiles-29500635-89vhm\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.320608 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-config-volume\") pod \"collect-profiles-29500635-89vhm\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.321701 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-config-volume\") pod \"collect-profiles-29500635-89vhm\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.328308 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-secret-volume\") pod \"collect-profiles-29500635-89vhm\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.350174 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqn7d\" (UniqueName: \"kubernetes.io/projected/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-kube-api-access-mqn7d\") pod \"collect-profiles-29500635-89vhm\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.517394 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:00 crc kubenswrapper[4846]: I0202 13:15:00.935219 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm"] Feb 02 13:15:01 crc kubenswrapper[4846]: I0202 13:15:01.025177 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" event={"ID":"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b","Type":"ContainerStarted","Data":"a75d46c775735080ad6396d159153b8a750897272b79593d1225b782a4322d84"} Feb 02 13:15:02 crc kubenswrapper[4846]: I0202 13:15:02.044206 4846 generic.go:334] "Generic (PLEG): container finished" podID="cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b" containerID="ddc740203fd5cf82960549078d5f6395a8cb9e8a6df3d7e01a826068e292b53f" exitCode=0 Feb 02 13:15:02 crc kubenswrapper[4846]: I0202 13:15:02.044348 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" event={"ID":"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b","Type":"ContainerDied","Data":"ddc740203fd5cf82960549078d5f6395a8cb9e8a6df3d7e01a826068e292b53f"} Feb 02 13:15:03 crc kubenswrapper[4846]: I0202 13:15:03.405773 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:03 crc kubenswrapper[4846]: I0202 13:15:03.465675 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-config-volume\") pod \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " Feb 02 13:15:03 crc kubenswrapper[4846]: I0202 13:15:03.465746 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-secret-volume\") pod \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " Feb 02 13:15:03 crc kubenswrapper[4846]: I0202 13:15:03.465799 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqn7d\" (UniqueName: \"kubernetes.io/projected/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-kube-api-access-mqn7d\") pod \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\" (UID: \"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b\") " Feb 02 13:15:03 crc kubenswrapper[4846]: I0202 13:15:03.466416 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-config-volume" (OuterVolumeSpecName: "config-volume") pod "cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b" (UID: "cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:15:03 crc kubenswrapper[4846]: I0202 13:15:03.466832 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 13:15:03 crc kubenswrapper[4846]: I0202 13:15:03.472658 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b" (UID: "cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:15:03 crc kubenswrapper[4846]: I0202 13:15:03.484940 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-kube-api-access-mqn7d" (OuterVolumeSpecName: "kube-api-access-mqn7d") pod "cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b" (UID: "cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b"). InnerVolumeSpecName "kube-api-access-mqn7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:15:03 crc kubenswrapper[4846]: I0202 13:15:03.568482 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 13:15:03 crc kubenswrapper[4846]: I0202 13:15:03.568523 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqn7d\" (UniqueName: \"kubernetes.io/projected/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b-kube-api-access-mqn7d\") on node \"crc\" DevicePath \"\"" Feb 02 13:15:04 crc kubenswrapper[4846]: I0202 13:15:04.059078 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" event={"ID":"cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b","Type":"ContainerDied","Data":"a75d46c775735080ad6396d159153b8a750897272b79593d1225b782a4322d84"} Feb 02 13:15:04 crc kubenswrapper[4846]: I0202 13:15:04.059391 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a75d46c775735080ad6396d159153b8a750897272b79593d1225b782a4322d84" Feb 02 13:15:04 crc kubenswrapper[4846]: I0202 13:15:04.059199 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm" Feb 02 13:15:04 crc kubenswrapper[4846]: I0202 13:15:04.478018 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs"] Feb 02 13:15:04 crc kubenswrapper[4846]: I0202 13:15:04.483789 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500590-vpxfs"] Feb 02 13:15:05 crc kubenswrapper[4846]: I0202 13:15:05.434775 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1c1ec69-7585-45e7-812b-df08a8ede792" path="/var/lib/kubelet/pods/e1c1ec69-7585-45e7-812b-df08a8ede792/volumes" Feb 02 13:15:37 crc kubenswrapper[4846]: I0202 13:15:37.380896 4846 scope.go:117] "RemoveContainer" containerID="b2355786fff0258ff67af38a657cb1150394658b2101b78593ceac673efc5f8f" Feb 02 13:16:30 crc kubenswrapper[4846]: I0202 13:16:30.479204 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:16:30 crc kubenswrapper[4846]: I0202 13:16:30.479751 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.304972 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mds9h"] Feb 02 13:16:37 crc kubenswrapper[4846]: E0202 13:16:37.306726 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b" containerName="collect-profiles" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.306760 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b" containerName="collect-profiles" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.307105 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b" containerName="collect-profiles" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.309222 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.318338 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mds9h"] Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.354354 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-catalog-content\") pod \"community-operators-mds9h\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.354428 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-utilities\") pod \"community-operators-mds9h\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.354468 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn7qw\" (UniqueName: \"kubernetes.io/projected/76cbbc95-f545-4929-8a5a-307da82dfd81-kube-api-access-gn7qw\") pod \"community-operators-mds9h\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.456056 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-catalog-content\") pod \"community-operators-mds9h\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.456103 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-utilities\") pod \"community-operators-mds9h\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.456130 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn7qw\" (UniqueName: \"kubernetes.io/projected/76cbbc95-f545-4929-8a5a-307da82dfd81-kube-api-access-gn7qw\") pod \"community-operators-mds9h\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.457426 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-utilities\") pod \"community-operators-mds9h\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.457494 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-catalog-content\") pod \"community-operators-mds9h\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.478062 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn7qw\" (UniqueName: \"kubernetes.io/projected/76cbbc95-f545-4929-8a5a-307da82dfd81-kube-api-access-gn7qw\") pod \"community-operators-mds9h\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:37 crc kubenswrapper[4846]: I0202 13:16:37.630169 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:38 crc kubenswrapper[4846]: I0202 13:16:38.166756 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mds9h"] Feb 02 13:16:38 crc kubenswrapper[4846]: E0202 13:16:38.442804 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76cbbc95_f545_4929_8a5a_307da82dfd81.slice/crio-conmon-e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86.scope\": RecentStats: unable to find data in memory cache]" Feb 02 13:16:38 crc kubenswrapper[4846]: I0202 13:16:38.810255 4846 generic.go:334] "Generic (PLEG): container finished" podID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerID="e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86" exitCode=0 Feb 02 13:16:38 crc kubenswrapper[4846]: I0202 13:16:38.810297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9h" event={"ID":"76cbbc95-f545-4929-8a5a-307da82dfd81","Type":"ContainerDied","Data":"e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86"} Feb 02 13:16:38 crc kubenswrapper[4846]: I0202 13:16:38.810326 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9h" event={"ID":"76cbbc95-f545-4929-8a5a-307da82dfd81","Type":"ContainerStarted","Data":"fc2c09e2033ca9467bdbf9e3dfc285f7cbac8979277d433dc4918694a2247c9e"} Feb 02 13:16:38 crc kubenswrapper[4846]: I0202 13:16:38.813925 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 13:16:39 crc kubenswrapper[4846]: I0202 13:16:39.825499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9h" event={"ID":"76cbbc95-f545-4929-8a5a-307da82dfd81","Type":"ContainerStarted","Data":"79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4"} Feb 02 13:16:40 crc kubenswrapper[4846]: I0202 13:16:40.833437 4846 generic.go:334] "Generic (PLEG): container finished" podID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerID="79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4" exitCode=0 Feb 02 13:16:40 crc kubenswrapper[4846]: I0202 13:16:40.833636 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9h" event={"ID":"76cbbc95-f545-4929-8a5a-307da82dfd81","Type":"ContainerDied","Data":"79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4"} Feb 02 13:16:40 crc kubenswrapper[4846]: I0202 13:16:40.833779 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9h" event={"ID":"76cbbc95-f545-4929-8a5a-307da82dfd81","Type":"ContainerStarted","Data":"8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55"} Feb 02 13:16:47 crc kubenswrapper[4846]: I0202 13:16:47.630907 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:47 crc kubenswrapper[4846]: I0202 13:16:47.631409 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:47 crc kubenswrapper[4846]: I0202 13:16:47.683337 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:47 crc kubenswrapper[4846]: I0202 13:16:47.705382 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mds9h" podStartSLOduration=9.284259706 podStartE2EDuration="10.705358855s" podCreationTimestamp="2026-02-02 13:16:37 +0000 UTC" firstStartedPulling="2026-02-02 13:16:38.81346223 +0000 UTC m=+4030.042049103" lastFinishedPulling="2026-02-02 13:16:40.234561389 +0000 UTC m=+4031.463148252" observedRunningTime="2026-02-02 13:16:40.85065626 +0000 UTC m=+4032.079243143" watchObservedRunningTime="2026-02-02 13:16:47.705358855 +0000 UTC m=+4038.933945718" Feb 02 13:16:47 crc kubenswrapper[4846]: I0202 13:16:47.921414 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:47 crc kubenswrapper[4846]: I0202 13:16:47.965229 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mds9h"] Feb 02 13:16:49 crc kubenswrapper[4846]: I0202 13:16:49.896061 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mds9h" podUID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerName="registry-server" containerID="cri-o://8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55" gracePeriod=2 Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.751052 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.907761 4846 generic.go:334] "Generic (PLEG): container finished" podID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerID="8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55" exitCode=0 Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.907821 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mds9h" Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.907833 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9h" event={"ID":"76cbbc95-f545-4929-8a5a-307da82dfd81","Type":"ContainerDied","Data":"8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55"} Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.907872 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9h" event={"ID":"76cbbc95-f545-4929-8a5a-307da82dfd81","Type":"ContainerDied","Data":"fc2c09e2033ca9467bdbf9e3dfc285f7cbac8979277d433dc4918694a2247c9e"} Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.907908 4846 scope.go:117] "RemoveContainer" containerID="8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55" Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.926014 4846 scope.go:117] "RemoveContainer" containerID="79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4" Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.944176 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-utilities\") pod \"76cbbc95-f545-4929-8a5a-307da82dfd81\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.944316 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-catalog-content\") pod \"76cbbc95-f545-4929-8a5a-307da82dfd81\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.944430 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn7qw\" (UniqueName: \"kubernetes.io/projected/76cbbc95-f545-4929-8a5a-307da82dfd81-kube-api-access-gn7qw\") pod \"76cbbc95-f545-4929-8a5a-307da82dfd81\" (UID: \"76cbbc95-f545-4929-8a5a-307da82dfd81\") " Feb 02 13:16:50 crc kubenswrapper[4846]: I0202 13:16:50.945509 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-utilities" (OuterVolumeSpecName: "utilities") pod "76cbbc95-f545-4929-8a5a-307da82dfd81" (UID: "76cbbc95-f545-4929-8a5a-307da82dfd81"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.000428 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76cbbc95-f545-4929-8a5a-307da82dfd81-kube-api-access-gn7qw" (OuterVolumeSpecName: "kube-api-access-gn7qw") pod "76cbbc95-f545-4929-8a5a-307da82dfd81" (UID: "76cbbc95-f545-4929-8a5a-307da82dfd81"). InnerVolumeSpecName "kube-api-access-gn7qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.011827 4846 scope.go:117] "RemoveContainer" containerID="e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.035081 4846 scope.go:117] "RemoveContainer" containerID="8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55" Feb 02 13:16:51 crc kubenswrapper[4846]: E0202 13:16:51.035676 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55\": container with ID starting with 8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55 not found: ID does not exist" containerID="8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.035716 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55"} err="failed to get container status \"8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55\": rpc error: code = NotFound desc = could not find container \"8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55\": container with ID starting with 8119baa22edac63864ac7d54769b3eaab70b191d3d8f7f99c8d5dbdb6f1dad55 not found: ID does not exist" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.035744 4846 scope.go:117] "RemoveContainer" containerID="79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4" Feb 02 13:16:51 crc kubenswrapper[4846]: E0202 13:16:51.036041 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4\": container with ID starting with 79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4 not found: ID does not exist" containerID="79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.036078 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4"} err="failed to get container status \"79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4\": rpc error: code = NotFound desc = could not find container \"79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4\": container with ID starting with 79249200c25d900891f646f5b6e706e8dcfb05ee4c2c6c642ddc32fc22982ef4 not found: ID does not exist" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.036098 4846 scope.go:117] "RemoveContainer" containerID="e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86" Feb 02 13:16:51 crc kubenswrapper[4846]: E0202 13:16:51.036594 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86\": container with ID starting with e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86 not found: ID does not exist" containerID="e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.036649 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86"} err="failed to get container status \"e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86\": rpc error: code = NotFound desc = could not find container \"e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86\": container with ID starting with e28de2c302cdbbcb33ba412f4d08ecb0b160a7b9132feee34666d26323a41d86 not found: ID does not exist" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.045733 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn7qw\" (UniqueName: \"kubernetes.io/projected/76cbbc95-f545-4929-8a5a-307da82dfd81-kube-api-access-gn7qw\") on node \"crc\" DevicePath \"\"" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.045760 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.661381 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76cbbc95-f545-4929-8a5a-307da82dfd81" (UID: "76cbbc95-f545-4929-8a5a-307da82dfd81"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.756135 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76cbbc95-f545-4929-8a5a-307da82dfd81-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.843728 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mds9h"] Feb 02 13:16:51 crc kubenswrapper[4846]: I0202 13:16:51.852502 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mds9h"] Feb 02 13:16:53 crc kubenswrapper[4846]: I0202 13:16:53.449961 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76cbbc95-f545-4929-8a5a-307da82dfd81" path="/var/lib/kubelet/pods/76cbbc95-f545-4929-8a5a-307da82dfd81/volumes" Feb 02 13:17:00 crc kubenswrapper[4846]: I0202 13:17:00.479263 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:17:00 crc kubenswrapper[4846]: I0202 13:17:00.480705 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:17:30 crc kubenswrapper[4846]: I0202 13:17:30.479070 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:17:30 crc kubenswrapper[4846]: I0202 13:17:30.479894 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:17:30 crc kubenswrapper[4846]: I0202 13:17:30.479950 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 13:17:30 crc kubenswrapper[4846]: I0202 13:17:30.480561 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b74edd9a518dbcb657b4541bd37c1507460aaa8aebee2269d785534a7f1cfd6b"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 13:17:30 crc kubenswrapper[4846]: I0202 13:17:30.480929 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://b74edd9a518dbcb657b4541bd37c1507460aaa8aebee2269d785534a7f1cfd6b" gracePeriod=600 Feb 02 13:17:31 crc kubenswrapper[4846]: I0202 13:17:31.188815 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="b74edd9a518dbcb657b4541bd37c1507460aaa8aebee2269d785534a7f1cfd6b" exitCode=0 Feb 02 13:17:31 crc kubenswrapper[4846]: I0202 13:17:31.189100 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"b74edd9a518dbcb657b4541bd37c1507460aaa8aebee2269d785534a7f1cfd6b"} Feb 02 13:17:31 crc kubenswrapper[4846]: I0202 13:17:31.189125 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4"} Feb 02 13:17:31 crc kubenswrapper[4846]: I0202 13:17:31.189140 4846 scope.go:117] "RemoveContainer" containerID="0aeefc6f7caeaa53b8324dc22fa30c397cd61ea6052cb9dc4cadf26aed40dc2c" Feb 02 13:19:30 crc kubenswrapper[4846]: I0202 13:19:30.479648 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:19:30 crc kubenswrapper[4846]: I0202 13:19:30.480077 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:20:00 crc kubenswrapper[4846]: I0202 13:20:00.479072 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:20:00 crc kubenswrapper[4846]: I0202 13:20:00.479476 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:20:30 crc kubenswrapper[4846]: I0202 13:20:30.479650 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:20:30 crc kubenswrapper[4846]: I0202 13:20:30.480149 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:20:30 crc kubenswrapper[4846]: I0202 13:20:30.480195 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 13:20:30 crc kubenswrapper[4846]: I0202 13:20:30.480919 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 13:20:30 crc kubenswrapper[4846]: I0202 13:20:30.480984 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" gracePeriod=600 Feb 02 13:20:30 crc kubenswrapper[4846]: E0202 13:20:30.608427 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:20:31 crc kubenswrapper[4846]: I0202 13:20:31.477662 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" exitCode=0 Feb 02 13:20:31 crc kubenswrapper[4846]: I0202 13:20:31.477708 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4"} Feb 02 13:20:31 crc kubenswrapper[4846]: I0202 13:20:31.477736 4846 scope.go:117] "RemoveContainer" containerID="b74edd9a518dbcb657b4541bd37c1507460aaa8aebee2269d785534a7f1cfd6b" Feb 02 13:20:31 crc kubenswrapper[4846]: I0202 13:20:31.478288 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:20:31 crc kubenswrapper[4846]: E0202 13:20:31.478662 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:20:45 crc kubenswrapper[4846]: I0202 13:20:45.423996 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:20:45 crc kubenswrapper[4846]: E0202 13:20:45.424765 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:20:58 crc kubenswrapper[4846]: I0202 13:20:58.424829 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:20:58 crc kubenswrapper[4846]: E0202 13:20:58.426284 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:21:13 crc kubenswrapper[4846]: I0202 13:21:13.424572 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:21:13 crc kubenswrapper[4846]: E0202 13:21:13.425350 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:21:25 crc kubenswrapper[4846]: I0202 13:21:25.424968 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:21:25 crc kubenswrapper[4846]: E0202 13:21:25.426386 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:21:36 crc kubenswrapper[4846]: I0202 13:21:36.423793 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:21:36 crc kubenswrapper[4846]: E0202 13:21:36.425234 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.558820 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xsr5s"] Feb 02 13:21:48 crc kubenswrapper[4846]: E0202 13:21:48.560116 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerName="extract-utilities" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.560136 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerName="extract-utilities" Feb 02 13:21:48 crc kubenswrapper[4846]: E0202 13:21:48.560164 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerName="registry-server" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.560174 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerName="registry-server" Feb 02 13:21:48 crc kubenswrapper[4846]: E0202 13:21:48.560188 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerName="extract-content" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.560197 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerName="extract-content" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.560380 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="76cbbc95-f545-4929-8a5a-307da82dfd81" containerName="registry-server" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.564557 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.567653 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xsr5s"] Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.653530 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-utilities\") pod \"redhat-operators-xsr5s\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.653647 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj9q9\" (UniqueName: \"kubernetes.io/projected/724f0750-43ec-4502-b7b1-13bf275c0436-kube-api-access-bj9q9\") pod \"redhat-operators-xsr5s\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.653694 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-catalog-content\") pod \"redhat-operators-xsr5s\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.755081 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-utilities\") pod \"redhat-operators-xsr5s\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.755145 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj9q9\" (UniqueName: \"kubernetes.io/projected/724f0750-43ec-4502-b7b1-13bf275c0436-kube-api-access-bj9q9\") pod \"redhat-operators-xsr5s\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.755164 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-catalog-content\") pod \"redhat-operators-xsr5s\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.755923 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-utilities\") pod \"redhat-operators-xsr5s\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.755937 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-catalog-content\") pod \"redhat-operators-xsr5s\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.776360 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj9q9\" (UniqueName: \"kubernetes.io/projected/724f0750-43ec-4502-b7b1-13bf275c0436-kube-api-access-bj9q9\") pod \"redhat-operators-xsr5s\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:48 crc kubenswrapper[4846]: I0202 13:21:48.884339 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:49 crc kubenswrapper[4846]: I0202 13:21:49.125439 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xsr5s"] Feb 02 13:21:49 crc kubenswrapper[4846]: I0202 13:21:49.305074 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsr5s" event={"ID":"724f0750-43ec-4502-b7b1-13bf275c0436","Type":"ContainerStarted","Data":"57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d"} Feb 02 13:21:49 crc kubenswrapper[4846]: I0202 13:21:49.305475 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsr5s" event={"ID":"724f0750-43ec-4502-b7b1-13bf275c0436","Type":"ContainerStarted","Data":"1f0f20ab23a9d810d40ace87f00f667ee3fa05fa16082081af0bdb915a19d24a"} Feb 02 13:21:49 crc kubenswrapper[4846]: I0202 13:21:49.428758 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:21:49 crc kubenswrapper[4846]: E0202 13:21:49.429258 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:21:50 crc kubenswrapper[4846]: I0202 13:21:50.312384 4846 generic.go:334] "Generic (PLEG): container finished" podID="724f0750-43ec-4502-b7b1-13bf275c0436" containerID="57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d" exitCode=0 Feb 02 13:21:50 crc kubenswrapper[4846]: I0202 13:21:50.312450 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsr5s" event={"ID":"724f0750-43ec-4502-b7b1-13bf275c0436","Type":"ContainerDied","Data":"57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d"} Feb 02 13:21:50 crc kubenswrapper[4846]: I0202 13:21:50.314511 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 13:21:51 crc kubenswrapper[4846]: I0202 13:21:51.323290 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsr5s" event={"ID":"724f0750-43ec-4502-b7b1-13bf275c0436","Type":"ContainerStarted","Data":"f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2"} Feb 02 13:21:52 crc kubenswrapper[4846]: I0202 13:21:52.332538 4846 generic.go:334] "Generic (PLEG): container finished" podID="724f0750-43ec-4502-b7b1-13bf275c0436" containerID="f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2" exitCode=0 Feb 02 13:21:52 crc kubenswrapper[4846]: I0202 13:21:52.332614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsr5s" event={"ID":"724f0750-43ec-4502-b7b1-13bf275c0436","Type":"ContainerDied","Data":"f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2"} Feb 02 13:21:53 crc kubenswrapper[4846]: I0202 13:21:53.343951 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsr5s" event={"ID":"724f0750-43ec-4502-b7b1-13bf275c0436","Type":"ContainerStarted","Data":"f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040"} Feb 02 13:21:53 crc kubenswrapper[4846]: I0202 13:21:53.363696 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xsr5s" podStartSLOduration=2.945210319 podStartE2EDuration="5.363637858s" podCreationTimestamp="2026-02-02 13:21:48 +0000 UTC" firstStartedPulling="2026-02-02 13:21:50.314164023 +0000 UTC m=+4341.542750886" lastFinishedPulling="2026-02-02 13:21:52.732591562 +0000 UTC m=+4343.961178425" observedRunningTime="2026-02-02 13:21:53.359138188 +0000 UTC m=+4344.587725071" watchObservedRunningTime="2026-02-02 13:21:53.363637858 +0000 UTC m=+4344.592224721" Feb 02 13:21:58 crc kubenswrapper[4846]: I0202 13:21:58.884933 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:58 crc kubenswrapper[4846]: I0202 13:21:58.885555 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:58 crc kubenswrapper[4846]: I0202 13:21:58.935546 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:59 crc kubenswrapper[4846]: I0202 13:21:59.439052 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:21:59 crc kubenswrapper[4846]: I0202 13:21:59.493598 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xsr5s"] Feb 02 13:22:01 crc kubenswrapper[4846]: I0202 13:22:01.399338 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xsr5s" podUID="724f0750-43ec-4502-b7b1-13bf275c0436" containerName="registry-server" containerID="cri-o://f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040" gracePeriod=2 Feb 02 13:22:01 crc kubenswrapper[4846]: I0202 13:22:01.817067 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:22:01 crc kubenswrapper[4846]: I0202 13:22:01.944166 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-utilities\") pod \"724f0750-43ec-4502-b7b1-13bf275c0436\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " Feb 02 13:22:01 crc kubenswrapper[4846]: I0202 13:22:01.944290 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-catalog-content\") pod \"724f0750-43ec-4502-b7b1-13bf275c0436\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " Feb 02 13:22:01 crc kubenswrapper[4846]: I0202 13:22:01.944330 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj9q9\" (UniqueName: \"kubernetes.io/projected/724f0750-43ec-4502-b7b1-13bf275c0436-kube-api-access-bj9q9\") pod \"724f0750-43ec-4502-b7b1-13bf275c0436\" (UID: \"724f0750-43ec-4502-b7b1-13bf275c0436\") " Feb 02 13:22:01 crc kubenswrapper[4846]: I0202 13:22:01.945404 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-utilities" (OuterVolumeSpecName: "utilities") pod "724f0750-43ec-4502-b7b1-13bf275c0436" (UID: "724f0750-43ec-4502-b7b1-13bf275c0436"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:22:01 crc kubenswrapper[4846]: I0202 13:22:01.950102 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/724f0750-43ec-4502-b7b1-13bf275c0436-kube-api-access-bj9q9" (OuterVolumeSpecName: "kube-api-access-bj9q9") pod "724f0750-43ec-4502-b7b1-13bf275c0436" (UID: "724f0750-43ec-4502-b7b1-13bf275c0436"). InnerVolumeSpecName "kube-api-access-bj9q9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.045809 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj9q9\" (UniqueName: \"kubernetes.io/projected/724f0750-43ec-4502-b7b1-13bf275c0436-kube-api-access-bj9q9\") on node \"crc\" DevicePath \"\"" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.045853 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.409528 4846 generic.go:334] "Generic (PLEG): container finished" podID="724f0750-43ec-4502-b7b1-13bf275c0436" containerID="f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040" exitCode=0 Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.409584 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsr5s" event={"ID":"724f0750-43ec-4502-b7b1-13bf275c0436","Type":"ContainerDied","Data":"f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040"} Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.409624 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsr5s" event={"ID":"724f0750-43ec-4502-b7b1-13bf275c0436","Type":"ContainerDied","Data":"1f0f20ab23a9d810d40ace87f00f667ee3fa05fa16082081af0bdb915a19d24a"} Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.409668 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xsr5s" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.409672 4846 scope.go:117] "RemoveContainer" containerID="f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.428558 4846 scope.go:117] "RemoveContainer" containerID="f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.445271 4846 scope.go:117] "RemoveContainer" containerID="57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.473928 4846 scope.go:117] "RemoveContainer" containerID="f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040" Feb 02 13:22:02 crc kubenswrapper[4846]: E0202 13:22:02.474424 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040\": container with ID starting with f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040 not found: ID does not exist" containerID="f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.474481 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040"} err="failed to get container status \"f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040\": rpc error: code = NotFound desc = could not find container \"f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040\": container with ID starting with f76d2764f51b5f4a88cd76b771eb298b091484ae3675798569013546ae658040 not found: ID does not exist" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.474518 4846 scope.go:117] "RemoveContainer" containerID="f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2" Feb 02 13:22:02 crc kubenswrapper[4846]: E0202 13:22:02.475248 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2\": container with ID starting with f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2 not found: ID does not exist" containerID="f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.475359 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2"} err="failed to get container status \"f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2\": rpc error: code = NotFound desc = could not find container \"f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2\": container with ID starting with f271d4600aacf02aacfd79176bbab980994af2767ca764b2c04a46b343fc75f2 not found: ID does not exist" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.475380 4846 scope.go:117] "RemoveContainer" containerID="57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d" Feb 02 13:22:02 crc kubenswrapper[4846]: E0202 13:22:02.475732 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d\": container with ID starting with 57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d not found: ID does not exist" containerID="57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d" Feb 02 13:22:02 crc kubenswrapper[4846]: I0202 13:22:02.475760 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d"} err="failed to get container status \"57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d\": rpc error: code = NotFound desc = could not find container \"57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d\": container with ID starting with 57912de2863f5529cd56c056eac10d350f85fa84c35062120dd6d2d90f08479d not found: ID does not exist" Feb 02 13:22:03 crc kubenswrapper[4846]: I0202 13:22:03.138799 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "724f0750-43ec-4502-b7b1-13bf275c0436" (UID: "724f0750-43ec-4502-b7b1-13bf275c0436"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:22:03 crc kubenswrapper[4846]: I0202 13:22:03.162152 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/724f0750-43ec-4502-b7b1-13bf275c0436-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:22:03 crc kubenswrapper[4846]: I0202 13:22:03.356618 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xsr5s"] Feb 02 13:22:03 crc kubenswrapper[4846]: I0202 13:22:03.367477 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xsr5s"] Feb 02 13:22:03 crc kubenswrapper[4846]: I0202 13:22:03.432602 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="724f0750-43ec-4502-b7b1-13bf275c0436" path="/var/lib/kubelet/pods/724f0750-43ec-4502-b7b1-13bf275c0436/volumes" Feb 02 13:22:04 crc kubenswrapper[4846]: I0202 13:22:04.423586 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:22:04 crc kubenswrapper[4846]: E0202 13:22:04.424148 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:22:17 crc kubenswrapper[4846]: I0202 13:22:17.423948 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:22:17 crc kubenswrapper[4846]: E0202 13:22:17.424894 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:22:30 crc kubenswrapper[4846]: I0202 13:22:30.423650 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:22:30 crc kubenswrapper[4846]: E0202 13:22:30.427063 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:22:43 crc kubenswrapper[4846]: I0202 13:22:43.424257 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:22:43 crc kubenswrapper[4846]: E0202 13:22:43.425442 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:22:57 crc kubenswrapper[4846]: I0202 13:22:57.423554 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:22:57 crc kubenswrapper[4846]: E0202 13:22:57.424392 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:23:10 crc kubenswrapper[4846]: I0202 13:23:10.423985 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:23:10 crc kubenswrapper[4846]: E0202 13:23:10.424761 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:23:25 crc kubenswrapper[4846]: I0202 13:23:25.423743 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:23:25 crc kubenswrapper[4846]: E0202 13:23:25.425461 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:23:38 crc kubenswrapper[4846]: I0202 13:23:38.423412 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:23:38 crc kubenswrapper[4846]: E0202 13:23:38.424192 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:23:50 crc kubenswrapper[4846]: I0202 13:23:50.424345 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:23:50 crc kubenswrapper[4846]: E0202 13:23:50.425389 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:24:01 crc kubenswrapper[4846]: I0202 13:24:01.423985 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:24:01 crc kubenswrapper[4846]: E0202 13:24:01.424823 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:24:12 crc kubenswrapper[4846]: I0202 13:24:12.424176 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:24:12 crc kubenswrapper[4846]: E0202 13:24:12.425164 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:24:26 crc kubenswrapper[4846]: I0202 13:24:26.425213 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:24:26 crc kubenswrapper[4846]: E0202 13:24:26.426410 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:24:41 crc kubenswrapper[4846]: I0202 13:24:41.577561 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:24:41 crc kubenswrapper[4846]: E0202 13:24:41.578229 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:24:53 crc kubenswrapper[4846]: I0202 13:24:53.423959 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:24:53 crc kubenswrapper[4846]: E0202 13:24:53.425388 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.389227 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bczss"] Feb 02 13:25:01 crc kubenswrapper[4846]: E0202 13:25:01.389968 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="724f0750-43ec-4502-b7b1-13bf275c0436" containerName="registry-server" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.389980 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="724f0750-43ec-4502-b7b1-13bf275c0436" containerName="registry-server" Feb 02 13:25:01 crc kubenswrapper[4846]: E0202 13:25:01.389995 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="724f0750-43ec-4502-b7b1-13bf275c0436" containerName="extract-content" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.390000 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="724f0750-43ec-4502-b7b1-13bf275c0436" containerName="extract-content" Feb 02 13:25:01 crc kubenswrapper[4846]: E0202 13:25:01.390021 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="724f0750-43ec-4502-b7b1-13bf275c0436" containerName="extract-utilities" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.390027 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="724f0750-43ec-4502-b7b1-13bf275c0436" containerName="extract-utilities" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.390172 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="724f0750-43ec-4502-b7b1-13bf275c0436" containerName="registry-server" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.391378 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.398908 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bczss"] Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.593736 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcpd4\" (UniqueName: \"kubernetes.io/projected/e66b9e56-77b8-4020-8bf5-1885765c3c72-kube-api-access-gcpd4\") pod \"certified-operators-bczss\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.593771 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-catalog-content\") pod \"certified-operators-bczss\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.594533 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-utilities\") pod \"certified-operators-bczss\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.695010 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcpd4\" (UniqueName: \"kubernetes.io/projected/e66b9e56-77b8-4020-8bf5-1885765c3c72-kube-api-access-gcpd4\") pod \"certified-operators-bczss\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.695056 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-catalog-content\") pod \"certified-operators-bczss\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.695124 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-utilities\") pod \"certified-operators-bczss\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.695587 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-utilities\") pod \"certified-operators-bczss\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.695695 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-catalog-content\") pod \"certified-operators-bczss\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.718001 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcpd4\" (UniqueName: \"kubernetes.io/projected/e66b9e56-77b8-4020-8bf5-1885765c3c72-kube-api-access-gcpd4\") pod \"certified-operators-bczss\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:01 crc kubenswrapper[4846]: I0202 13:25:01.720650 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:02 crc kubenswrapper[4846]: I0202 13:25:02.192970 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bczss"] Feb 02 13:25:02 crc kubenswrapper[4846]: I0202 13:25:02.740320 4846 generic.go:334] "Generic (PLEG): container finished" podID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerID="1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa" exitCode=0 Feb 02 13:25:02 crc kubenswrapper[4846]: I0202 13:25:02.740433 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bczss" event={"ID":"e66b9e56-77b8-4020-8bf5-1885765c3c72","Type":"ContainerDied","Data":"1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa"} Feb 02 13:25:02 crc kubenswrapper[4846]: I0202 13:25:02.740633 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bczss" event={"ID":"e66b9e56-77b8-4020-8bf5-1885765c3c72","Type":"ContainerStarted","Data":"8f0539ed25b41f86bdf28c075ca014e70d362ce0477b3fcd9a8e3df8559f1a52"} Feb 02 13:25:03 crc kubenswrapper[4846]: I0202 13:25:03.747409 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bczss" event={"ID":"e66b9e56-77b8-4020-8bf5-1885765c3c72","Type":"ContainerStarted","Data":"591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200"} Feb 02 13:25:04 crc kubenswrapper[4846]: I0202 13:25:04.757130 4846 generic.go:334] "Generic (PLEG): container finished" podID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerID="591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200" exitCode=0 Feb 02 13:25:04 crc kubenswrapper[4846]: I0202 13:25:04.757179 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bczss" event={"ID":"e66b9e56-77b8-4020-8bf5-1885765c3c72","Type":"ContainerDied","Data":"591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200"} Feb 02 13:25:05 crc kubenswrapper[4846]: I0202 13:25:05.767309 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bczss" event={"ID":"e66b9e56-77b8-4020-8bf5-1885765c3c72","Type":"ContainerStarted","Data":"2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f"} Feb 02 13:25:06 crc kubenswrapper[4846]: I0202 13:25:06.423774 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:25:06 crc kubenswrapper[4846]: E0202 13:25:06.424036 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:25:11 crc kubenswrapper[4846]: I0202 13:25:11.721235 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:11 crc kubenswrapper[4846]: I0202 13:25:11.721785 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:11 crc kubenswrapper[4846]: I0202 13:25:11.766764 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:11 crc kubenswrapper[4846]: I0202 13:25:11.787113 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bczss" podStartSLOduration=8.312421216 podStartE2EDuration="10.787097001s" podCreationTimestamp="2026-02-02 13:25:01 +0000 UTC" firstStartedPulling="2026-02-02 13:25:02.741685263 +0000 UTC m=+4533.970272126" lastFinishedPulling="2026-02-02 13:25:05.216361048 +0000 UTC m=+4536.444947911" observedRunningTime="2026-02-02 13:25:05.806122274 +0000 UTC m=+4537.034709137" watchObservedRunningTime="2026-02-02 13:25:11.787097001 +0000 UTC m=+4543.015683864" Feb 02 13:25:11 crc kubenswrapper[4846]: I0202 13:25:11.841842 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:12 crc kubenswrapper[4846]: I0202 13:25:12.003021 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bczss"] Feb 02 13:25:13 crc kubenswrapper[4846]: I0202 13:25:13.819165 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bczss" podUID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerName="registry-server" containerID="cri-o://2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f" gracePeriod=2 Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.247087 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.392004 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcpd4\" (UniqueName: \"kubernetes.io/projected/e66b9e56-77b8-4020-8bf5-1885765c3c72-kube-api-access-gcpd4\") pod \"e66b9e56-77b8-4020-8bf5-1885765c3c72\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.392105 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-utilities\") pod \"e66b9e56-77b8-4020-8bf5-1885765c3c72\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.392261 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-catalog-content\") pod \"e66b9e56-77b8-4020-8bf5-1885765c3c72\" (UID: \"e66b9e56-77b8-4020-8bf5-1885765c3c72\") " Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.393119 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-utilities" (OuterVolumeSpecName: "utilities") pod "e66b9e56-77b8-4020-8bf5-1885765c3c72" (UID: "e66b9e56-77b8-4020-8bf5-1885765c3c72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.397314 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e66b9e56-77b8-4020-8bf5-1885765c3c72-kube-api-access-gcpd4" (OuterVolumeSpecName: "kube-api-access-gcpd4") pod "e66b9e56-77b8-4020-8bf5-1885765c3c72" (UID: "e66b9e56-77b8-4020-8bf5-1885765c3c72"). InnerVolumeSpecName "kube-api-access-gcpd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.436996 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e66b9e56-77b8-4020-8bf5-1885765c3c72" (UID: "e66b9e56-77b8-4020-8bf5-1885765c3c72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.493714 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.493750 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e66b9e56-77b8-4020-8bf5-1885765c3c72-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.493763 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcpd4\" (UniqueName: \"kubernetes.io/projected/e66b9e56-77b8-4020-8bf5-1885765c3c72-kube-api-access-gcpd4\") on node \"crc\" DevicePath \"\"" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.827505 4846 generic.go:334] "Generic (PLEG): container finished" podID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerID="2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f" exitCode=0 Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.827556 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bczss" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.827556 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bczss" event={"ID":"e66b9e56-77b8-4020-8bf5-1885765c3c72","Type":"ContainerDied","Data":"2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f"} Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.827664 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bczss" event={"ID":"e66b9e56-77b8-4020-8bf5-1885765c3c72","Type":"ContainerDied","Data":"8f0539ed25b41f86bdf28c075ca014e70d362ce0477b3fcd9a8e3df8559f1a52"} Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.827686 4846 scope.go:117] "RemoveContainer" containerID="2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.850022 4846 scope.go:117] "RemoveContainer" containerID="591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.886697 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bczss"] Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.898752 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bczss"] Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.920947 4846 scope.go:117] "RemoveContainer" containerID="1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.972818 4846 scope.go:117] "RemoveContainer" containerID="2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f" Feb 02 13:25:14 crc kubenswrapper[4846]: E0202 13:25:14.973491 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f\": container with ID starting with 2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f not found: ID does not exist" containerID="2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.973536 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f"} err="failed to get container status \"2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f\": rpc error: code = NotFound desc = could not find container \"2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f\": container with ID starting with 2978eb66ba7af98b46d74d1bb885dbcc2ff3bbf824ca9c9a628444c4f72c5d0f not found: ID does not exist" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.973583 4846 scope.go:117] "RemoveContainer" containerID="591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200" Feb 02 13:25:14 crc kubenswrapper[4846]: E0202 13:25:14.977030 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200\": container with ID starting with 591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200 not found: ID does not exist" containerID="591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.977080 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200"} err="failed to get container status \"591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200\": rpc error: code = NotFound desc = could not find container \"591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200\": container with ID starting with 591d1808c689f271f554f9323e0f6dd71fd30d5dacb423540b5e042a49211200 not found: ID does not exist" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.977114 4846 scope.go:117] "RemoveContainer" containerID="1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa" Feb 02 13:25:14 crc kubenswrapper[4846]: E0202 13:25:14.977821 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa\": container with ID starting with 1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa not found: ID does not exist" containerID="1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa" Feb 02 13:25:14 crc kubenswrapper[4846]: I0202 13:25:14.977894 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa"} err="failed to get container status \"1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa\": rpc error: code = NotFound desc = could not find container \"1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa\": container with ID starting with 1151487b8bfd0de73b1cc7a773f66d5af92d4859a978b86571ba7aeb1393e3aa not found: ID does not exist" Feb 02 13:25:15 crc kubenswrapper[4846]: I0202 13:25:15.433590 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e66b9e56-77b8-4020-8bf5-1885765c3c72" path="/var/lib/kubelet/pods/e66b9e56-77b8-4020-8bf5-1885765c3c72/volumes" Feb 02 13:25:20 crc kubenswrapper[4846]: I0202 13:25:20.424100 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:25:20 crc kubenswrapper[4846]: E0202 13:25:20.424916 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:25:34 crc kubenswrapper[4846]: I0202 13:25:34.422942 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:25:34 crc kubenswrapper[4846]: I0202 13:25:34.967000 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"118141657a51d573c44958b377ad8411707adb515a180c423602a0ce85f6f046"} Feb 02 13:28:00 crc kubenswrapper[4846]: I0202 13:28:00.478599 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:28:00 crc kubenswrapper[4846]: I0202 13:28:00.479293 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:28:02 crc kubenswrapper[4846]: I0202 13:28:02.826652 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c9c26"] Feb 02 13:28:02 crc kubenswrapper[4846]: E0202 13:28:02.827582 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerName="extract-utilities" Feb 02 13:28:02 crc kubenswrapper[4846]: I0202 13:28:02.827615 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerName="extract-utilities" Feb 02 13:28:02 crc kubenswrapper[4846]: E0202 13:28:02.827680 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerName="registry-server" Feb 02 13:28:02 crc kubenswrapper[4846]: I0202 13:28:02.827697 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerName="registry-server" Feb 02 13:28:02 crc kubenswrapper[4846]: E0202 13:28:02.827741 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerName="extract-content" Feb 02 13:28:02 crc kubenswrapper[4846]: I0202 13:28:02.827783 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerName="extract-content" Feb 02 13:28:02 crc kubenswrapper[4846]: I0202 13:28:02.828131 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e66b9e56-77b8-4020-8bf5-1885765c3c72" containerName="registry-server" Feb 02 13:28:02 crc kubenswrapper[4846]: I0202 13:28:02.830579 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:02 crc kubenswrapper[4846]: I0202 13:28:02.837702 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c9c26"] Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.016978 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l96mr\" (UniqueName: \"kubernetes.io/projected/adff2d32-6dbc-44eb-9354-2b7465990677-kube-api-access-l96mr\") pod \"community-operators-c9c26\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.017102 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-catalog-content\") pod \"community-operators-c9c26\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.017222 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-utilities\") pod \"community-operators-c9c26\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.118988 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l96mr\" (UniqueName: \"kubernetes.io/projected/adff2d32-6dbc-44eb-9354-2b7465990677-kube-api-access-l96mr\") pod \"community-operators-c9c26\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.119073 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-catalog-content\") pod \"community-operators-c9c26\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.119131 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-utilities\") pod \"community-operators-c9c26\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.119655 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-catalog-content\") pod \"community-operators-c9c26\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.119785 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-utilities\") pod \"community-operators-c9c26\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.150588 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l96mr\" (UniqueName: \"kubernetes.io/projected/adff2d32-6dbc-44eb-9354-2b7465990677-kube-api-access-l96mr\") pod \"community-operators-c9c26\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.189369 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:03 crc kubenswrapper[4846]: I0202 13:28:03.704509 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c9c26"] Feb 02 13:28:04 crc kubenswrapper[4846]: I0202 13:28:04.085224 4846 generic.go:334] "Generic (PLEG): container finished" podID="adff2d32-6dbc-44eb-9354-2b7465990677" containerID="d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879" exitCode=0 Feb 02 13:28:04 crc kubenswrapper[4846]: I0202 13:28:04.085341 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9c26" event={"ID":"adff2d32-6dbc-44eb-9354-2b7465990677","Type":"ContainerDied","Data":"d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879"} Feb 02 13:28:04 crc kubenswrapper[4846]: I0202 13:28:04.085670 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9c26" event={"ID":"adff2d32-6dbc-44eb-9354-2b7465990677","Type":"ContainerStarted","Data":"4b2b6e08054fb6bd437ab2e81f4dc2b74078c3c69dd2d32cf2ac3dcd9e26286a"} Feb 02 13:28:04 crc kubenswrapper[4846]: I0202 13:28:04.087717 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.223459 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f67mn"] Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.225835 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.248254 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f67mn"] Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.350882 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-catalog-content\") pod \"redhat-marketplace-f67mn\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.350961 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-utilities\") pod \"redhat-marketplace-f67mn\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.350991 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r74qv\" (UniqueName: \"kubernetes.io/projected/e552b9d5-b185-41ac-a3a7-740e4895833d-kube-api-access-r74qv\") pod \"redhat-marketplace-f67mn\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.452105 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-catalog-content\") pod \"redhat-marketplace-f67mn\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.452547 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-utilities\") pod \"redhat-marketplace-f67mn\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.452708 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r74qv\" (UniqueName: \"kubernetes.io/projected/e552b9d5-b185-41ac-a3a7-740e4895833d-kube-api-access-r74qv\") pod \"redhat-marketplace-f67mn\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.454082 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-catalog-content\") pod \"redhat-marketplace-f67mn\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.454152 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-utilities\") pod \"redhat-marketplace-f67mn\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.478128 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r74qv\" (UniqueName: \"kubernetes.io/projected/e552b9d5-b185-41ac-a3a7-740e4895833d-kube-api-access-r74qv\") pod \"redhat-marketplace-f67mn\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.543532 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:05 crc kubenswrapper[4846]: I0202 13:28:05.972249 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f67mn"] Feb 02 13:28:06 crc kubenswrapper[4846]: I0202 13:28:06.098790 4846 generic.go:334] "Generic (PLEG): container finished" podID="adff2d32-6dbc-44eb-9354-2b7465990677" containerID="036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4" exitCode=0 Feb 02 13:28:06 crc kubenswrapper[4846]: I0202 13:28:06.098900 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9c26" event={"ID":"adff2d32-6dbc-44eb-9354-2b7465990677","Type":"ContainerDied","Data":"036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4"} Feb 02 13:28:06 crc kubenswrapper[4846]: I0202 13:28:06.100080 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f67mn" event={"ID":"e552b9d5-b185-41ac-a3a7-740e4895833d","Type":"ContainerStarted","Data":"ebb5e6f4c630c9408f9af24598920bc98c0e680b929faaf9b53a03154063fb68"} Feb 02 13:28:07 crc kubenswrapper[4846]: I0202 13:28:07.108601 4846 generic.go:334] "Generic (PLEG): container finished" podID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerID="6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc" exitCode=0 Feb 02 13:28:07 crc kubenswrapper[4846]: I0202 13:28:07.108666 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f67mn" event={"ID":"e552b9d5-b185-41ac-a3a7-740e4895833d","Type":"ContainerDied","Data":"6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc"} Feb 02 13:28:07 crc kubenswrapper[4846]: I0202 13:28:07.111792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9c26" event={"ID":"adff2d32-6dbc-44eb-9354-2b7465990677","Type":"ContainerStarted","Data":"565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd"} Feb 02 13:28:08 crc kubenswrapper[4846]: I0202 13:28:08.120018 4846 generic.go:334] "Generic (PLEG): container finished" podID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerID="6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8" exitCode=0 Feb 02 13:28:08 crc kubenswrapper[4846]: I0202 13:28:08.120118 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f67mn" event={"ID":"e552b9d5-b185-41ac-a3a7-740e4895833d","Type":"ContainerDied","Data":"6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8"} Feb 02 13:28:08 crc kubenswrapper[4846]: I0202 13:28:08.145479 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c9c26" podStartSLOduration=3.642305252 podStartE2EDuration="6.145454983s" podCreationTimestamp="2026-02-02 13:28:02 +0000 UTC" firstStartedPulling="2026-02-02 13:28:04.087499905 +0000 UTC m=+4715.316086758" lastFinishedPulling="2026-02-02 13:28:06.590649626 +0000 UTC m=+4717.819236489" observedRunningTime="2026-02-02 13:28:07.153951975 +0000 UTC m=+4718.382538838" watchObservedRunningTime="2026-02-02 13:28:08.145454983 +0000 UTC m=+4719.374041856" Feb 02 13:28:09 crc kubenswrapper[4846]: I0202 13:28:09.129466 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f67mn" event={"ID":"e552b9d5-b185-41ac-a3a7-740e4895833d","Type":"ContainerStarted","Data":"816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f"} Feb 02 13:28:09 crc kubenswrapper[4846]: I0202 13:28:09.156477 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f67mn" podStartSLOduration=2.739249603 podStartE2EDuration="4.156452377s" podCreationTimestamp="2026-02-02 13:28:05 +0000 UTC" firstStartedPulling="2026-02-02 13:28:07.110159935 +0000 UTC m=+4718.338746798" lastFinishedPulling="2026-02-02 13:28:08.527362709 +0000 UTC m=+4719.755949572" observedRunningTime="2026-02-02 13:28:09.147346535 +0000 UTC m=+4720.375933408" watchObservedRunningTime="2026-02-02 13:28:09.156452377 +0000 UTC m=+4720.385039260" Feb 02 13:28:13 crc kubenswrapper[4846]: I0202 13:28:13.189834 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:13 crc kubenswrapper[4846]: I0202 13:28:13.190845 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:13 crc kubenswrapper[4846]: I0202 13:28:13.255065 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:14 crc kubenswrapper[4846]: I0202 13:28:14.215247 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:14 crc kubenswrapper[4846]: I0202 13:28:14.253162 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c9c26"] Feb 02 13:28:15 crc kubenswrapper[4846]: I0202 13:28:15.543780 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:15 crc kubenswrapper[4846]: I0202 13:28:15.544525 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:15 crc kubenswrapper[4846]: I0202 13:28:15.587182 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.181404 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c9c26" podUID="adff2d32-6dbc-44eb-9354-2b7465990677" containerName="registry-server" containerID="cri-o://565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd" gracePeriod=2 Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.226396 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.611727 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.715864 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-utilities\") pod \"adff2d32-6dbc-44eb-9354-2b7465990677\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.715914 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-catalog-content\") pod \"adff2d32-6dbc-44eb-9354-2b7465990677\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.715961 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l96mr\" (UniqueName: \"kubernetes.io/projected/adff2d32-6dbc-44eb-9354-2b7465990677-kube-api-access-l96mr\") pod \"adff2d32-6dbc-44eb-9354-2b7465990677\" (UID: \"adff2d32-6dbc-44eb-9354-2b7465990677\") " Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.717879 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-utilities" (OuterVolumeSpecName: "utilities") pod "adff2d32-6dbc-44eb-9354-2b7465990677" (UID: "adff2d32-6dbc-44eb-9354-2b7465990677"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.733042 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adff2d32-6dbc-44eb-9354-2b7465990677-kube-api-access-l96mr" (OuterVolumeSpecName: "kube-api-access-l96mr") pod "adff2d32-6dbc-44eb-9354-2b7465990677" (UID: "adff2d32-6dbc-44eb-9354-2b7465990677"). InnerVolumeSpecName "kube-api-access-l96mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.772127 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adff2d32-6dbc-44eb-9354-2b7465990677" (UID: "adff2d32-6dbc-44eb-9354-2b7465990677"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.817930 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l96mr\" (UniqueName: \"kubernetes.io/projected/adff2d32-6dbc-44eb-9354-2b7465990677-kube-api-access-l96mr\") on node \"crc\" DevicePath \"\"" Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.817980 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:28:16 crc kubenswrapper[4846]: I0202 13:28:16.817994 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adff2d32-6dbc-44eb-9354-2b7465990677-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.088087 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f67mn"] Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.189135 4846 generic.go:334] "Generic (PLEG): container finished" podID="adff2d32-6dbc-44eb-9354-2b7465990677" containerID="565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd" exitCode=0 Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.189194 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9c26" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.189244 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9c26" event={"ID":"adff2d32-6dbc-44eb-9354-2b7465990677","Type":"ContainerDied","Data":"565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd"} Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.189290 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9c26" event={"ID":"adff2d32-6dbc-44eb-9354-2b7465990677","Type":"ContainerDied","Data":"4b2b6e08054fb6bd437ab2e81f4dc2b74078c3c69dd2d32cf2ac3dcd9e26286a"} Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.189312 4846 scope.go:117] "RemoveContainer" containerID="565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.206707 4846 scope.go:117] "RemoveContainer" containerID="036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.222927 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c9c26"] Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.229427 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c9c26"] Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.231388 4846 scope.go:117] "RemoveContainer" containerID="d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.249799 4846 scope.go:117] "RemoveContainer" containerID="565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd" Feb 02 13:28:17 crc kubenswrapper[4846]: E0202 13:28:17.250172 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd\": container with ID starting with 565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd not found: ID does not exist" containerID="565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.250221 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd"} err="failed to get container status \"565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd\": rpc error: code = NotFound desc = could not find container \"565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd\": container with ID starting with 565dad78b98c80ffcfa2e51131deff452f946501d2e15857b64ef0cdc6cd19cd not found: ID does not exist" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.250256 4846 scope.go:117] "RemoveContainer" containerID="036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4" Feb 02 13:28:17 crc kubenswrapper[4846]: E0202 13:28:17.250787 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4\": container with ID starting with 036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4 not found: ID does not exist" containerID="036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.250814 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4"} err="failed to get container status \"036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4\": rpc error: code = NotFound desc = could not find container \"036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4\": container with ID starting with 036f7eb17671e30fa8bbdfff414cacadca20fb50d981aa8747342a11ae8153d4 not found: ID does not exist" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.250832 4846 scope.go:117] "RemoveContainer" containerID="d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879" Feb 02 13:28:17 crc kubenswrapper[4846]: E0202 13:28:17.251036 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879\": container with ID starting with d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879 not found: ID does not exist" containerID="d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.251060 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879"} err="failed to get container status \"d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879\": rpc error: code = NotFound desc = could not find container \"d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879\": container with ID starting with d0d1976699630103b535ffb43af40509e91780bdb3167000f388f3ee47d3e879 not found: ID does not exist" Feb 02 13:28:17 crc kubenswrapper[4846]: I0202 13:28:17.447456 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adff2d32-6dbc-44eb-9354-2b7465990677" path="/var/lib/kubelet/pods/adff2d32-6dbc-44eb-9354-2b7465990677/volumes" Feb 02 13:28:18 crc kubenswrapper[4846]: I0202 13:28:18.196945 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f67mn" podUID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerName="registry-server" containerID="cri-o://816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f" gracePeriod=2 Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.177550 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.205593 4846 generic.go:334] "Generic (PLEG): container finished" podID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerID="816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f" exitCode=0 Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.205647 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f67mn" event={"ID":"e552b9d5-b185-41ac-a3a7-740e4895833d","Type":"ContainerDied","Data":"816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f"} Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.205669 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f67mn" event={"ID":"e552b9d5-b185-41ac-a3a7-740e4895833d","Type":"ContainerDied","Data":"ebb5e6f4c630c9408f9af24598920bc98c0e680b929faaf9b53a03154063fb68"} Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.205685 4846 scope.go:117] "RemoveContainer" containerID="816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.205770 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f67mn" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.224925 4846 scope.go:117] "RemoveContainer" containerID="6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.254230 4846 scope.go:117] "RemoveContainer" containerID="6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.274195 4846 scope.go:117] "RemoveContainer" containerID="816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f" Feb 02 13:28:19 crc kubenswrapper[4846]: E0202 13:28:19.274752 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f\": container with ID starting with 816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f not found: ID does not exist" containerID="816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.274791 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f"} err="failed to get container status \"816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f\": rpc error: code = NotFound desc = could not find container \"816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f\": container with ID starting with 816db186362d4a715405a0e66c8f71d46694cfd6d2fcc40c3ffdc543430b781f not found: ID does not exist" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.274817 4846 scope.go:117] "RemoveContainer" containerID="6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8" Feb 02 13:28:19 crc kubenswrapper[4846]: E0202 13:28:19.275204 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8\": container with ID starting with 6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8 not found: ID does not exist" containerID="6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.275234 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8"} err="failed to get container status \"6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8\": rpc error: code = NotFound desc = could not find container \"6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8\": container with ID starting with 6ebde9bf1d34898d471e781ea89f3995e645100b634ad8b78f26bfd3b646e8d8 not found: ID does not exist" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.275255 4846 scope.go:117] "RemoveContainer" containerID="6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc" Feb 02 13:28:19 crc kubenswrapper[4846]: E0202 13:28:19.275570 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc\": container with ID starting with 6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc not found: ID does not exist" containerID="6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.275586 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc"} err="failed to get container status \"6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc\": rpc error: code = NotFound desc = could not find container \"6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc\": container with ID starting with 6a1e06ede11f0da3e0ef79c2aa5e2d8058865dc042bd4c58a1a3253f2dfc2edc not found: ID does not exist" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.351339 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-utilities\") pod \"e552b9d5-b185-41ac-a3a7-740e4895833d\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.351439 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-catalog-content\") pod \"e552b9d5-b185-41ac-a3a7-740e4895833d\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.351466 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r74qv\" (UniqueName: \"kubernetes.io/projected/e552b9d5-b185-41ac-a3a7-740e4895833d-kube-api-access-r74qv\") pod \"e552b9d5-b185-41ac-a3a7-740e4895833d\" (UID: \"e552b9d5-b185-41ac-a3a7-740e4895833d\") " Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.356748 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e552b9d5-b185-41ac-a3a7-740e4895833d-kube-api-access-r74qv" (OuterVolumeSpecName: "kube-api-access-r74qv") pod "e552b9d5-b185-41ac-a3a7-740e4895833d" (UID: "e552b9d5-b185-41ac-a3a7-740e4895833d"). InnerVolumeSpecName "kube-api-access-r74qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.368868 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-utilities" (OuterVolumeSpecName: "utilities") pod "e552b9d5-b185-41ac-a3a7-740e4895833d" (UID: "e552b9d5-b185-41ac-a3a7-740e4895833d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.376735 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e552b9d5-b185-41ac-a3a7-740e4895833d" (UID: "e552b9d5-b185-41ac-a3a7-740e4895833d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.454451 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.454836 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e552b9d5-b185-41ac-a3a7-740e4895833d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.454914 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r74qv\" (UniqueName: \"kubernetes.io/projected/e552b9d5-b185-41ac-a3a7-740e4895833d-kube-api-access-r74qv\") on node \"crc\" DevicePath \"\"" Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.768761 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f67mn"] Feb 02 13:28:19 crc kubenswrapper[4846]: I0202 13:28:19.774258 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f67mn"] Feb 02 13:28:21 crc kubenswrapper[4846]: I0202 13:28:21.432175 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e552b9d5-b185-41ac-a3a7-740e4895833d" path="/var/lib/kubelet/pods/e552b9d5-b185-41ac-a3a7-740e4895833d/volumes" Feb 02 13:28:30 crc kubenswrapper[4846]: I0202 13:28:30.478913 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:28:30 crc kubenswrapper[4846]: I0202 13:28:30.479410 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:29:00 crc kubenswrapper[4846]: I0202 13:29:00.478835 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:29:00 crc kubenswrapper[4846]: I0202 13:29:00.479413 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:29:00 crc kubenswrapper[4846]: I0202 13:29:00.479468 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 13:29:00 crc kubenswrapper[4846]: I0202 13:29:00.480134 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"118141657a51d573c44958b377ad8411707adb515a180c423602a0ce85f6f046"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 13:29:00 crc kubenswrapper[4846]: I0202 13:29:00.480201 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://118141657a51d573c44958b377ad8411707adb515a180c423602a0ce85f6f046" gracePeriod=600 Feb 02 13:29:00 crc kubenswrapper[4846]: I0202 13:29:00.802236 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="118141657a51d573c44958b377ad8411707adb515a180c423602a0ce85f6f046" exitCode=0 Feb 02 13:29:00 crc kubenswrapper[4846]: I0202 13:29:00.802329 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"118141657a51d573c44958b377ad8411707adb515a180c423602a0ce85f6f046"} Feb 02 13:29:00 crc kubenswrapper[4846]: I0202 13:29:00.802571 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5"} Feb 02 13:29:00 crc kubenswrapper[4846]: I0202 13:29:00.802593 4846 scope.go:117] "RemoveContainer" containerID="d7de587d6ebfd210bdee7abd295e4e3add23f6fdb229ca1fe3e660b5b698eab4" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.799341 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-md5nm"] Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.806504 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-md5nm"] Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.910305 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-7m72z"] Feb 02 13:29:24 crc kubenswrapper[4846]: E0202 13:29:24.910591 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adff2d32-6dbc-44eb-9354-2b7465990677" containerName="extract-content" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.910608 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="adff2d32-6dbc-44eb-9354-2b7465990677" containerName="extract-content" Feb 02 13:29:24 crc kubenswrapper[4846]: E0202 13:29:24.910650 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerName="extract-content" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.910657 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerName="extract-content" Feb 02 13:29:24 crc kubenswrapper[4846]: E0202 13:29:24.910672 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerName="extract-utilities" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.910680 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerName="extract-utilities" Feb 02 13:29:24 crc kubenswrapper[4846]: E0202 13:29:24.910698 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerName="registry-server" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.910706 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerName="registry-server" Feb 02 13:29:24 crc kubenswrapper[4846]: E0202 13:29:24.910717 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adff2d32-6dbc-44eb-9354-2b7465990677" containerName="extract-utilities" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.910724 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="adff2d32-6dbc-44eb-9354-2b7465990677" containerName="extract-utilities" Feb 02 13:29:24 crc kubenswrapper[4846]: E0202 13:29:24.910739 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adff2d32-6dbc-44eb-9354-2b7465990677" containerName="registry-server" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.910746 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="adff2d32-6dbc-44eb-9354-2b7465990677" containerName="registry-server" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.910911 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e552b9d5-b185-41ac-a3a7-740e4895833d" containerName="registry-server" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.910936 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="adff2d32-6dbc-44eb-9354-2b7465990677" containerName="registry-server" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.911449 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.914975 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.916604 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.916865 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.917590 4846 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-l8clt" Feb 02 13:29:24 crc kubenswrapper[4846]: I0202 13:29:24.927674 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-7m72z"] Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.075545 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-crc-storage\") pod \"crc-storage-crc-7m72z\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.075659 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-node-mnt\") pod \"crc-storage-crc-7m72z\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.075726 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lf4x\" (UniqueName: \"kubernetes.io/projected/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-kube-api-access-5lf4x\") pod \"crc-storage-crc-7m72z\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.177087 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-node-mnt\") pod \"crc-storage-crc-7m72z\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.177165 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lf4x\" (UniqueName: \"kubernetes.io/projected/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-kube-api-access-5lf4x\") pod \"crc-storage-crc-7m72z\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.177234 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-crc-storage\") pod \"crc-storage-crc-7m72z\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.177493 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-node-mnt\") pod \"crc-storage-crc-7m72z\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.178229 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-crc-storage\") pod \"crc-storage-crc-7m72z\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.205500 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lf4x\" (UniqueName: \"kubernetes.io/projected/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-kube-api-access-5lf4x\") pod \"crc-storage-crc-7m72z\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.237140 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.438117 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4686b9f2-68b9-4de6-a3ce-aa95d6730abe" path="/var/lib/kubelet/pods/4686b9f2-68b9-4de6-a3ce-aa95d6730abe/volumes" Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.697477 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-7m72z"] Feb 02 13:29:25 crc kubenswrapper[4846]: I0202 13:29:25.983657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-7m72z" event={"ID":"8fee6f85-c7a1-4912-8852-f62aba2b4c6a","Type":"ContainerStarted","Data":"ee70b38d2b9a6469833106ab5ce1c4fce209b5d17cdc465fe9a25edef0e33770"} Feb 02 13:29:26 crc kubenswrapper[4846]: I0202 13:29:26.997232 4846 generic.go:334] "Generic (PLEG): container finished" podID="8fee6f85-c7a1-4912-8852-f62aba2b4c6a" containerID="d3dfe31162f6806d3acc550e0d36502902440b56801f3a2f311ce6d90f93062c" exitCode=0 Feb 02 13:29:26 crc kubenswrapper[4846]: I0202 13:29:26.997365 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-7m72z" event={"ID":"8fee6f85-c7a1-4912-8852-f62aba2b4c6a","Type":"ContainerDied","Data":"d3dfe31162f6806d3acc550e0d36502902440b56801f3a2f311ce6d90f93062c"} Feb 02 13:29:28 crc kubenswrapper[4846]: I0202 13:29:28.289001 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:28 crc kubenswrapper[4846]: I0202 13:29:28.424751 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-crc-storage\") pod \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " Feb 02 13:29:28 crc kubenswrapper[4846]: I0202 13:29:28.424815 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lf4x\" (UniqueName: \"kubernetes.io/projected/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-kube-api-access-5lf4x\") pod \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " Feb 02 13:29:28 crc kubenswrapper[4846]: I0202 13:29:28.424967 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-node-mnt\") pod \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\" (UID: \"8fee6f85-c7a1-4912-8852-f62aba2b4c6a\") " Feb 02 13:29:28 crc kubenswrapper[4846]: I0202 13:29:28.425137 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "8fee6f85-c7a1-4912-8852-f62aba2b4c6a" (UID: "8fee6f85-c7a1-4912-8852-f62aba2b4c6a"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:29:28 crc kubenswrapper[4846]: I0202 13:29:28.425403 4846 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-node-mnt\") on node \"crc\" DevicePath \"\"" Feb 02 13:29:28 crc kubenswrapper[4846]: I0202 13:29:28.434868 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-kube-api-access-5lf4x" (OuterVolumeSpecName: "kube-api-access-5lf4x") pod "8fee6f85-c7a1-4912-8852-f62aba2b4c6a" (UID: "8fee6f85-c7a1-4912-8852-f62aba2b4c6a"). InnerVolumeSpecName "kube-api-access-5lf4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:29:28 crc kubenswrapper[4846]: I0202 13:29:28.452666 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "8fee6f85-c7a1-4912-8852-f62aba2b4c6a" (UID: "8fee6f85-c7a1-4912-8852-f62aba2b4c6a"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:29:28 crc kubenswrapper[4846]: I0202 13:29:28.527019 4846 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-crc-storage\") on node \"crc\" DevicePath \"\"" Feb 02 13:29:28 crc kubenswrapper[4846]: I0202 13:29:28.527070 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lf4x\" (UniqueName: \"kubernetes.io/projected/8fee6f85-c7a1-4912-8852-f62aba2b4c6a-kube-api-access-5lf4x\") on node \"crc\" DevicePath \"\"" Feb 02 13:29:29 crc kubenswrapper[4846]: I0202 13:29:29.014887 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-7m72z" event={"ID":"8fee6f85-c7a1-4912-8852-f62aba2b4c6a","Type":"ContainerDied","Data":"ee70b38d2b9a6469833106ab5ce1c4fce209b5d17cdc465fe9a25edef0e33770"} Feb 02 13:29:29 crc kubenswrapper[4846]: I0202 13:29:29.015184 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee70b38d2b9a6469833106ab5ce1c4fce209b5d17cdc465fe9a25edef0e33770" Feb 02 13:29:29 crc kubenswrapper[4846]: I0202 13:29:29.014942 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7m72z" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.481966 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-7m72z"] Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.487342 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-7m72z"] Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.628061 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-7drv9"] Feb 02 13:29:30 crc kubenswrapper[4846]: E0202 13:29:30.628339 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fee6f85-c7a1-4912-8852-f62aba2b4c6a" containerName="storage" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.628353 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fee6f85-c7a1-4912-8852-f62aba2b4c6a" containerName="storage" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.628496 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fee6f85-c7a1-4912-8852-f62aba2b4c6a" containerName="storage" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.628946 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.631230 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.631265 4846 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-l8clt" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.631331 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.631436 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.642915 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-7drv9"] Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.754910 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dc05aaca-8ca7-4148-88d2-a0a613e02edb-crc-storage\") pod \"crc-storage-crc-7drv9\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.755284 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dc05aaca-8ca7-4148-88d2-a0a613e02edb-node-mnt\") pod \"crc-storage-crc-7drv9\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.755417 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t48sr\" (UniqueName: \"kubernetes.io/projected/dc05aaca-8ca7-4148-88d2-a0a613e02edb-kube-api-access-t48sr\") pod \"crc-storage-crc-7drv9\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.856924 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dc05aaca-8ca7-4148-88d2-a0a613e02edb-crc-storage\") pod \"crc-storage-crc-7drv9\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.857281 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dc05aaca-8ca7-4148-88d2-a0a613e02edb-node-mnt\") pod \"crc-storage-crc-7drv9\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.857423 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t48sr\" (UniqueName: \"kubernetes.io/projected/dc05aaca-8ca7-4148-88d2-a0a613e02edb-kube-api-access-t48sr\") pod \"crc-storage-crc-7drv9\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.857731 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dc05aaca-8ca7-4148-88d2-a0a613e02edb-crc-storage\") pod \"crc-storage-crc-7drv9\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.857770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dc05aaca-8ca7-4148-88d2-a0a613e02edb-node-mnt\") pod \"crc-storage-crc-7drv9\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.880079 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t48sr\" (UniqueName: \"kubernetes.io/projected/dc05aaca-8ca7-4148-88d2-a0a613e02edb-kube-api-access-t48sr\") pod \"crc-storage-crc-7drv9\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:30 crc kubenswrapper[4846]: I0202 13:29:30.982442 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:31 crc kubenswrapper[4846]: I0202 13:29:31.437791 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fee6f85-c7a1-4912-8852-f62aba2b4c6a" path="/var/lib/kubelet/pods/8fee6f85-c7a1-4912-8852-f62aba2b4c6a/volumes" Feb 02 13:29:31 crc kubenswrapper[4846]: I0202 13:29:31.439070 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-7drv9"] Feb 02 13:29:32 crc kubenswrapper[4846]: I0202 13:29:32.034726 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-7drv9" event={"ID":"dc05aaca-8ca7-4148-88d2-a0a613e02edb","Type":"ContainerStarted","Data":"22e8ef7e45501e4554aa0fa9d135e17f8e48a60163de27f8fc9a44ab35cae4cb"} Feb 02 13:29:33 crc kubenswrapper[4846]: I0202 13:29:33.042668 4846 generic.go:334] "Generic (PLEG): container finished" podID="dc05aaca-8ca7-4148-88d2-a0a613e02edb" containerID="dd203982810eea51d300bfeb6ec3f8c39a714d7e0dd88d9209ab4d0f1d3c77de" exitCode=0 Feb 02 13:29:33 crc kubenswrapper[4846]: I0202 13:29:33.042715 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-7drv9" event={"ID":"dc05aaca-8ca7-4148-88d2-a0a613e02edb","Type":"ContainerDied","Data":"dd203982810eea51d300bfeb6ec3f8c39a714d7e0dd88d9209ab4d0f1d3c77de"} Feb 02 13:29:34 crc kubenswrapper[4846]: I0202 13:29:34.417597 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:34 crc kubenswrapper[4846]: I0202 13:29:34.508914 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t48sr\" (UniqueName: \"kubernetes.io/projected/dc05aaca-8ca7-4148-88d2-a0a613e02edb-kube-api-access-t48sr\") pod \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " Feb 02 13:29:34 crc kubenswrapper[4846]: I0202 13:29:34.508975 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dc05aaca-8ca7-4148-88d2-a0a613e02edb-node-mnt\") pod \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " Feb 02 13:29:34 crc kubenswrapper[4846]: I0202 13:29:34.509048 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dc05aaca-8ca7-4148-88d2-a0a613e02edb-crc-storage\") pod \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\" (UID: \"dc05aaca-8ca7-4148-88d2-a0a613e02edb\") " Feb 02 13:29:34 crc kubenswrapper[4846]: I0202 13:29:34.509100 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc05aaca-8ca7-4148-88d2-a0a613e02edb-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "dc05aaca-8ca7-4148-88d2-a0a613e02edb" (UID: "dc05aaca-8ca7-4148-88d2-a0a613e02edb"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:29:34 crc kubenswrapper[4846]: I0202 13:29:34.510304 4846 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dc05aaca-8ca7-4148-88d2-a0a613e02edb-node-mnt\") on node \"crc\" DevicePath \"\"" Feb 02 13:29:34 crc kubenswrapper[4846]: I0202 13:29:34.516587 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc05aaca-8ca7-4148-88d2-a0a613e02edb-kube-api-access-t48sr" (OuterVolumeSpecName: "kube-api-access-t48sr") pod "dc05aaca-8ca7-4148-88d2-a0a613e02edb" (UID: "dc05aaca-8ca7-4148-88d2-a0a613e02edb"). InnerVolumeSpecName "kube-api-access-t48sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:29:34 crc kubenswrapper[4846]: I0202 13:29:34.544128 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc05aaca-8ca7-4148-88d2-a0a613e02edb-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "dc05aaca-8ca7-4148-88d2-a0a613e02edb" (UID: "dc05aaca-8ca7-4148-88d2-a0a613e02edb"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:29:34 crc kubenswrapper[4846]: I0202 13:29:34.611613 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t48sr\" (UniqueName: \"kubernetes.io/projected/dc05aaca-8ca7-4148-88d2-a0a613e02edb-kube-api-access-t48sr\") on node \"crc\" DevicePath \"\"" Feb 02 13:29:34 crc kubenswrapper[4846]: I0202 13:29:34.611658 4846 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dc05aaca-8ca7-4148-88d2-a0a613e02edb-crc-storage\") on node \"crc\" DevicePath \"\"" Feb 02 13:29:35 crc kubenswrapper[4846]: I0202 13:29:35.058353 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-7drv9" event={"ID":"dc05aaca-8ca7-4148-88d2-a0a613e02edb","Type":"ContainerDied","Data":"22e8ef7e45501e4554aa0fa9d135e17f8e48a60163de27f8fc9a44ab35cae4cb"} Feb 02 13:29:35 crc kubenswrapper[4846]: I0202 13:29:35.058401 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22e8ef7e45501e4554aa0fa9d135e17f8e48a60163de27f8fc9a44ab35cae4cb" Feb 02 13:29:35 crc kubenswrapper[4846]: I0202 13:29:35.058478 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-7drv9" Feb 02 13:29:37 crc kubenswrapper[4846]: I0202 13:29:37.674515 4846 scope.go:117] "RemoveContainer" containerID="6ce66fb16deff0f3558660d1093b0c9c70cae937604eddce8f29d9ee7d328df9" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.147495 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25"] Feb 02 13:30:00 crc kubenswrapper[4846]: E0202 13:30:00.148344 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc05aaca-8ca7-4148-88d2-a0a613e02edb" containerName="storage" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.148358 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc05aaca-8ca7-4148-88d2-a0a613e02edb" containerName="storage" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.148734 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc05aaca-8ca7-4148-88d2-a0a613e02edb" containerName="storage" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.149284 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.151769 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.153276 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.164495 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25"] Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.213891 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr65t\" (UniqueName: \"kubernetes.io/projected/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-kube-api-access-vr65t\") pod \"collect-profiles-29500650-2hs25\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.213947 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-config-volume\") pod \"collect-profiles-29500650-2hs25\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.214020 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-secret-volume\") pod \"collect-profiles-29500650-2hs25\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.315481 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-secret-volume\") pod \"collect-profiles-29500650-2hs25\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.315557 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr65t\" (UniqueName: \"kubernetes.io/projected/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-kube-api-access-vr65t\") pod \"collect-profiles-29500650-2hs25\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.315589 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-config-volume\") pod \"collect-profiles-29500650-2hs25\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.316520 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-config-volume\") pod \"collect-profiles-29500650-2hs25\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.326830 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-secret-volume\") pod \"collect-profiles-29500650-2hs25\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.334164 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr65t\" (UniqueName: \"kubernetes.io/projected/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-kube-api-access-vr65t\") pod \"collect-profiles-29500650-2hs25\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.479670 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:00 crc kubenswrapper[4846]: I0202 13:30:00.899393 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25"] Feb 02 13:30:01 crc kubenswrapper[4846]: I0202 13:30:01.248023 4846 generic.go:334] "Generic (PLEG): container finished" podID="c8022e42-1a6e-477b-aefc-7d5c76c26fd2" containerID="a53234168abee55a3e09d40a916442af33d348439dd91dfbdf9162fc3b4d3dc7" exitCode=0 Feb 02 13:30:01 crc kubenswrapper[4846]: I0202 13:30:01.248089 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" event={"ID":"c8022e42-1a6e-477b-aefc-7d5c76c26fd2","Type":"ContainerDied","Data":"a53234168abee55a3e09d40a916442af33d348439dd91dfbdf9162fc3b4d3dc7"} Feb 02 13:30:01 crc kubenswrapper[4846]: I0202 13:30:01.248388 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" event={"ID":"c8022e42-1a6e-477b-aefc-7d5c76c26fd2","Type":"ContainerStarted","Data":"0fc9b9ff53c1ca0d297461431a3ff7fdb23eee3ce46c41ebbb595b36369ffdd2"} Feb 02 13:30:02 crc kubenswrapper[4846]: I0202 13:30:02.506562 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:02 crc kubenswrapper[4846]: I0202 13:30:02.651955 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-config-volume\") pod \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " Feb 02 13:30:02 crc kubenswrapper[4846]: I0202 13:30:02.652050 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr65t\" (UniqueName: \"kubernetes.io/projected/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-kube-api-access-vr65t\") pod \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " Feb 02 13:30:02 crc kubenswrapper[4846]: I0202 13:30:02.652107 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-secret-volume\") pod \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\" (UID: \"c8022e42-1a6e-477b-aefc-7d5c76c26fd2\") " Feb 02 13:30:02 crc kubenswrapper[4846]: I0202 13:30:02.652555 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-config-volume" (OuterVolumeSpecName: "config-volume") pod "c8022e42-1a6e-477b-aefc-7d5c76c26fd2" (UID: "c8022e42-1a6e-477b-aefc-7d5c76c26fd2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:30:02 crc kubenswrapper[4846]: I0202 13:30:02.657647 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c8022e42-1a6e-477b-aefc-7d5c76c26fd2" (UID: "c8022e42-1a6e-477b-aefc-7d5c76c26fd2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:30:02 crc kubenswrapper[4846]: I0202 13:30:02.658083 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-kube-api-access-vr65t" (OuterVolumeSpecName: "kube-api-access-vr65t") pod "c8022e42-1a6e-477b-aefc-7d5c76c26fd2" (UID: "c8022e42-1a6e-477b-aefc-7d5c76c26fd2"). InnerVolumeSpecName "kube-api-access-vr65t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:30:02 crc kubenswrapper[4846]: I0202 13:30:02.753465 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 13:30:02 crc kubenswrapper[4846]: I0202 13:30:02.753509 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 13:30:02 crc kubenswrapper[4846]: I0202 13:30:02.753528 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr65t\" (UniqueName: \"kubernetes.io/projected/c8022e42-1a6e-477b-aefc-7d5c76c26fd2-kube-api-access-vr65t\") on node \"crc\" DevicePath \"\"" Feb 02 13:30:03 crc kubenswrapper[4846]: I0202 13:30:03.262682 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" Feb 02 13:30:03 crc kubenswrapper[4846]: I0202 13:30:03.262671 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25" event={"ID":"c8022e42-1a6e-477b-aefc-7d5c76c26fd2","Type":"ContainerDied","Data":"0fc9b9ff53c1ca0d297461431a3ff7fdb23eee3ce46c41ebbb595b36369ffdd2"} Feb 02 13:30:03 crc kubenswrapper[4846]: I0202 13:30:03.263074 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fc9b9ff53c1ca0d297461431a3ff7fdb23eee3ce46c41ebbb595b36369ffdd2" Feb 02 13:30:03 crc kubenswrapper[4846]: I0202 13:30:03.571872 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn"] Feb 02 13:30:03 crc kubenswrapper[4846]: I0202 13:30:03.577015 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500605-swkkn"] Feb 02 13:30:05 crc kubenswrapper[4846]: I0202 13:30:05.433863 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f814e99-ae56-4c39-a852-5efa6e466b81" path="/var/lib/kubelet/pods/0f814e99-ae56-4c39-a852-5efa6e466b81/volumes" Feb 02 13:30:37 crc kubenswrapper[4846]: I0202 13:30:37.749479 4846 scope.go:117] "RemoveContainer" containerID="8b6165518b94d6d62387f0620967b1c1b5c7f81696b97b4f2f79733d4fce3ebb" Feb 02 13:31:00 crc kubenswrapper[4846]: I0202 13:31:00.479094 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:31:00 crc kubenswrapper[4846]: I0202 13:31:00.479814 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:31:30 crc kubenswrapper[4846]: I0202 13:31:30.479055 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:31:30 crc kubenswrapper[4846]: I0202 13:31:30.479885 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.942945 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f98b88745-thmb7"] Feb 02 13:31:34 crc kubenswrapper[4846]: E0202 13:31:34.943545 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8022e42-1a6e-477b-aefc-7d5c76c26fd2" containerName="collect-profiles" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.943563 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8022e42-1a6e-477b-aefc-7d5c76c26fd2" containerName="collect-profiles" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.943736 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8022e42-1a6e-477b-aefc-7d5c76c26fd2" containerName="collect-profiles" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.944443 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f98b88745-thmb7" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.947045 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-db59r" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.947251 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.947392 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.947603 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.951749 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4641d24-28fe-48f7-8ce5-54986e03715a-config\") pod \"dnsmasq-dns-6f98b88745-thmb7\" (UID: \"d4641d24-28fe-48f7-8ce5-54986e03715a\") " pod="openstack/dnsmasq-dns-6f98b88745-thmb7" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.951816 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pnb6\" (UniqueName: \"kubernetes.io/projected/d4641d24-28fe-48f7-8ce5-54986e03715a-kube-api-access-6pnb6\") pod \"dnsmasq-dns-6f98b88745-thmb7\" (UID: \"d4641d24-28fe-48f7-8ce5-54986e03715a\") " pod="openstack/dnsmasq-dns-6f98b88745-thmb7" Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.979347 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f98b88745-thmb7"] Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.993520 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9d69655f7-cpr2z"] Feb 02 13:31:34 crc kubenswrapper[4846]: I0202 13:31:34.998534 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.002192 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.008574 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9d69655f7-cpr2z"] Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.053222 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4641d24-28fe-48f7-8ce5-54986e03715a-config\") pod \"dnsmasq-dns-6f98b88745-thmb7\" (UID: \"d4641d24-28fe-48f7-8ce5-54986e03715a\") " pod="openstack/dnsmasq-dns-6f98b88745-thmb7" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.053274 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-config\") pod \"dnsmasq-dns-9d69655f7-cpr2z\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.053340 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pnb6\" (UniqueName: \"kubernetes.io/projected/d4641d24-28fe-48f7-8ce5-54986e03715a-kube-api-access-6pnb6\") pod \"dnsmasq-dns-6f98b88745-thmb7\" (UID: \"d4641d24-28fe-48f7-8ce5-54986e03715a\") " pod="openstack/dnsmasq-dns-6f98b88745-thmb7" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.053434 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j98rm\" (UniqueName: \"kubernetes.io/projected/9e1ca213-0b3d-489e-8b40-aeeac4571a76-kube-api-access-j98rm\") pod \"dnsmasq-dns-9d69655f7-cpr2z\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.053459 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-dns-svc\") pod \"dnsmasq-dns-9d69655f7-cpr2z\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.054280 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4641d24-28fe-48f7-8ce5-54986e03715a-config\") pod \"dnsmasq-dns-6f98b88745-thmb7\" (UID: \"d4641d24-28fe-48f7-8ce5-54986e03715a\") " pod="openstack/dnsmasq-dns-6f98b88745-thmb7" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.082245 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pnb6\" (UniqueName: \"kubernetes.io/projected/d4641d24-28fe-48f7-8ce5-54986e03715a-kube-api-access-6pnb6\") pod \"dnsmasq-dns-6f98b88745-thmb7\" (UID: \"d4641d24-28fe-48f7-8ce5-54986e03715a\") " pod="openstack/dnsmasq-dns-6f98b88745-thmb7" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.138235 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f98b88745-thmb7"] Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.138835 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f98b88745-thmb7" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.154635 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j98rm\" (UniqueName: \"kubernetes.io/projected/9e1ca213-0b3d-489e-8b40-aeeac4571a76-kube-api-access-j98rm\") pod \"dnsmasq-dns-9d69655f7-cpr2z\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.154714 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-dns-svc\") pod \"dnsmasq-dns-9d69655f7-cpr2z\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.154759 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-config\") pod \"dnsmasq-dns-9d69655f7-cpr2z\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.155721 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-dns-svc\") pod \"dnsmasq-dns-9d69655f7-cpr2z\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.155727 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-config\") pod \"dnsmasq-dns-9d69655f7-cpr2z\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.168121 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5699fdb769-drstp"] Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.169181 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.188951 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j98rm\" (UniqueName: \"kubernetes.io/projected/9e1ca213-0b3d-489e-8b40-aeeac4571a76-kube-api-access-j98rm\") pod \"dnsmasq-dns-9d69655f7-cpr2z\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.223823 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5699fdb769-drstp"] Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.257185 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twkfq\" (UniqueName: \"kubernetes.io/projected/1567e5b1-a396-4b10-9440-ae3a43835be9-kube-api-access-twkfq\") pod \"dnsmasq-dns-5699fdb769-drstp\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.257275 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-dns-svc\") pod \"dnsmasq-dns-5699fdb769-drstp\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.257312 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-config\") pod \"dnsmasq-dns-5699fdb769-drstp\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.313939 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.360900 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twkfq\" (UniqueName: \"kubernetes.io/projected/1567e5b1-a396-4b10-9440-ae3a43835be9-kube-api-access-twkfq\") pod \"dnsmasq-dns-5699fdb769-drstp\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.360972 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-dns-svc\") pod \"dnsmasq-dns-5699fdb769-drstp\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.361001 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-config\") pod \"dnsmasq-dns-5699fdb769-drstp\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.361977 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-config\") pod \"dnsmasq-dns-5699fdb769-drstp\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.362416 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-dns-svc\") pod \"dnsmasq-dns-5699fdb769-drstp\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.381752 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twkfq\" (UniqueName: \"kubernetes.io/projected/1567e5b1-a396-4b10-9440-ae3a43835be9-kube-api-access-twkfq\") pod \"dnsmasq-dns-5699fdb769-drstp\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.453087 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f98b88745-thmb7"] Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.539414 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9d69655f7-cpr2z"] Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.555490 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.564450 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-589cf688cc-vdpcl"] Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.565555 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.587430 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-589cf688cc-vdpcl"] Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.674615 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-config\") pod \"dnsmasq-dns-589cf688cc-vdpcl\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.674797 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjnjv\" (UniqueName: \"kubernetes.io/projected/0247731c-e91f-45f2-8106-69b3e3c55dc6-kube-api-access-wjnjv\") pod \"dnsmasq-dns-589cf688cc-vdpcl\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.674836 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-dns-svc\") pod \"dnsmasq-dns-589cf688cc-vdpcl\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.777860 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjnjv\" (UniqueName: \"kubernetes.io/projected/0247731c-e91f-45f2-8106-69b3e3c55dc6-kube-api-access-wjnjv\") pod \"dnsmasq-dns-589cf688cc-vdpcl\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.777946 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-dns-svc\") pod \"dnsmasq-dns-589cf688cc-vdpcl\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.777982 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-config\") pod \"dnsmasq-dns-589cf688cc-vdpcl\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.779298 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-config\") pod \"dnsmasq-dns-589cf688cc-vdpcl\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.780091 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-dns-svc\") pod \"dnsmasq-dns-589cf688cc-vdpcl\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.797434 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjnjv\" (UniqueName: \"kubernetes.io/projected/0247731c-e91f-45f2-8106-69b3e3c55dc6-kube-api-access-wjnjv\") pod \"dnsmasq-dns-589cf688cc-vdpcl\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.832101 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9d69655f7-cpr2z"] Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.904910 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.939086 4846 generic.go:334] "Generic (PLEG): container finished" podID="d4641d24-28fe-48f7-8ce5-54986e03715a" containerID="62f27c46e60145f6bbfa137acd9c48a87f5302b9ef4166a7582566956ad35cca" exitCode=0 Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.939161 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f98b88745-thmb7" event={"ID":"d4641d24-28fe-48f7-8ce5-54986e03715a","Type":"ContainerDied","Data":"62f27c46e60145f6bbfa137acd9c48a87f5302b9ef4166a7582566956ad35cca"} Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.939190 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f98b88745-thmb7" event={"ID":"d4641d24-28fe-48f7-8ce5-54986e03715a","Type":"ContainerStarted","Data":"e2bffece3acaee1b94341dd5c9dd7c6b706549d390b6480e1ed67667b38c09f9"} Feb 02 13:31:35 crc kubenswrapper[4846]: I0202 13:31:35.946945 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" event={"ID":"9e1ca213-0b3d-489e-8b40-aeeac4571a76","Type":"ContainerStarted","Data":"e8500b0c151dd1202989f945b19b0c59816b3a6cb8e29d492208dedbbb9542e4"} Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.074151 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5699fdb769-drstp"] Feb 02 13:31:36 crc kubenswrapper[4846]: W0202 13:31:36.106321 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1567e5b1_a396_4b10_9440_ae3a43835be9.slice/crio-9af3d7b4fd7ec300a4a9fa9c45d671d3f43f88d614149ff3b2056493ede29bb8 WatchSource:0}: Error finding container 9af3d7b4fd7ec300a4a9fa9c45d671d3f43f88d614149ff3b2056493ede29bb8: Status 404 returned error can't find the container with id 9af3d7b4fd7ec300a4a9fa9c45d671d3f43f88d614149ff3b2056493ede29bb8 Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.267164 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f98b88745-thmb7" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.287920 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pnb6\" (UniqueName: \"kubernetes.io/projected/d4641d24-28fe-48f7-8ce5-54986e03715a-kube-api-access-6pnb6\") pod \"d4641d24-28fe-48f7-8ce5-54986e03715a\" (UID: \"d4641d24-28fe-48f7-8ce5-54986e03715a\") " Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.288017 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4641d24-28fe-48f7-8ce5-54986e03715a-config\") pod \"d4641d24-28fe-48f7-8ce5-54986e03715a\" (UID: \"d4641d24-28fe-48f7-8ce5-54986e03715a\") " Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.310044 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4641d24-28fe-48f7-8ce5-54986e03715a-kube-api-access-6pnb6" (OuterVolumeSpecName: "kube-api-access-6pnb6") pod "d4641d24-28fe-48f7-8ce5-54986e03715a" (UID: "d4641d24-28fe-48f7-8ce5-54986e03715a"). InnerVolumeSpecName "kube-api-access-6pnb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.327148 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4641d24-28fe-48f7-8ce5-54986e03715a-config" (OuterVolumeSpecName: "config") pod "d4641d24-28fe-48f7-8ce5-54986e03715a" (UID: "d4641d24-28fe-48f7-8ce5-54986e03715a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.346455 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 13:31:36 crc kubenswrapper[4846]: E0202 13:31:36.346794 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4641d24-28fe-48f7-8ce5-54986e03715a" containerName="init" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.346816 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4641d24-28fe-48f7-8ce5-54986e03715a" containerName="init" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.347025 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4641d24-28fe-48f7-8ce5-54986e03715a" containerName="init" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.347802 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.350508 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-s6zvs" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.350904 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.351536 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.351666 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.351685 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.354801 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.354812 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.357070 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389110 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389191 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389220 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389253 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389270 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389285 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389308 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389334 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389357 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdkk6\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-kube-api-access-zdkk6\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389375 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389400 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389435 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pnb6\" (UniqueName: \"kubernetes.io/projected/d4641d24-28fe-48f7-8ce5-54986e03715a-kube-api-access-6pnb6\") on node \"crc\" DevicePath \"\"" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.389447 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4641d24-28fe-48f7-8ce5-54986e03715a-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.428456 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-589cf688cc-vdpcl"] Feb 02 13:31:36 crc kubenswrapper[4846]: W0202 13:31:36.434222 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0247731c_e91f_45f2_8106_69b3e3c55dc6.slice/crio-61787784d5c0f56464be566c43bd63d3905271e40bbf71482384b7fc0af5f757 WatchSource:0}: Error finding container 61787784d5c0f56464be566c43bd63d3905271e40bbf71482384b7fc0af5f757: Status 404 returned error can't find the container with id 61787784d5c0f56464be566c43bd63d3905271e40bbf71482384b7fc0af5f757 Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492261 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492342 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492392 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492421 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492465 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492488 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492512 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492551 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492594 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492667 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdkk6\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-kube-api-access-zdkk6\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.492698 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.493137 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.493813 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.495545 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.495559 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.495807 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.497316 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.499098 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.500073 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.500103 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c5cf7ac03c1e4654a9eb7f16be0bafe95391676eee61ca36eecad9f3b956cc3e/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.500953 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.502789 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.511026 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdkk6\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-kube-api-access-zdkk6\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.533903 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") pod \"rabbitmq-cell1-server-0\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.668727 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.711869 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.713280 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.718451 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.718619 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.718678 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.718749 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.718842 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.718846 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.718946 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-94sh9" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.727202 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.898276 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-config-data\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.898676 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43634f49-b272-42a2-9e88-bfa2220498b7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.898723 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43634f49-b272-42a2-9e88-bfa2220498b7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.898787 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.898832 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.898864 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbzr8\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-kube-api-access-lbzr8\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.898915 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-63c31d1b-8df9-4273-9675-16d5003068d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.898949 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.898974 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.899004 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.899024 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.954297 4846 generic.go:334] "Generic (PLEG): container finished" podID="0247731c-e91f-45f2-8106-69b3e3c55dc6" containerID="81e7147869ec2463b15a4897692271a4ef96bfb852d09ba51098763cd66b2d24" exitCode=0 Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.954359 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" event={"ID":"0247731c-e91f-45f2-8106-69b3e3c55dc6","Type":"ContainerDied","Data":"81e7147869ec2463b15a4897692271a4ef96bfb852d09ba51098763cd66b2d24"} Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.954382 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" event={"ID":"0247731c-e91f-45f2-8106-69b3e3c55dc6","Type":"ContainerStarted","Data":"61787784d5c0f56464be566c43bd63d3905271e40bbf71482384b7fc0af5f757"} Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.957662 4846 generic.go:334] "Generic (PLEG): container finished" podID="1567e5b1-a396-4b10-9440-ae3a43835be9" containerID="935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a" exitCode=0 Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.957798 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5699fdb769-drstp" event={"ID":"1567e5b1-a396-4b10-9440-ae3a43835be9","Type":"ContainerDied","Data":"935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a"} Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.957945 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5699fdb769-drstp" event={"ID":"1567e5b1-a396-4b10-9440-ae3a43835be9","Type":"ContainerStarted","Data":"9af3d7b4fd7ec300a4a9fa9c45d671d3f43f88d614149ff3b2056493ede29bb8"} Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.959532 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f98b88745-thmb7" event={"ID":"d4641d24-28fe-48f7-8ce5-54986e03715a","Type":"ContainerDied","Data":"e2bffece3acaee1b94341dd5c9dd7c6b706549d390b6480e1ed67667b38c09f9"} Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.959574 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f98b88745-thmb7" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.959592 4846 scope.go:117] "RemoveContainer" containerID="62f27c46e60145f6bbfa137acd9c48a87f5302b9ef4166a7582566956ad35cca" Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.961493 4846 generic.go:334] "Generic (PLEG): container finished" podID="9e1ca213-0b3d-489e-8b40-aeeac4571a76" containerID="1f6e4879c14e429daa6e56ee50d5ee66ad1ddcf0b9df6458e6e6ba59401c6e28" exitCode=0 Feb 02 13:31:36 crc kubenswrapper[4846]: I0202 13:31:36.961574 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" event={"ID":"9e1ca213-0b3d-489e-8b40-aeeac4571a76","Type":"ContainerDied","Data":"1f6e4879c14e429daa6e56ee50d5ee66ad1ddcf0b9df6458e6e6ba59401c6e28"} Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.005235 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.006335 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.006416 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.006445 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-config-data\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.006476 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43634f49-b272-42a2-9e88-bfa2220498b7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.006517 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43634f49-b272-42a2-9e88-bfa2220498b7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.006574 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.006615 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.006665 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbzr8\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-kube-api-access-lbzr8\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.006731 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-63c31d1b-8df9-4273-9675-16d5003068d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.006773 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.007172 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.007494 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-config-data\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.007864 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.008499 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.008989 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.011020 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.015444 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.015515 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-63c31d1b-8df9-4273-9675-16d5003068d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/486b899c90445695a6faaf1f10d5b80276ce921cece83762353508bece56da44/globalmount\"" pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.016449 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43634f49-b272-42a2-9e88-bfa2220498b7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.020462 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.022611 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43634f49-b272-42a2-9e88-bfa2220498b7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.027600 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbzr8\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-kube-api-access-lbzr8\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.062923 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f98b88745-thmb7"] Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.070170 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-63c31d1b-8df9-4273-9675-16d5003068d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") pod \"rabbitmq-server-0\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.072190 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f98b88745-thmb7"] Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.133658 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.324959 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.331664 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 13:31:37 crc kubenswrapper[4846]: E0202 13:31:37.338684 4846 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Feb 02 13:31:37 crc kubenswrapper[4846]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/1567e5b1-a396-4b10-9440-ae3a43835be9/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 02 13:31:37 crc kubenswrapper[4846]: > podSandboxID="9af3d7b4fd7ec300a4a9fa9c45d671d3f43f88d614149ff3b2056493ede29bb8" Feb 02 13:31:37 crc kubenswrapper[4846]: E0202 13:31:37.338878 4846 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 02 13:31:37 crc kubenswrapper[4846]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:f391b842000dadaeb692eb6b5e845c2aa3125ef24679fbb4af2c8b98252de4b2,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb6hc5h68h68h594h659hdbh679h65ch5f6hdch6h5b9h8fh55hfhf8h57fhc7h56ch687h669h559h678h5dhc7hf7h697h5d6h9ch669h54fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-twkfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5699fdb769-drstp_openstack(1567e5b1-a396-4b10-9440-ae3a43835be9): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/1567e5b1-a396-4b10-9440-ae3a43835be9/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 02 13:31:37 crc kubenswrapper[4846]: > logger="UnhandledError" Feb 02 13:31:37 crc kubenswrapper[4846]: E0202 13:31:37.340115 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/1567e5b1-a396-4b10-9440-ae3a43835be9/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5699fdb769-drstp" podUID="1567e5b1-a396-4b10-9440-ae3a43835be9" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.435247 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4641d24-28fe-48f7-8ce5-54986e03715a" path="/var/lib/kubelet/pods/d4641d24-28fe-48f7-8ce5-54986e03715a/volumes" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.513422 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j98rm\" (UniqueName: \"kubernetes.io/projected/9e1ca213-0b3d-489e-8b40-aeeac4571a76-kube-api-access-j98rm\") pod \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.513789 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-dns-svc\") pod \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.513929 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-config\") pod \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\" (UID: \"9e1ca213-0b3d-489e-8b40-aeeac4571a76\") " Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.519206 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e1ca213-0b3d-489e-8b40-aeeac4571a76-kube-api-access-j98rm" (OuterVolumeSpecName: "kube-api-access-j98rm") pod "9e1ca213-0b3d-489e-8b40-aeeac4571a76" (UID: "9e1ca213-0b3d-489e-8b40-aeeac4571a76"). InnerVolumeSpecName "kube-api-access-j98rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.531235 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e1ca213-0b3d-489e-8b40-aeeac4571a76" (UID: "9e1ca213-0b3d-489e-8b40-aeeac4571a76"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.535077 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-config" (OuterVolumeSpecName: "config") pod "9e1ca213-0b3d-489e-8b40-aeeac4571a76" (UID: "9e1ca213-0b3d-489e-8b40-aeeac4571a76"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.563347 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 13:31:37 crc kubenswrapper[4846]: W0202 13:31:37.569336 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43634f49_b272_42a2_9e88_bfa2220498b7.slice/crio-a88bd0ff8ecdc22983e52c8875d681630f388cbf2a0fdb987d31e1ba9cc4ae63 WatchSource:0}: Error finding container a88bd0ff8ecdc22983e52c8875d681630f388cbf2a0fdb987d31e1ba9cc4ae63: Status 404 returned error can't find the container with id a88bd0ff8ecdc22983e52c8875d681630f388cbf2a0fdb987d31e1ba9cc4ae63 Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.616572 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j98rm\" (UniqueName: \"kubernetes.io/projected/9e1ca213-0b3d-489e-8b40-aeeac4571a76-kube-api-access-j98rm\") on node \"crc\" DevicePath \"\"" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.616613 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.616629 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e1ca213-0b3d-489e-8b40-aeeac4571a76-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.657817 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 02 13:31:37 crc kubenswrapper[4846]: E0202 13:31:37.658158 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e1ca213-0b3d-489e-8b40-aeeac4571a76" containerName="init" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.658178 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e1ca213-0b3d-489e-8b40-aeeac4571a76" containerName="init" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.658361 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e1ca213-0b3d-489e-8b40-aeeac4571a76" containerName="init" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.661466 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.663910 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-qlwcn" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.664824 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.667040 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.667917 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.668576 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.674312 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.819669 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bed62f97-3e65-40e9-bc60-1b0338fc4672-operator-scripts\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.819719 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bed62f97-3e65-40e9-bc60-1b0338fc4672-config-data-default\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.819752 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bed62f97-3e65-40e9-bc60-1b0338fc4672-kolla-config\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.819866 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bed62f97-3e65-40e9-bc60-1b0338fc4672-config-data-generated\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.820003 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-24357ce1-1fed-4d34-ba3c-2fdc20a3e660\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-24357ce1-1fed-4d34-ba3c-2fdc20a3e660\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.820031 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed62f97-3e65-40e9-bc60-1b0338fc4672-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.820066 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6xmt\" (UniqueName: \"kubernetes.io/projected/bed62f97-3e65-40e9-bc60-1b0338fc4672-kube-api-access-v6xmt\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.820090 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bed62f97-3e65-40e9-bc60-1b0338fc4672-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.921327 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bed62f97-3e65-40e9-bc60-1b0338fc4672-operator-scripts\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.921373 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bed62f97-3e65-40e9-bc60-1b0338fc4672-config-data-default\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.921403 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bed62f97-3e65-40e9-bc60-1b0338fc4672-kolla-config\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.921434 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bed62f97-3e65-40e9-bc60-1b0338fc4672-config-data-generated\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.921479 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-24357ce1-1fed-4d34-ba3c-2fdc20a3e660\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-24357ce1-1fed-4d34-ba3c-2fdc20a3e660\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.921497 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed62f97-3e65-40e9-bc60-1b0338fc4672-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.921519 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6xmt\" (UniqueName: \"kubernetes.io/projected/bed62f97-3e65-40e9-bc60-1b0338fc4672-kube-api-access-v6xmt\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.921535 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bed62f97-3e65-40e9-bc60-1b0338fc4672-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.921947 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bed62f97-3e65-40e9-bc60-1b0338fc4672-config-data-generated\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.922502 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bed62f97-3e65-40e9-bc60-1b0338fc4672-config-data-default\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.922844 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bed62f97-3e65-40e9-bc60-1b0338fc4672-operator-scripts\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.924048 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bed62f97-3e65-40e9-bc60-1b0338fc4672-kolla-config\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.924211 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.924288 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-24357ce1-1fed-4d34-ba3c-2fdc20a3e660\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-24357ce1-1fed-4d34-ba3c-2fdc20a3e660\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b7d76cd28a20d6c5b5e3e5d5807d13ba9882d01152f862268aae6c82798bea45/globalmount\"" pod="openstack/openstack-galera-0" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.970490 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" event={"ID":"0247731c-e91f-45f2-8106-69b3e3c55dc6","Type":"ContainerStarted","Data":"1df810e40e64a52836ac0e8e3b00831c733ffeb575ce07c24efa8dd191aff9ab"} Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.971584 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.972729 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"43634f49-b272-42a2-9e88-bfa2220498b7","Type":"ContainerStarted","Data":"a88bd0ff8ecdc22983e52c8875d681630f388cbf2a0fdb987d31e1ba9cc4ae63"} Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.975709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9c2deb33-5a4e-4b54-9273-d2c7ddf11668","Type":"ContainerStarted","Data":"03cf2e6fc8c9b4194049af615337492d2d8166ff0e8ecf68298f9977b0477d9a"} Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.978621 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.978620 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d69655f7-cpr2z" event={"ID":"9e1ca213-0b3d-489e-8b40-aeeac4571a76","Type":"ContainerDied","Data":"e8500b0c151dd1202989f945b19b0c59816b3a6cb8e29d492208dedbbb9542e4"} Feb 02 13:31:37 crc kubenswrapper[4846]: I0202 13:31:37.978782 4846 scope.go:117] "RemoveContainer" containerID="1f6e4879c14e429daa6e56ee50d5ee66ad1ddcf0b9df6458e6e6ba59401c6e28" Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.002525 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bed62f97-3e65-40e9-bc60-1b0338fc4672-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.006386 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed62f97-3e65-40e9-bc60-1b0338fc4672-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.019141 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6xmt\" (UniqueName: \"kubernetes.io/projected/bed62f97-3e65-40e9-bc60-1b0338fc4672-kube-api-access-v6xmt\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.020491 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" podStartSLOduration=3.02047067 podStartE2EDuration="3.02047067s" podCreationTimestamp="2026-02-02 13:31:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:31:37.991403223 +0000 UTC m=+4929.219990086" watchObservedRunningTime="2026-02-02 13:31:38.02047067 +0000 UTC m=+4929.249057533" Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.064972 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9d69655f7-cpr2z"] Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.070425 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9d69655f7-cpr2z"] Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.153177 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-24357ce1-1fed-4d34-ba3c-2fdc20a3e660\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-24357ce1-1fed-4d34-ba3c-2fdc20a3e660\") pod \"openstack-galera-0\" (UID: \"bed62f97-3e65-40e9-bc60-1b0338fc4672\") " pod="openstack/openstack-galera-0" Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.279455 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.763919 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 02 13:31:38 crc kubenswrapper[4846]: W0202 13:31:38.766447 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbed62f97_3e65_40e9_bc60_1b0338fc4672.slice/crio-396c8ff3455b97ff637bd36fc6d86d536bf650d008c412d2b4d5db91239258f6 WatchSource:0}: Error finding container 396c8ff3455b97ff637bd36fc6d86d536bf650d008c412d2b4d5db91239258f6: Status 404 returned error can't find the container with id 396c8ff3455b97ff637bd36fc6d86d536bf650d008c412d2b4d5db91239258f6 Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.986322 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bed62f97-3e65-40e9-bc60-1b0338fc4672","Type":"ContainerStarted","Data":"96d23ca2aa23c5701ef18db6ecb31f33d30fafd86940e07b5c7ef29a0ed1c4a7"} Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.986680 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bed62f97-3e65-40e9-bc60-1b0338fc4672","Type":"ContainerStarted","Data":"396c8ff3455b97ff637bd36fc6d86d536bf650d008c412d2b4d5db91239258f6"} Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.988079 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5699fdb769-drstp" event={"ID":"1567e5b1-a396-4b10-9440-ae3a43835be9","Type":"ContainerStarted","Data":"332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c"} Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.988349 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.989300 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"43634f49-b272-42a2-9e88-bfa2220498b7","Type":"ContainerStarted","Data":"80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e"} Feb 02 13:31:38 crc kubenswrapper[4846]: I0202 13:31:38.990472 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9c2deb33-5a4e-4b54-9273-d2c7ddf11668","Type":"ContainerStarted","Data":"6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67"} Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.123430 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5699fdb769-drstp" podStartSLOduration=4.123407947 podStartE2EDuration="4.123407947s" podCreationTimestamp="2026-02-02 13:31:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:31:39.119782548 +0000 UTC m=+4930.348369421" watchObservedRunningTime="2026-02-02 13:31:39.123407947 +0000 UTC m=+4930.351994830" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.226415 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.227677 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.230349 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-nn2jp" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.230518 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.232229 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.238683 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.243157 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.348171 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3003f765-6322-4adf-a73d-1ded30560bf9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.348258 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-230a2978-e4d1-419e-93f2-0eeaf1f24d95\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-230a2978-e4d1-419e-93f2-0eeaf1f24d95\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.348286 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw857\" (UniqueName: \"kubernetes.io/projected/3003f765-6322-4adf-a73d-1ded30560bf9-kube-api-access-gw857\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.348308 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3003f765-6322-4adf-a73d-1ded30560bf9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.348343 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3003f765-6322-4adf-a73d-1ded30560bf9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.348363 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3003f765-6322-4adf-a73d-1ded30560bf9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.348444 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3003f765-6322-4adf-a73d-1ded30560bf9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.348519 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3003f765-6322-4adf-a73d-1ded30560bf9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.432141 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e1ca213-0b3d-489e-8b40-aeeac4571a76" path="/var/lib/kubelet/pods/9e1ca213-0b3d-489e-8b40-aeeac4571a76/volumes" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.450246 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3003f765-6322-4adf-a73d-1ded30560bf9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.450621 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3003f765-6322-4adf-a73d-1ded30560bf9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.450753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3003f765-6322-4adf-a73d-1ded30560bf9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.450875 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-230a2978-e4d1-419e-93f2-0eeaf1f24d95\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-230a2978-e4d1-419e-93f2-0eeaf1f24d95\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.450981 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw857\" (UniqueName: \"kubernetes.io/projected/3003f765-6322-4adf-a73d-1ded30560bf9-kube-api-access-gw857\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.451138 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3003f765-6322-4adf-a73d-1ded30560bf9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.451257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3003f765-6322-4adf-a73d-1ded30560bf9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.451342 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3003f765-6322-4adf-a73d-1ded30560bf9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.451790 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3003f765-6322-4adf-a73d-1ded30560bf9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.451913 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3003f765-6322-4adf-a73d-1ded30560bf9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.452718 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3003f765-6322-4adf-a73d-1ded30560bf9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.453116 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3003f765-6322-4adf-a73d-1ded30560bf9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.455047 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.455221 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-230a2978-e4d1-419e-93f2-0eeaf1f24d95\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-230a2978-e4d1-419e-93f2-0eeaf1f24d95\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4801570c4af8bb7ec08d6f5be5372eec605256f33f0a4f63c81df4effd4cd521/globalmount\"" pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.455247 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3003f765-6322-4adf-a73d-1ded30560bf9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.455612 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3003f765-6322-4adf-a73d-1ded30560bf9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.471686 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw857\" (UniqueName: \"kubernetes.io/projected/3003f765-6322-4adf-a73d-1ded30560bf9-kube-api-access-gw857\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.478813 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-230a2978-e4d1-419e-93f2-0eeaf1f24d95\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-230a2978-e4d1-419e-93f2-0eeaf1f24d95\") pod \"openstack-cell1-galera-0\" (UID: \"3003f765-6322-4adf-a73d-1ded30560bf9\") " pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.542851 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.797023 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.798603 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.809183 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.809313 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.809437 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-nlb9f" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.812654 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.861689 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/705ad25e-8635-426c-8069-f09e3b16712f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.861734 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw7pv\" (UniqueName: \"kubernetes.io/projected/705ad25e-8635-426c-8069-f09e3b16712f-kube-api-access-jw7pv\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.861784 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/705ad25e-8635-426c-8069-f09e3b16712f-config-data\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.861816 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/705ad25e-8635-426c-8069-f09e3b16712f-kolla-config\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.861852 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/705ad25e-8635-426c-8069-f09e3b16712f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.963149 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/705ad25e-8635-426c-8069-f09e3b16712f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.963209 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw7pv\" (UniqueName: \"kubernetes.io/projected/705ad25e-8635-426c-8069-f09e3b16712f-kube-api-access-jw7pv\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.963278 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/705ad25e-8635-426c-8069-f09e3b16712f-config-data\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.963331 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/705ad25e-8635-426c-8069-f09e3b16712f-kolla-config\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.963366 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/705ad25e-8635-426c-8069-f09e3b16712f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.964399 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/705ad25e-8635-426c-8069-f09e3b16712f-config-data\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.964918 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/705ad25e-8635-426c-8069-f09e3b16712f-kolla-config\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.967410 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/705ad25e-8635-426c-8069-f09e3b16712f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.967455 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/705ad25e-8635-426c-8069-f09e3b16712f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:39 crc kubenswrapper[4846]: I0202 13:31:39.982045 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw7pv\" (UniqueName: \"kubernetes.io/projected/705ad25e-8635-426c-8069-f09e3b16712f-kube-api-access-jw7pv\") pod \"memcached-0\" (UID: \"705ad25e-8635-426c-8069-f09e3b16712f\") " pod="openstack/memcached-0" Feb 02 13:31:40 crc kubenswrapper[4846]: I0202 13:31:40.081274 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 02 13:31:40 crc kubenswrapper[4846]: I0202 13:31:40.128590 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 02 13:31:40 crc kubenswrapper[4846]: I0202 13:31:40.600398 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 02 13:31:41 crc kubenswrapper[4846]: I0202 13:31:41.016362 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3003f765-6322-4adf-a73d-1ded30560bf9","Type":"ContainerStarted","Data":"d61e4cbfe155248317ac65f0db14530ff7fd22fbad19c349f3d1747d9a07dccf"} Feb 02 13:31:41 crc kubenswrapper[4846]: I0202 13:31:41.016407 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3003f765-6322-4adf-a73d-1ded30560bf9","Type":"ContainerStarted","Data":"13f4218e041e160c7dc628ba814a1938b960f76d24a3f9ec95bcf2d2d42d7e32"} Feb 02 13:31:41 crc kubenswrapper[4846]: I0202 13:31:41.018406 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"705ad25e-8635-426c-8069-f09e3b16712f","Type":"ContainerStarted","Data":"aec70c7967182980e1d14995a020e23f97a3b671eb5e4c4f1a41ae73d17881e8"} Feb 02 13:31:41 crc kubenswrapper[4846]: I0202 13:31:41.018445 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"705ad25e-8635-426c-8069-f09e3b16712f","Type":"ContainerStarted","Data":"6ee7c81f3704c2178f85b02c9909aa6a53274a2635f85b6890d35071d41d4e28"} Feb 02 13:31:41 crc kubenswrapper[4846]: I0202 13:31:41.018557 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 02 13:31:41 crc kubenswrapper[4846]: I0202 13:31:41.061576 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.061557346 podStartE2EDuration="2.061557346s" podCreationTimestamp="2026-02-02 13:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:31:41.056320308 +0000 UTC m=+4932.284907191" watchObservedRunningTime="2026-02-02 13:31:41.061557346 +0000 UTC m=+4932.290144209" Feb 02 13:31:43 crc kubenswrapper[4846]: I0202 13:31:43.034573 4846 generic.go:334] "Generic (PLEG): container finished" podID="bed62f97-3e65-40e9-bc60-1b0338fc4672" containerID="96d23ca2aa23c5701ef18db6ecb31f33d30fafd86940e07b5c7ef29a0ed1c4a7" exitCode=0 Feb 02 13:31:43 crc kubenswrapper[4846]: I0202 13:31:43.034653 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bed62f97-3e65-40e9-bc60-1b0338fc4672","Type":"ContainerDied","Data":"96d23ca2aa23c5701ef18db6ecb31f33d30fafd86940e07b5c7ef29a0ed1c4a7"} Feb 02 13:31:44 crc kubenswrapper[4846]: I0202 13:31:44.043584 4846 generic.go:334] "Generic (PLEG): container finished" podID="3003f765-6322-4adf-a73d-1ded30560bf9" containerID="d61e4cbfe155248317ac65f0db14530ff7fd22fbad19c349f3d1747d9a07dccf" exitCode=0 Feb 02 13:31:44 crc kubenswrapper[4846]: I0202 13:31:44.043677 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3003f765-6322-4adf-a73d-1ded30560bf9","Type":"ContainerDied","Data":"d61e4cbfe155248317ac65f0db14530ff7fd22fbad19c349f3d1747d9a07dccf"} Feb 02 13:31:44 crc kubenswrapper[4846]: I0202 13:31:44.048252 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"bed62f97-3e65-40e9-bc60-1b0338fc4672","Type":"ContainerStarted","Data":"5369fec9a60e7349a78147df678a5a60820d226afd25cc4c80cbbe624f589e90"} Feb 02 13:31:44 crc kubenswrapper[4846]: I0202 13:31:44.094267 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.094247278 podStartE2EDuration="8.094247278s" podCreationTimestamp="2026-02-02 13:31:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:31:44.087941014 +0000 UTC m=+4935.316527887" watchObservedRunningTime="2026-02-02 13:31:44.094247278 +0000 UTC m=+4935.322834141" Feb 02 13:31:45 crc kubenswrapper[4846]: I0202 13:31:45.058989 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3003f765-6322-4adf-a73d-1ded30560bf9","Type":"ContainerStarted","Data":"87938e1e0a0d431602414d7afed0ea3e1ceb2df7f75367a3a1250bd2e364a647"} Feb 02 13:31:45 crc kubenswrapper[4846]: I0202 13:31:45.091500 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.091470606 podStartE2EDuration="7.091470606s" podCreationTimestamp="2026-02-02 13:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:31:45.08423043 +0000 UTC m=+4936.312817313" watchObservedRunningTime="2026-02-02 13:31:45.091470606 +0000 UTC m=+4936.320057509" Feb 02 13:31:45 crc kubenswrapper[4846]: I0202 13:31:45.129807 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 02 13:31:45 crc kubenswrapper[4846]: I0202 13:31:45.557399 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:45 crc kubenswrapper[4846]: I0202 13:31:45.906829 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:31:45 crc kubenswrapper[4846]: I0202 13:31:45.962316 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5699fdb769-drstp"] Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.066083 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5699fdb769-drstp" podUID="1567e5b1-a396-4b10-9440-ae3a43835be9" containerName="dnsmasq-dns" containerID="cri-o://332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c" gracePeriod=10 Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.552831 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.675772 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-config\") pod \"1567e5b1-a396-4b10-9440-ae3a43835be9\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.676199 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-dns-svc\") pod \"1567e5b1-a396-4b10-9440-ae3a43835be9\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.676286 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twkfq\" (UniqueName: \"kubernetes.io/projected/1567e5b1-a396-4b10-9440-ae3a43835be9-kube-api-access-twkfq\") pod \"1567e5b1-a396-4b10-9440-ae3a43835be9\" (UID: \"1567e5b1-a396-4b10-9440-ae3a43835be9\") " Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.685088 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1567e5b1-a396-4b10-9440-ae3a43835be9-kube-api-access-twkfq" (OuterVolumeSpecName: "kube-api-access-twkfq") pod "1567e5b1-a396-4b10-9440-ae3a43835be9" (UID: "1567e5b1-a396-4b10-9440-ae3a43835be9"). InnerVolumeSpecName "kube-api-access-twkfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.710381 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-config" (OuterVolumeSpecName: "config") pod "1567e5b1-a396-4b10-9440-ae3a43835be9" (UID: "1567e5b1-a396-4b10-9440-ae3a43835be9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.723886 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1567e5b1-a396-4b10-9440-ae3a43835be9" (UID: "1567e5b1-a396-4b10-9440-ae3a43835be9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.778230 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.778276 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twkfq\" (UniqueName: \"kubernetes.io/projected/1567e5b1-a396-4b10-9440-ae3a43835be9-kube-api-access-twkfq\") on node \"crc\" DevicePath \"\"" Feb 02 13:31:46 crc kubenswrapper[4846]: I0202 13:31:46.778289 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1567e5b1-a396-4b10-9440-ae3a43835be9-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.073851 4846 generic.go:334] "Generic (PLEG): container finished" podID="1567e5b1-a396-4b10-9440-ae3a43835be9" containerID="332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c" exitCode=0 Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.073897 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5699fdb769-drstp" event={"ID":"1567e5b1-a396-4b10-9440-ae3a43835be9","Type":"ContainerDied","Data":"332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c"} Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.073926 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5699fdb769-drstp" event={"ID":"1567e5b1-a396-4b10-9440-ae3a43835be9","Type":"ContainerDied","Data":"9af3d7b4fd7ec300a4a9fa9c45d671d3f43f88d614149ff3b2056493ede29bb8"} Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.073945 4846 scope.go:117] "RemoveContainer" containerID="332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c" Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.074084 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5699fdb769-drstp" Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.094105 4846 scope.go:117] "RemoveContainer" containerID="935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a" Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.110539 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5699fdb769-drstp"] Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.116199 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5699fdb769-drstp"] Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.130044 4846 scope.go:117] "RemoveContainer" containerID="332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c" Feb 02 13:31:47 crc kubenswrapper[4846]: E0202 13:31:47.130422 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c\": container with ID starting with 332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c not found: ID does not exist" containerID="332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c" Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.130454 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c"} err="failed to get container status \"332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c\": rpc error: code = NotFound desc = could not find container \"332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c\": container with ID starting with 332bf313afde58148a623798495f5fb35d2ca68d4b3afa4571f4ad594be0665c not found: ID does not exist" Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.130472 4846 scope.go:117] "RemoveContainer" containerID="935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a" Feb 02 13:31:47 crc kubenswrapper[4846]: E0202 13:31:47.130772 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a\": container with ID starting with 935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a not found: ID does not exist" containerID="935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a" Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.130790 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a"} err="failed to get container status \"935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a\": rpc error: code = NotFound desc = could not find container \"935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a\": container with ID starting with 935f94580951028d3c3531d3244cf6a2c15f01c20d8964eb80d3b347ed467b5a not found: ID does not exist" Feb 02 13:31:47 crc kubenswrapper[4846]: I0202 13:31:47.430876 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1567e5b1-a396-4b10-9440-ae3a43835be9" path="/var/lib/kubelet/pods/1567e5b1-a396-4b10-9440-ae3a43835be9/volumes" Feb 02 13:31:48 crc kubenswrapper[4846]: I0202 13:31:48.279996 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 02 13:31:48 crc kubenswrapper[4846]: I0202 13:31:48.280463 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 02 13:31:48 crc kubenswrapper[4846]: I0202 13:31:48.373766 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 02 13:31:49 crc kubenswrapper[4846]: I0202 13:31:49.151385 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 02 13:31:49 crc kubenswrapper[4846]: I0202 13:31:49.543191 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:49 crc kubenswrapper[4846]: I0202 13:31:49.543236 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:51 crc kubenswrapper[4846]: I0202 13:31:51.840025 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:51 crc kubenswrapper[4846]: I0202 13:31:51.925212 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.650256 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-b2jpg"] Feb 02 13:31:56 crc kubenswrapper[4846]: E0202 13:31:56.651372 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1567e5b1-a396-4b10-9440-ae3a43835be9" containerName="init" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.651397 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1567e5b1-a396-4b10-9440-ae3a43835be9" containerName="init" Feb 02 13:31:56 crc kubenswrapper[4846]: E0202 13:31:56.651444 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1567e5b1-a396-4b10-9440-ae3a43835be9" containerName="dnsmasq-dns" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.651456 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1567e5b1-a396-4b10-9440-ae3a43835be9" containerName="dnsmasq-dns" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.651702 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1567e5b1-a396-4b10-9440-ae3a43835be9" containerName="dnsmasq-dns" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.652472 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-b2jpg" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.654652 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.656710 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-b2jpg"] Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.761138 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnzf5\" (UniqueName: \"kubernetes.io/projected/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-kube-api-access-qnzf5\") pod \"root-account-create-update-b2jpg\" (UID: \"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093\") " pod="openstack/root-account-create-update-b2jpg" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.761199 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-operator-scripts\") pod \"root-account-create-update-b2jpg\" (UID: \"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093\") " pod="openstack/root-account-create-update-b2jpg" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.862546 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnzf5\" (UniqueName: \"kubernetes.io/projected/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-kube-api-access-qnzf5\") pod \"root-account-create-update-b2jpg\" (UID: \"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093\") " pod="openstack/root-account-create-update-b2jpg" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.862591 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-operator-scripts\") pod \"root-account-create-update-b2jpg\" (UID: \"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093\") " pod="openstack/root-account-create-update-b2jpg" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.863339 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-operator-scripts\") pod \"root-account-create-update-b2jpg\" (UID: \"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093\") " pod="openstack/root-account-create-update-b2jpg" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.882759 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnzf5\" (UniqueName: \"kubernetes.io/projected/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-kube-api-access-qnzf5\") pod \"root-account-create-update-b2jpg\" (UID: \"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093\") " pod="openstack/root-account-create-update-b2jpg" Feb 02 13:31:56 crc kubenswrapper[4846]: I0202 13:31:56.984700 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-b2jpg" Feb 02 13:31:57 crc kubenswrapper[4846]: W0202 13:31:57.438323 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab7ee4cc_cb04_4aa5_ab64_dd3e75b68093.slice/crio-f5338948e465e24c35dfbd957e73d6a0a09e5e88a7ed79184770c750b6689478 WatchSource:0}: Error finding container f5338948e465e24c35dfbd957e73d6a0a09e5e88a7ed79184770c750b6689478: Status 404 returned error can't find the container with id f5338948e465e24c35dfbd957e73d6a0a09e5e88a7ed79184770c750b6689478 Feb 02 13:31:57 crc kubenswrapper[4846]: I0202 13:31:57.440751 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-b2jpg"] Feb 02 13:31:58 crc kubenswrapper[4846]: I0202 13:31:58.151508 4846 generic.go:334] "Generic (PLEG): container finished" podID="ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093" containerID="c8b08e4c031e4ebef7e1a74092d58616de6c71c9dc6920c55aca4c82f85f0e6e" exitCode=0 Feb 02 13:31:58 crc kubenswrapper[4846]: I0202 13:31:58.151776 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-b2jpg" event={"ID":"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093","Type":"ContainerDied","Data":"c8b08e4c031e4ebef7e1a74092d58616de6c71c9dc6920c55aca4c82f85f0e6e"} Feb 02 13:31:58 crc kubenswrapper[4846]: I0202 13:31:58.151837 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-b2jpg" event={"ID":"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093","Type":"ContainerStarted","Data":"f5338948e465e24c35dfbd957e73d6a0a09e5e88a7ed79184770c750b6689478"} Feb 02 13:31:59 crc kubenswrapper[4846]: I0202 13:31:59.468313 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-b2jpg" Feb 02 13:31:59 crc kubenswrapper[4846]: I0202 13:31:59.511275 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-operator-scripts\") pod \"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093\" (UID: \"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093\") " Feb 02 13:31:59 crc kubenswrapper[4846]: I0202 13:31:59.511473 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnzf5\" (UniqueName: \"kubernetes.io/projected/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-kube-api-access-qnzf5\") pod \"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093\" (UID: \"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093\") " Feb 02 13:31:59 crc kubenswrapper[4846]: I0202 13:31:59.512545 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093" (UID: "ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:31:59 crc kubenswrapper[4846]: I0202 13:31:59.520906 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-kube-api-access-qnzf5" (OuterVolumeSpecName: "kube-api-access-qnzf5") pod "ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093" (UID: "ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093"). InnerVolumeSpecName "kube-api-access-qnzf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:31:59 crc kubenswrapper[4846]: I0202 13:31:59.613438 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnzf5\" (UniqueName: \"kubernetes.io/projected/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-kube-api-access-qnzf5\") on node \"crc\" DevicePath \"\"" Feb 02 13:31:59 crc kubenswrapper[4846]: I0202 13:31:59.613486 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:00 crc kubenswrapper[4846]: I0202 13:32:00.169935 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-b2jpg" event={"ID":"ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093","Type":"ContainerDied","Data":"f5338948e465e24c35dfbd957e73d6a0a09e5e88a7ed79184770c750b6689478"} Feb 02 13:32:00 crc kubenswrapper[4846]: I0202 13:32:00.169976 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5338948e465e24c35dfbd957e73d6a0a09e5e88a7ed79184770c750b6689478" Feb 02 13:32:00 crc kubenswrapper[4846]: I0202 13:32:00.170053 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-b2jpg" Feb 02 13:32:00 crc kubenswrapper[4846]: I0202 13:32:00.479280 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:32:00 crc kubenswrapper[4846]: I0202 13:32:00.479369 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:32:00 crc kubenswrapper[4846]: I0202 13:32:00.479436 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 13:32:00 crc kubenswrapper[4846]: I0202 13:32:00.480344 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 13:32:00 crc kubenswrapper[4846]: I0202 13:32:00.480474 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" gracePeriod=600 Feb 02 13:32:00 crc kubenswrapper[4846]: E0202 13:32:00.602144 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:32:01 crc kubenswrapper[4846]: I0202 13:32:01.177568 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" exitCode=0 Feb 02 13:32:01 crc kubenswrapper[4846]: I0202 13:32:01.177610 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5"} Feb 02 13:32:01 crc kubenswrapper[4846]: I0202 13:32:01.177661 4846 scope.go:117] "RemoveContainer" containerID="118141657a51d573c44958b377ad8411707adb515a180c423602a0ce85f6f046" Feb 02 13:32:01 crc kubenswrapper[4846]: I0202 13:32:01.178187 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:32:01 crc kubenswrapper[4846]: E0202 13:32:01.178400 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:32:03 crc kubenswrapper[4846]: I0202 13:32:03.218402 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-b2jpg"] Feb 02 13:32:03 crc kubenswrapper[4846]: I0202 13:32:03.224677 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-b2jpg"] Feb 02 13:32:03 crc kubenswrapper[4846]: I0202 13:32:03.433560 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093" path="/var/lib/kubelet/pods/ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093/volumes" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.226638 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-mqzll"] Feb 02 13:32:08 crc kubenswrapper[4846]: E0202 13:32:08.227485 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093" containerName="mariadb-account-create-update" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.227501 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093" containerName="mariadb-account-create-update" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.227675 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab7ee4cc-cb04-4aa5-ab64-dd3e75b68093" containerName="mariadb-account-create-update" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.228233 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-mqzll" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.235001 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.250292 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-mqzll"] Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.347432 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s274d\" (UniqueName: \"kubernetes.io/projected/8ab60cda-aa99-4608-9011-6e9e569ca203-kube-api-access-s274d\") pod \"root-account-create-update-mqzll\" (UID: \"8ab60cda-aa99-4608-9011-6e9e569ca203\") " pod="openstack/root-account-create-update-mqzll" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.347849 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ab60cda-aa99-4608-9011-6e9e569ca203-operator-scripts\") pod \"root-account-create-update-mqzll\" (UID: \"8ab60cda-aa99-4608-9011-6e9e569ca203\") " pod="openstack/root-account-create-update-mqzll" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.449017 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s274d\" (UniqueName: \"kubernetes.io/projected/8ab60cda-aa99-4608-9011-6e9e569ca203-kube-api-access-s274d\") pod \"root-account-create-update-mqzll\" (UID: \"8ab60cda-aa99-4608-9011-6e9e569ca203\") " pod="openstack/root-account-create-update-mqzll" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.449083 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ab60cda-aa99-4608-9011-6e9e569ca203-operator-scripts\") pod \"root-account-create-update-mqzll\" (UID: \"8ab60cda-aa99-4608-9011-6e9e569ca203\") " pod="openstack/root-account-create-update-mqzll" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.449939 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ab60cda-aa99-4608-9011-6e9e569ca203-operator-scripts\") pod \"root-account-create-update-mqzll\" (UID: \"8ab60cda-aa99-4608-9011-6e9e569ca203\") " pod="openstack/root-account-create-update-mqzll" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.468034 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s274d\" (UniqueName: \"kubernetes.io/projected/8ab60cda-aa99-4608-9011-6e9e569ca203-kube-api-access-s274d\") pod \"root-account-create-update-mqzll\" (UID: \"8ab60cda-aa99-4608-9011-6e9e569ca203\") " pod="openstack/root-account-create-update-mqzll" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.563576 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-mqzll" Feb 02 13:32:08 crc kubenswrapper[4846]: I0202 13:32:08.983412 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-mqzll"] Feb 02 13:32:09 crc kubenswrapper[4846]: I0202 13:32:09.256895 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-mqzll" event={"ID":"8ab60cda-aa99-4608-9011-6e9e569ca203","Type":"ContainerStarted","Data":"2c0fd3b791357a80593703b92ed8aa00a4c584b1e20dcf0f65e60ee14a6d6f58"} Feb 02 13:32:09 crc kubenswrapper[4846]: I0202 13:32:09.257246 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-mqzll" event={"ID":"8ab60cda-aa99-4608-9011-6e9e569ca203","Type":"ContainerStarted","Data":"28bf62aab22fdb26595861daba6300a10b27b2a93a197e0686c96de939b098f9"} Feb 02 13:32:09 crc kubenswrapper[4846]: I0202 13:32:09.284957 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-mqzll" podStartSLOduration=1.284936233 podStartE2EDuration="1.284936233s" podCreationTimestamp="2026-02-02 13:32:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:32:09.276816306 +0000 UTC m=+4960.505403179" watchObservedRunningTime="2026-02-02 13:32:09.284936233 +0000 UTC m=+4960.513523116" Feb 02 13:32:10 crc kubenswrapper[4846]: I0202 13:32:10.264945 4846 generic.go:334] "Generic (PLEG): container finished" podID="8ab60cda-aa99-4608-9011-6e9e569ca203" containerID="2c0fd3b791357a80593703b92ed8aa00a4c584b1e20dcf0f65e60ee14a6d6f58" exitCode=0 Feb 02 13:32:10 crc kubenswrapper[4846]: I0202 13:32:10.265007 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-mqzll" event={"ID":"8ab60cda-aa99-4608-9011-6e9e569ca203","Type":"ContainerDied","Data":"2c0fd3b791357a80593703b92ed8aa00a4c584b1e20dcf0f65e60ee14a6d6f58"} Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.272214 4846 generic.go:334] "Generic (PLEG): container finished" podID="43634f49-b272-42a2-9e88-bfa2220498b7" containerID="80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e" exitCode=0 Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.272294 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"43634f49-b272-42a2-9e88-bfa2220498b7","Type":"ContainerDied","Data":"80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e"} Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.274265 4846 generic.go:334] "Generic (PLEG): container finished" podID="9c2deb33-5a4e-4b54-9273-d2c7ddf11668" containerID="6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67" exitCode=0 Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.274360 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9c2deb33-5a4e-4b54-9273-d2c7ddf11668","Type":"ContainerDied","Data":"6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67"} Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.571083 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-mqzll" Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.701304 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s274d\" (UniqueName: \"kubernetes.io/projected/8ab60cda-aa99-4608-9011-6e9e569ca203-kube-api-access-s274d\") pod \"8ab60cda-aa99-4608-9011-6e9e569ca203\" (UID: \"8ab60cda-aa99-4608-9011-6e9e569ca203\") " Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.701512 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ab60cda-aa99-4608-9011-6e9e569ca203-operator-scripts\") pod \"8ab60cda-aa99-4608-9011-6e9e569ca203\" (UID: \"8ab60cda-aa99-4608-9011-6e9e569ca203\") " Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.701986 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ab60cda-aa99-4608-9011-6e9e569ca203-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8ab60cda-aa99-4608-9011-6e9e569ca203" (UID: "8ab60cda-aa99-4608-9011-6e9e569ca203"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.705923 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ab60cda-aa99-4608-9011-6e9e569ca203-kube-api-access-s274d" (OuterVolumeSpecName: "kube-api-access-s274d") pod "8ab60cda-aa99-4608-9011-6e9e569ca203" (UID: "8ab60cda-aa99-4608-9011-6e9e569ca203"). InnerVolumeSpecName "kube-api-access-s274d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.802810 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s274d\" (UniqueName: \"kubernetes.io/projected/8ab60cda-aa99-4608-9011-6e9e569ca203-kube-api-access-s274d\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:11 crc kubenswrapper[4846]: I0202 13:32:11.802848 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ab60cda-aa99-4608-9011-6e9e569ca203-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:12 crc kubenswrapper[4846]: I0202 13:32:12.282244 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9c2deb33-5a4e-4b54-9273-d2c7ddf11668","Type":"ContainerStarted","Data":"44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629"} Feb 02 13:32:12 crc kubenswrapper[4846]: I0202 13:32:12.282469 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:12 crc kubenswrapper[4846]: I0202 13:32:12.284079 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-mqzll" Feb 02 13:32:12 crc kubenswrapper[4846]: I0202 13:32:12.284072 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-mqzll" event={"ID":"8ab60cda-aa99-4608-9011-6e9e569ca203","Type":"ContainerDied","Data":"28bf62aab22fdb26595861daba6300a10b27b2a93a197e0686c96de939b098f9"} Feb 02 13:32:12 crc kubenswrapper[4846]: I0202 13:32:12.284454 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28bf62aab22fdb26595861daba6300a10b27b2a93a197e0686c96de939b098f9" Feb 02 13:32:12 crc kubenswrapper[4846]: I0202 13:32:12.286001 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"43634f49-b272-42a2-9e88-bfa2220498b7","Type":"ContainerStarted","Data":"fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf"} Feb 02 13:32:12 crc kubenswrapper[4846]: I0202 13:32:12.286205 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 02 13:32:12 crc kubenswrapper[4846]: I0202 13:32:12.321022 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.321002427 podStartE2EDuration="37.321002427s" podCreationTimestamp="2026-02-02 13:31:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:32:12.313733551 +0000 UTC m=+4963.542320464" watchObservedRunningTime="2026-02-02 13:32:12.321002427 +0000 UTC m=+4963.549589290" Feb 02 13:32:12 crc kubenswrapper[4846]: I0202 13:32:12.344295 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.344273803 podStartE2EDuration="37.344273803s" podCreationTimestamp="2026-02-02 13:31:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:32:12.335730074 +0000 UTC m=+4963.564316957" watchObservedRunningTime="2026-02-02 13:32:12.344273803 +0000 UTC m=+4963.572860686" Feb 02 13:32:13 crc kubenswrapper[4846]: I0202 13:32:13.429082 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:32:13 crc kubenswrapper[4846]: E0202 13:32:13.429328 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:32:25 crc kubenswrapper[4846]: I0202 13:32:25.423797 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:32:25 crc kubenswrapper[4846]: E0202 13:32:25.424849 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:32:26 crc kubenswrapper[4846]: I0202 13:32:26.674731 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:27 crc kubenswrapper[4846]: I0202 13:32:27.334791 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 02 13:32:30 crc kubenswrapper[4846]: I0202 13:32:30.771867 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54dc9c94cc-jwf4j"] Feb 02 13:32:30 crc kubenswrapper[4846]: E0202 13:32:30.772763 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab60cda-aa99-4608-9011-6e9e569ca203" containerName="mariadb-account-create-update" Feb 02 13:32:30 crc kubenswrapper[4846]: I0202 13:32:30.772783 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab60cda-aa99-4608-9011-6e9e569ca203" containerName="mariadb-account-create-update" Feb 02 13:32:30 crc kubenswrapper[4846]: I0202 13:32:30.772965 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ab60cda-aa99-4608-9011-6e9e569ca203" containerName="mariadb-account-create-update" Feb 02 13:32:30 crc kubenswrapper[4846]: I0202 13:32:30.774024 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:30 crc kubenswrapper[4846]: I0202 13:32:30.778910 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54dc9c94cc-jwf4j"] Feb 02 13:32:30 crc kubenswrapper[4846]: I0202 13:32:30.920215 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-config\") pod \"dnsmasq-dns-54dc9c94cc-jwf4j\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:30 crc kubenswrapper[4846]: I0202 13:32:30.920351 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-dns-svc\") pod \"dnsmasq-dns-54dc9c94cc-jwf4j\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:30 crc kubenswrapper[4846]: I0202 13:32:30.920396 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84xtg\" (UniqueName: \"kubernetes.io/projected/02fe5a7e-e42b-48f0-997c-13465a788818-kube-api-access-84xtg\") pod \"dnsmasq-dns-54dc9c94cc-jwf4j\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:31 crc kubenswrapper[4846]: I0202 13:32:31.022535 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-config\") pod \"dnsmasq-dns-54dc9c94cc-jwf4j\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:31 crc kubenswrapper[4846]: I0202 13:32:31.022651 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-dns-svc\") pod \"dnsmasq-dns-54dc9c94cc-jwf4j\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:31 crc kubenswrapper[4846]: I0202 13:32:31.022681 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84xtg\" (UniqueName: \"kubernetes.io/projected/02fe5a7e-e42b-48f0-997c-13465a788818-kube-api-access-84xtg\") pod \"dnsmasq-dns-54dc9c94cc-jwf4j\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:31 crc kubenswrapper[4846]: I0202 13:32:31.023479 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-dns-svc\") pod \"dnsmasq-dns-54dc9c94cc-jwf4j\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:31 crc kubenswrapper[4846]: I0202 13:32:31.023751 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-config\") pod \"dnsmasq-dns-54dc9c94cc-jwf4j\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:31 crc kubenswrapper[4846]: I0202 13:32:31.043119 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84xtg\" (UniqueName: \"kubernetes.io/projected/02fe5a7e-e42b-48f0-997c-13465a788818-kube-api-access-84xtg\") pod \"dnsmasq-dns-54dc9c94cc-jwf4j\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:31 crc kubenswrapper[4846]: I0202 13:32:31.108196 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:31 crc kubenswrapper[4846]: I0202 13:32:31.392320 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54dc9c94cc-jwf4j"] Feb 02 13:32:31 crc kubenswrapper[4846]: W0202 13:32:31.398362 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02fe5a7e_e42b_48f0_997c_13465a788818.slice/crio-b15c550733f4cad63befdd938a9d46a018f3fe30abba396d09200c03265ff980 WatchSource:0}: Error finding container b15c550733f4cad63befdd938a9d46a018f3fe30abba396d09200c03265ff980: Status 404 returned error can't find the container with id b15c550733f4cad63befdd938a9d46a018f3fe30abba396d09200c03265ff980 Feb 02 13:32:31 crc kubenswrapper[4846]: I0202 13:32:31.469263 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 13:32:31 crc kubenswrapper[4846]: I0202 13:32:31.482913 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" event={"ID":"02fe5a7e-e42b-48f0-997c-13465a788818","Type":"ContainerStarted","Data":"b15c550733f4cad63befdd938a9d46a018f3fe30abba396d09200c03265ff980"} Feb 02 13:32:32 crc kubenswrapper[4846]: I0202 13:32:32.133182 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 13:32:32 crc kubenswrapper[4846]: I0202 13:32:32.491143 4846 generic.go:334] "Generic (PLEG): container finished" podID="02fe5a7e-e42b-48f0-997c-13465a788818" containerID="7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054" exitCode=0 Feb 02 13:32:32 crc kubenswrapper[4846]: I0202 13:32:32.491187 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" event={"ID":"02fe5a7e-e42b-48f0-997c-13465a788818","Type":"ContainerDied","Data":"7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054"} Feb 02 13:32:33 crc kubenswrapper[4846]: I0202 13:32:33.499068 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" event={"ID":"02fe5a7e-e42b-48f0-997c-13465a788818","Type":"ContainerStarted","Data":"9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833"} Feb 02 13:32:33 crc kubenswrapper[4846]: I0202 13:32:33.499426 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:33 crc kubenswrapper[4846]: I0202 13:32:33.517663 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" podStartSLOduration=3.517646724 podStartE2EDuration="3.517646724s" podCreationTimestamp="2026-02-02 13:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:32:33.514354274 +0000 UTC m=+4984.742941147" watchObservedRunningTime="2026-02-02 13:32:33.517646724 +0000 UTC m=+4984.746233587" Feb 02 13:32:35 crc kubenswrapper[4846]: I0202 13:32:35.368193 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="43634f49-b272-42a2-9e88-bfa2220498b7" containerName="rabbitmq" containerID="cri-o://fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf" gracePeriod=604797 Feb 02 13:32:35 crc kubenswrapper[4846]: I0202 13:32:35.799124 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="9c2deb33-5a4e-4b54-9273-d2c7ddf11668" containerName="rabbitmq" containerID="cri-o://44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629" gracePeriod=604797 Feb 02 13:32:36 crc kubenswrapper[4846]: I0202 13:32:36.669871 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9c2deb33-5a4e-4b54-9273-d2c7ddf11668" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.250:5671: connect: connection refused" Feb 02 13:32:37 crc kubenswrapper[4846]: I0202 13:32:37.333019 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="43634f49-b272-42a2-9e88-bfa2220498b7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.251:5671: connect: connection refused" Feb 02 13:32:40 crc kubenswrapper[4846]: I0202 13:32:40.423373 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:32:40 crc kubenswrapper[4846]: E0202 13:32:40.423971 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.109890 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.165779 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-589cf688cc-vdpcl"] Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.166005 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" podUID="0247731c-e91f-45f2-8106-69b3e3c55dc6" containerName="dnsmasq-dns" containerID="cri-o://1df810e40e64a52836ac0e8e3b00831c733ffeb575ce07c24efa8dd191aff9ab" gracePeriod=10 Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.566014 4846 generic.go:334] "Generic (PLEG): container finished" podID="0247731c-e91f-45f2-8106-69b3e3c55dc6" containerID="1df810e40e64a52836ac0e8e3b00831c733ffeb575ce07c24efa8dd191aff9ab" exitCode=0 Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.566088 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" event={"ID":"0247731c-e91f-45f2-8106-69b3e3c55dc6","Type":"ContainerDied","Data":"1df810e40e64a52836ac0e8e3b00831c733ffeb575ce07c24efa8dd191aff9ab"} Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.566294 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" event={"ID":"0247731c-e91f-45f2-8106-69b3e3c55dc6","Type":"ContainerDied","Data":"61787784d5c0f56464be566c43bd63d3905271e40bbf71482384b7fc0af5f757"} Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.566308 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61787784d5c0f56464be566c43bd63d3905271e40bbf71482384b7fc0af5f757" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.614401 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.778999 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-dns-svc\") pod \"0247731c-e91f-45f2-8106-69b3e3c55dc6\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.779049 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-config\") pod \"0247731c-e91f-45f2-8106-69b3e3c55dc6\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.779090 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjnjv\" (UniqueName: \"kubernetes.io/projected/0247731c-e91f-45f2-8106-69b3e3c55dc6-kube-api-access-wjnjv\") pod \"0247731c-e91f-45f2-8106-69b3e3c55dc6\" (UID: \"0247731c-e91f-45f2-8106-69b3e3c55dc6\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.800538 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0247731c-e91f-45f2-8106-69b3e3c55dc6-kube-api-access-wjnjv" (OuterVolumeSpecName: "kube-api-access-wjnjv") pod "0247731c-e91f-45f2-8106-69b3e3c55dc6" (UID: "0247731c-e91f-45f2-8106-69b3e3c55dc6"). InnerVolumeSpecName "kube-api-access-wjnjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.821100 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-config" (OuterVolumeSpecName: "config") pod "0247731c-e91f-45f2-8106-69b3e3c55dc6" (UID: "0247731c-e91f-45f2-8106-69b3e3c55dc6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.830487 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0247731c-e91f-45f2-8106-69b3e3c55dc6" (UID: "0247731c-e91f-45f2-8106-69b3e3c55dc6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.860121 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.883344 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjnjv\" (UniqueName: \"kubernetes.io/projected/0247731c-e91f-45f2-8106-69b3e3c55dc6-kube-api-access-wjnjv\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.883382 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.883396 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0247731c-e91f-45f2-8106-69b3e3c55dc6-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.984936 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-config-data\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985260 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-plugins-conf\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985340 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-plugins\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985360 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-confd\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985378 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43634f49-b272-42a2-9e88-bfa2220498b7-pod-info\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985405 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43634f49-b272-42a2-9e88-bfa2220498b7-erlang-cookie-secret\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985565 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985605 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbzr8\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-kube-api-access-lbzr8\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985636 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-erlang-cookie\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985674 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-server-conf\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985708 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-tls\") pod \"43634f49-b272-42a2-9e88-bfa2220498b7\" (UID: \"43634f49-b272-42a2-9e88-bfa2220498b7\") " Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.985877 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.986167 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.986541 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.987457 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.988429 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.989814 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43634f49-b272-42a2-9e88-bfa2220498b7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.990138 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-kube-api-access-lbzr8" (OuterVolumeSpecName: "kube-api-access-lbzr8") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "kube-api-access-lbzr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:32:41 crc kubenswrapper[4846]: I0202 13:32:41.990667 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/43634f49-b272-42a2-9e88-bfa2220498b7-pod-info" (OuterVolumeSpecName: "pod-info") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.000408 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1" (OuterVolumeSpecName: "persistence") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "pvc-63c31d1b-8df9-4273-9675-16d5003068d1". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.012578 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-config-data" (OuterVolumeSpecName: "config-data") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.022634 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-server-conf" (OuterVolumeSpecName: "server-conf") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.073410 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "43634f49-b272-42a2-9e88-bfa2220498b7" (UID: "43634f49-b272-42a2-9e88-bfa2220498b7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.088626 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.088676 4846 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.088689 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.088700 4846 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43634f49-b272-42a2-9e88-bfa2220498b7-pod-info\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.088713 4846 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43634f49-b272-42a2-9e88-bfa2220498b7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.088762 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-63c31d1b-8df9-4273-9675-16d5003068d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") on node \"crc\" " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.088776 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbzr8\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-kube-api-access-lbzr8\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.088788 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.088798 4846 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43634f49-b272-42a2-9e88-bfa2220498b7-server-conf\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.088808 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43634f49-b272-42a2-9e88-bfa2220498b7-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.109581 4846 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.109758 4846 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-63c31d1b-8df9-4273-9675-16d5003068d1" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1") on node "crc" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.189937 4846 reconciler_common.go:293] "Volume detached for volume \"pvc-63c31d1b-8df9-4273-9675-16d5003068d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.211275 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.391558 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-pod-info\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.391650 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdkk6\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-kube-api-access-zdkk6\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.391689 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-erlang-cookie-secret\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.391737 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-config-data\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.391766 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-confd\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.391785 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-plugins\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.391814 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-erlang-cookie\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.391986 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.392016 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-tls\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.392032 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-plugins-conf\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.392052 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-server-conf\") pod \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\" (UID: \"9c2deb33-5a4e-4b54-9273-d2c7ddf11668\") " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.392762 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.392794 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.392952 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.395185 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.395315 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-kube-api-access-zdkk6" (OuterVolumeSpecName: "kube-api-access-zdkk6") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "kube-api-access-zdkk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.397123 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.403893 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-pod-info" (OuterVolumeSpecName: "pod-info") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.404219 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82" (OuterVolumeSpecName: "persistence") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.410201 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-config-data" (OuterVolumeSpecName: "config-data") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.428613 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-server-conf" (OuterVolumeSpecName: "server-conf") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.459808 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9c2deb33-5a4e-4b54-9273-d2c7ddf11668" (UID: "9c2deb33-5a4e-4b54-9273-d2c7ddf11668"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493835 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493867 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493903 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") on node \"crc\" " Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493916 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493925 4846 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493934 4846 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-server-conf\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493942 4846 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-pod-info\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493952 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdkk6\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-kube-api-access-zdkk6\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493961 4846 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493969 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.493977 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9c2deb33-5a4e-4b54-9273-d2c7ddf11668-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.526026 4846 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.526216 4846 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82") on node "crc" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.574451 4846 generic.go:334] "Generic (PLEG): container finished" podID="43634f49-b272-42a2-9e88-bfa2220498b7" containerID="fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf" exitCode=0 Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.574515 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.574523 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"43634f49-b272-42a2-9e88-bfa2220498b7","Type":"ContainerDied","Data":"fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf"} Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.574548 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"43634f49-b272-42a2-9e88-bfa2220498b7","Type":"ContainerDied","Data":"a88bd0ff8ecdc22983e52c8875d681630f388cbf2a0fdb987d31e1ba9cc4ae63"} Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.574564 4846 scope.go:117] "RemoveContainer" containerID="fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.578672 4846 generic.go:334] "Generic (PLEG): container finished" podID="9c2deb33-5a4e-4b54-9273-d2c7ddf11668" containerID="44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629" exitCode=0 Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.578742 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.578772 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-589cf688cc-vdpcl" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.578780 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9c2deb33-5a4e-4b54-9273-d2c7ddf11668","Type":"ContainerDied","Data":"44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629"} Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.578821 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9c2deb33-5a4e-4b54-9273-d2c7ddf11668","Type":"ContainerDied","Data":"03cf2e6fc8c9b4194049af615337492d2d8166ff0e8ecf68298f9977b0477d9a"} Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.594995 4846 reconciler_common.go:293] "Volume detached for volume \"pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") on node \"crc\" DevicePath \"\"" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.606858 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.618834 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.624810 4846 scope.go:117] "RemoveContainer" containerID="80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.628573 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-589cf688cc-vdpcl"] Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.634806 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-589cf688cc-vdpcl"] Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.640295 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 13:32:42 crc kubenswrapper[4846]: E0202 13:32:42.640595 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43634f49-b272-42a2-9e88-bfa2220498b7" containerName="rabbitmq" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.640606 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="43634f49-b272-42a2-9e88-bfa2220498b7" containerName="rabbitmq" Feb 02 13:32:42 crc kubenswrapper[4846]: E0202 13:32:42.640624 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c2deb33-5a4e-4b54-9273-d2c7ddf11668" containerName="rabbitmq" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.640629 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c2deb33-5a4e-4b54-9273-d2c7ddf11668" containerName="rabbitmq" Feb 02 13:32:42 crc kubenswrapper[4846]: E0202 13:32:42.640642 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43634f49-b272-42a2-9e88-bfa2220498b7" containerName="setup-container" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.641124 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="43634f49-b272-42a2-9e88-bfa2220498b7" containerName="setup-container" Feb 02 13:32:42 crc kubenswrapper[4846]: E0202 13:32:42.641147 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0247731c-e91f-45f2-8106-69b3e3c55dc6" containerName="dnsmasq-dns" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.641153 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0247731c-e91f-45f2-8106-69b3e3c55dc6" containerName="dnsmasq-dns" Feb 02 13:32:42 crc kubenswrapper[4846]: E0202 13:32:42.641170 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c2deb33-5a4e-4b54-9273-d2c7ddf11668" containerName="setup-container" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.641177 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c2deb33-5a4e-4b54-9273-d2c7ddf11668" containerName="setup-container" Feb 02 13:32:42 crc kubenswrapper[4846]: E0202 13:32:42.641189 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0247731c-e91f-45f2-8106-69b3e3c55dc6" containerName="init" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.641196 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0247731c-e91f-45f2-8106-69b3e3c55dc6" containerName="init" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.641344 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0247731c-e91f-45f2-8106-69b3e3c55dc6" containerName="dnsmasq-dns" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.641354 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="43634f49-b272-42a2-9e88-bfa2220498b7" containerName="rabbitmq" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.641368 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c2deb33-5a4e-4b54-9273-d2c7ddf11668" containerName="rabbitmq" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.642212 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.647440 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-94sh9" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.647966 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.648161 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.648373 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.648598 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.648850 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.649644 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.657085 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.667155 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.677412 4846 scope.go:117] "RemoveContainer" containerID="fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf" Feb 02 13:32:42 crc kubenswrapper[4846]: E0202 13:32:42.679270 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf\": container with ID starting with fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf not found: ID does not exist" containerID="fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.679403 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf"} err="failed to get container status \"fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf\": rpc error: code = NotFound desc = could not find container \"fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf\": container with ID starting with fe7d2c5530264ae2008b714ba861e505d3da76198c2e4336b80a98e6be7eefbf not found: ID does not exist" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.679489 4846 scope.go:117] "RemoveContainer" containerID="80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e" Feb 02 13:32:42 crc kubenswrapper[4846]: E0202 13:32:42.681663 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e\": container with ID starting with 80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e not found: ID does not exist" containerID="80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.681792 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e"} err="failed to get container status \"80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e\": rpc error: code = NotFound desc = could not find container \"80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e\": container with ID starting with 80537f57e666a35fda24d2b8b5f885faddf8ea3102296eaef4210dd8d8eb810e not found: ID does not exist" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.681888 4846 scope.go:117] "RemoveContainer" containerID="44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.686824 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.706127 4846 scope.go:117] "RemoveContainer" containerID="6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.718059 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.721487 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.732266 4846 scope.go:117] "RemoveContainer" containerID="44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.732392 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.732394 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.732494 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 02 13:32:42 crc kubenswrapper[4846]: E0202 13:32:42.732764 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629\": container with ID starting with 44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629 not found: ID does not exist" containerID="44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.732808 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629"} err="failed to get container status \"44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629\": rpc error: code = NotFound desc = could not find container \"44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629\": container with ID starting with 44131608ce10074b08f4ade0d4cebb8b1f67e98c8e73b285bbc140fc7b223629 not found: ID does not exist" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.732837 4846 scope.go:117] "RemoveContainer" containerID="6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.732975 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.733040 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-s6zvs" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.733159 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 02 13:32:42 crc kubenswrapper[4846]: E0202 13:32:42.733186 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67\": container with ID starting with 6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67 not found: ID does not exist" containerID="6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.733205 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67"} err="failed to get container status \"6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67\": rpc error: code = NotFound desc = could not find container \"6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67\": container with ID starting with 6371cac9453ce42d61bed609f463e3ca28709453db45695023c9f83cb6308b67 not found: ID does not exist" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.733269 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.737500 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803298 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803365 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aaa5fe94-6e26-4ef5-884c-b0731723c243-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803388 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803408 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803429 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803465 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79lbn\" (UniqueName: \"kubernetes.io/projected/aaa5fe94-6e26-4ef5-884c-b0731723c243-kube-api-access-79lbn\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803521 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803538 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803673 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-config-data\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803741 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803803 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803829 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803873 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aaa5fe94-6e26-4ef5-884c-b0731723c243-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803904 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wkln\" (UniqueName: \"kubernetes.io/projected/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-kube-api-access-8wkln\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803940 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803957 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aaa5fe94-6e26-4ef5-884c-b0731723c243-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.803981 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-63c31d1b-8df9-4273-9675-16d5003068d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.804154 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.804225 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.804253 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aaa5fe94-6e26-4ef5-884c-b0731723c243-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.804278 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.804304 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aaa5fe94-6e26-4ef5-884c-b0731723c243-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906044 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906128 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aaa5fe94-6e26-4ef5-884c-b0731723c243-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906145 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906161 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906182 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906206 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79lbn\" (UniqueName: \"kubernetes.io/projected/aaa5fe94-6e26-4ef5-884c-b0731723c243-kube-api-access-79lbn\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906221 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906241 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906258 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-config-data\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906279 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906307 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906341 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906374 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aaa5fe94-6e26-4ef5-884c-b0731723c243-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906400 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wkln\" (UniqueName: \"kubernetes.io/projected/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-kube-api-access-8wkln\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906432 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aaa5fe94-6e26-4ef5-884c-b0731723c243-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906471 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-63c31d1b-8df9-4273-9675-16d5003068d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906492 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906513 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906530 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aaa5fe94-6e26-4ef5-884c-b0731723c243-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906548 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.906565 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aaa5fe94-6e26-4ef5-884c-b0731723c243-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.907173 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.907660 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.907856 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.908147 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.908153 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aaa5fe94-6e26-4ef5-884c-b0731723c243-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.908207 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-config-data\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.909548 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.910915 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.911956 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.912229 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.913125 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.913177 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aaa5fe94-6e26-4ef5-884c-b0731723c243-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.913988 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aaa5fe94-6e26-4ef5-884c-b0731723c243-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.914235 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aaa5fe94-6e26-4ef5-884c-b0731723c243-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.914576 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.914600 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.914613 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c5cf7ac03c1e4654a9eb7f16be0bafe95391676eee61ca36eecad9f3b956cc3e/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.914623 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-63c31d1b-8df9-4273-9675-16d5003068d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/486b899c90445695a6faaf1f10d5b80276ce921cece83762353508bece56da44/globalmount\"" pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.914840 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aaa5fe94-6e26-4ef5-884c-b0731723c243-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.915046 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.916049 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.917164 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aaa5fe94-6e26-4ef5-884c-b0731723c243-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.925874 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wkln\" (UniqueName: \"kubernetes.io/projected/3bf0590a-eb90-4b29-baaf-d4eaa8404e50-kube-api-access-8wkln\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.942785 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bf8ef551-62f2-4291-b619-4ed6e28a8b82\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.946037 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79lbn\" (UniqueName: \"kubernetes.io/projected/aaa5fe94-6e26-4ef5-884c-b0731723c243-kube-api-access-79lbn\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaa5fe94-6e26-4ef5-884c-b0731723c243\") " pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.949899 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-63c31d1b-8df9-4273-9675-16d5003068d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-63c31d1b-8df9-4273-9675-16d5003068d1\") pod \"rabbitmq-server-0\" (UID: \"3bf0590a-eb90-4b29-baaf-d4eaa8404e50\") " pod="openstack/rabbitmq-server-0" Feb 02 13:32:42 crc kubenswrapper[4846]: I0202 13:32:42.988056 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 02 13:32:43 crc kubenswrapper[4846]: I0202 13:32:43.051211 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:32:43 crc kubenswrapper[4846]: I0202 13:32:43.390168 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 02 13:32:43 crc kubenswrapper[4846]: W0202 13:32:43.398941 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bf0590a_eb90_4b29_baaf_d4eaa8404e50.slice/crio-054b3d956dd3cc7740414916be6ff4ce0e0c881937fd681c9a57972412896e1d WatchSource:0}: Error finding container 054b3d956dd3cc7740414916be6ff4ce0e0c881937fd681c9a57972412896e1d: Status 404 returned error can't find the container with id 054b3d956dd3cc7740414916be6ff4ce0e0c881937fd681c9a57972412896e1d Feb 02 13:32:43 crc kubenswrapper[4846]: I0202 13:32:43.440769 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0247731c-e91f-45f2-8106-69b3e3c55dc6" path="/var/lib/kubelet/pods/0247731c-e91f-45f2-8106-69b3e3c55dc6/volumes" Feb 02 13:32:43 crc kubenswrapper[4846]: I0202 13:32:43.441571 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43634f49-b272-42a2-9e88-bfa2220498b7" path="/var/lib/kubelet/pods/43634f49-b272-42a2-9e88-bfa2220498b7/volumes" Feb 02 13:32:43 crc kubenswrapper[4846]: I0202 13:32:43.442891 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c2deb33-5a4e-4b54-9273-d2c7ddf11668" path="/var/lib/kubelet/pods/9c2deb33-5a4e-4b54-9273-d2c7ddf11668/volumes" Feb 02 13:32:43 crc kubenswrapper[4846]: I0202 13:32:43.497122 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 02 13:32:43 crc kubenswrapper[4846]: W0202 13:32:43.505157 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaa5fe94_6e26_4ef5_884c_b0731723c243.slice/crio-2399920ec140c41483ede0e13c58015d716846ab14873d8a1465501ae84f37dc WatchSource:0}: Error finding container 2399920ec140c41483ede0e13c58015d716846ab14873d8a1465501ae84f37dc: Status 404 returned error can't find the container with id 2399920ec140c41483ede0e13c58015d716846ab14873d8a1465501ae84f37dc Feb 02 13:32:43 crc kubenswrapper[4846]: I0202 13:32:43.588983 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"aaa5fe94-6e26-4ef5-884c-b0731723c243","Type":"ContainerStarted","Data":"2399920ec140c41483ede0e13c58015d716846ab14873d8a1465501ae84f37dc"} Feb 02 13:32:43 crc kubenswrapper[4846]: I0202 13:32:43.590132 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3bf0590a-eb90-4b29-baaf-d4eaa8404e50","Type":"ContainerStarted","Data":"054b3d956dd3cc7740414916be6ff4ce0e0c881937fd681c9a57972412896e1d"} Feb 02 13:32:45 crc kubenswrapper[4846]: I0202 13:32:45.610821 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"aaa5fe94-6e26-4ef5-884c-b0731723c243","Type":"ContainerStarted","Data":"ca8314c1d609cc11fcb957f6a49d0462244c5a26ba0a29b795d0226b0d2aff91"} Feb 02 13:32:45 crc kubenswrapper[4846]: I0202 13:32:45.613517 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3bf0590a-eb90-4b29-baaf-d4eaa8404e50","Type":"ContainerStarted","Data":"c312069e86fdb4ce779f0f4abd1196709241bcc8a6af56445cf9ddb0ba6093a2"} Feb 02 13:32:53 crc kubenswrapper[4846]: I0202 13:32:53.424166 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:32:53 crc kubenswrapper[4846]: E0202 13:32:53.425295 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:33:05 crc kubenswrapper[4846]: I0202 13:33:05.423795 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:33:05 crc kubenswrapper[4846]: E0202 13:33:05.424714 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:33:17 crc kubenswrapper[4846]: I0202 13:33:17.844664 4846 generic.go:334] "Generic (PLEG): container finished" podID="aaa5fe94-6e26-4ef5-884c-b0731723c243" containerID="ca8314c1d609cc11fcb957f6a49d0462244c5a26ba0a29b795d0226b0d2aff91" exitCode=0 Feb 02 13:33:17 crc kubenswrapper[4846]: I0202 13:33:17.844767 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"aaa5fe94-6e26-4ef5-884c-b0731723c243","Type":"ContainerDied","Data":"ca8314c1d609cc11fcb957f6a49d0462244c5a26ba0a29b795d0226b0d2aff91"} Feb 02 13:33:17 crc kubenswrapper[4846]: I0202 13:33:17.849152 4846 generic.go:334] "Generic (PLEG): container finished" podID="3bf0590a-eb90-4b29-baaf-d4eaa8404e50" containerID="c312069e86fdb4ce779f0f4abd1196709241bcc8a6af56445cf9ddb0ba6093a2" exitCode=0 Feb 02 13:33:17 crc kubenswrapper[4846]: I0202 13:33:17.849201 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3bf0590a-eb90-4b29-baaf-d4eaa8404e50","Type":"ContainerDied","Data":"c312069e86fdb4ce779f0f4abd1196709241bcc8a6af56445cf9ddb0ba6093a2"} Feb 02 13:33:18 crc kubenswrapper[4846]: I0202 13:33:18.423071 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:33:18 crc kubenswrapper[4846]: E0202 13:33:18.423605 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:33:18 crc kubenswrapper[4846]: I0202 13:33:18.858338 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"aaa5fe94-6e26-4ef5-884c-b0731723c243","Type":"ContainerStarted","Data":"a92c41c2c12316a3f11213db19c025504560d2fbc92a1a59378353c3d2079478"} Feb 02 13:33:18 crc kubenswrapper[4846]: I0202 13:33:18.858534 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:33:18 crc kubenswrapper[4846]: I0202 13:33:18.861583 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3bf0590a-eb90-4b29-baaf-d4eaa8404e50","Type":"ContainerStarted","Data":"39e40fc9029247b00ddf40055f11634f510d2be1cb540af4c5ac047436486d34"} Feb 02 13:33:18 crc kubenswrapper[4846]: I0202 13:33:18.861877 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 02 13:33:18 crc kubenswrapper[4846]: I0202 13:33:18.888749 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.888725309 podStartE2EDuration="36.888725309s" podCreationTimestamp="2026-02-02 13:32:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:33:18.881014492 +0000 UTC m=+5030.109601355" watchObservedRunningTime="2026-02-02 13:33:18.888725309 +0000 UTC m=+5030.117312182" Feb 02 13:33:18 crc kubenswrapper[4846]: I0202 13:33:18.906952 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.906932641 podStartE2EDuration="36.906932641s" podCreationTimestamp="2026-02-02 13:32:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:33:18.905818935 +0000 UTC m=+5030.134405798" watchObservedRunningTime="2026-02-02 13:33:18.906932641 +0000 UTC m=+5030.135519504" Feb 02 13:33:31 crc kubenswrapper[4846]: I0202 13:33:31.424096 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:33:31 crc kubenswrapper[4846]: E0202 13:33:31.425231 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:33:32 crc kubenswrapper[4846]: I0202 13:33:32.990796 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 02 13:33:33 crc kubenswrapper[4846]: I0202 13:33:33.054053 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 02 13:33:37 crc kubenswrapper[4846]: I0202 13:33:37.925780 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Feb 02 13:33:37 crc kubenswrapper[4846]: I0202 13:33:37.942500 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:33:37 crc kubenswrapper[4846]: I0202 13:33:37.942647 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:33:37 crc kubenswrapper[4846]: I0202 13:33:37.944882 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5jfpr" Feb 02 13:33:38 crc kubenswrapper[4846]: I0202 13:33:38.059539 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh9sd\" (UniqueName: \"kubernetes.io/projected/6a9498fd-a304-4c6b-85a5-d9c02350b75b-kube-api-access-wh9sd\") pod \"mariadb-client\" (UID: \"6a9498fd-a304-4c6b-85a5-d9c02350b75b\") " pod="openstack/mariadb-client" Feb 02 13:33:38 crc kubenswrapper[4846]: I0202 13:33:38.161777 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh9sd\" (UniqueName: \"kubernetes.io/projected/6a9498fd-a304-4c6b-85a5-d9c02350b75b-kube-api-access-wh9sd\") pod \"mariadb-client\" (UID: \"6a9498fd-a304-4c6b-85a5-d9c02350b75b\") " pod="openstack/mariadb-client" Feb 02 13:33:38 crc kubenswrapper[4846]: I0202 13:33:38.180673 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh9sd\" (UniqueName: \"kubernetes.io/projected/6a9498fd-a304-4c6b-85a5-d9c02350b75b-kube-api-access-wh9sd\") pod \"mariadb-client\" (UID: \"6a9498fd-a304-4c6b-85a5-d9c02350b75b\") " pod="openstack/mariadb-client" Feb 02 13:33:38 crc kubenswrapper[4846]: I0202 13:33:38.265206 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:33:38 crc kubenswrapper[4846]: I0202 13:33:38.768993 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:33:38 crc kubenswrapper[4846]: I0202 13:33:38.773395 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 13:33:39 crc kubenswrapper[4846]: I0202 13:33:39.022541 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6a9498fd-a304-4c6b-85a5-d9c02350b75b","Type":"ContainerStarted","Data":"71419b9f52ae1368fdc0cbab14d09c4f16279c9ad9039616fb713221dde1dd5f"} Feb 02 13:33:40 crc kubenswrapper[4846]: I0202 13:33:40.031984 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6a9498fd-a304-4c6b-85a5-d9c02350b75b","Type":"ContainerStarted","Data":"c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d"} Feb 02 13:33:40 crc kubenswrapper[4846]: I0202 13:33:40.048474 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client" podStartSLOduration=2.41077281 podStartE2EDuration="3.048442627s" podCreationTimestamp="2026-02-02 13:33:37 +0000 UTC" firstStartedPulling="2026-02-02 13:33:38.773036313 +0000 UTC m=+5050.001623176" lastFinishedPulling="2026-02-02 13:33:39.41070613 +0000 UTC m=+5050.639292993" observedRunningTime="2026-02-02 13:33:40.047001142 +0000 UTC m=+5051.275588035" watchObservedRunningTime="2026-02-02 13:33:40.048442627 +0000 UTC m=+5051.277029530" Feb 02 13:33:45 crc kubenswrapper[4846]: I0202 13:33:45.423578 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:33:45 crc kubenswrapper[4846]: E0202 13:33:45.424160 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:33:54 crc kubenswrapper[4846]: I0202 13:33:54.096172 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:33:54 crc kubenswrapper[4846]: I0202 13:33:54.098977 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-client" podUID="6a9498fd-a304-4c6b-85a5-d9c02350b75b" containerName="mariadb-client" containerID="cri-o://c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d" gracePeriod=30 Feb 02 13:33:54 crc kubenswrapper[4846]: I0202 13:33:54.626477 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:33:54 crc kubenswrapper[4846]: I0202 13:33:54.765934 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh9sd\" (UniqueName: \"kubernetes.io/projected/6a9498fd-a304-4c6b-85a5-d9c02350b75b-kube-api-access-wh9sd\") pod \"6a9498fd-a304-4c6b-85a5-d9c02350b75b\" (UID: \"6a9498fd-a304-4c6b-85a5-d9c02350b75b\") " Feb 02 13:33:54 crc kubenswrapper[4846]: I0202 13:33:54.772204 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a9498fd-a304-4c6b-85a5-d9c02350b75b-kube-api-access-wh9sd" (OuterVolumeSpecName: "kube-api-access-wh9sd") pod "6a9498fd-a304-4c6b-85a5-d9c02350b75b" (UID: "6a9498fd-a304-4c6b-85a5-d9c02350b75b"). InnerVolumeSpecName "kube-api-access-wh9sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:33:54 crc kubenswrapper[4846]: I0202 13:33:54.869286 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh9sd\" (UniqueName: \"kubernetes.io/projected/6a9498fd-a304-4c6b-85a5-d9c02350b75b-kube-api-access-wh9sd\") on node \"crc\" DevicePath \"\"" Feb 02 13:33:55 crc kubenswrapper[4846]: I0202 13:33:55.132458 4846 generic.go:334] "Generic (PLEG): container finished" podID="6a9498fd-a304-4c6b-85a5-d9c02350b75b" containerID="c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d" exitCode=143 Feb 02 13:33:55 crc kubenswrapper[4846]: I0202 13:33:55.132510 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6a9498fd-a304-4c6b-85a5-d9c02350b75b","Type":"ContainerDied","Data":"c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d"} Feb 02 13:33:55 crc kubenswrapper[4846]: I0202 13:33:55.132557 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6a9498fd-a304-4c6b-85a5-d9c02350b75b","Type":"ContainerDied","Data":"71419b9f52ae1368fdc0cbab14d09c4f16279c9ad9039616fb713221dde1dd5f"} Feb 02 13:33:55 crc kubenswrapper[4846]: I0202 13:33:55.132586 4846 scope.go:117] "RemoveContainer" containerID="c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d" Feb 02 13:33:55 crc kubenswrapper[4846]: I0202 13:33:55.132581 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:33:55 crc kubenswrapper[4846]: I0202 13:33:55.170276 4846 scope.go:117] "RemoveContainer" containerID="c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d" Feb 02 13:33:55 crc kubenswrapper[4846]: E0202 13:33:55.172821 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d\": container with ID starting with c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d not found: ID does not exist" containerID="c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d" Feb 02 13:33:55 crc kubenswrapper[4846]: I0202 13:33:55.172914 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d"} err="failed to get container status \"c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d\": rpc error: code = NotFound desc = could not find container \"c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d\": container with ID starting with c86177fb55ead6e3951132967c4cba341418d2ead060209b5ebeb99a3119935d not found: ID does not exist" Feb 02 13:33:55 crc kubenswrapper[4846]: I0202 13:33:55.177883 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:33:55 crc kubenswrapper[4846]: I0202 13:33:55.187361 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:33:55 crc kubenswrapper[4846]: I0202 13:33:55.437785 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a9498fd-a304-4c6b-85a5-d9c02350b75b" path="/var/lib/kubelet/pods/6a9498fd-a304-4c6b-85a5-d9c02350b75b/volumes" Feb 02 13:34:00 crc kubenswrapper[4846]: I0202 13:34:00.423496 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:34:00 crc kubenswrapper[4846]: E0202 13:34:00.424221 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:34:14 crc kubenswrapper[4846]: I0202 13:34:14.423840 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:34:14 crc kubenswrapper[4846]: E0202 13:34:14.424596 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:34:26 crc kubenswrapper[4846]: I0202 13:34:26.423704 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:34:26 crc kubenswrapper[4846]: E0202 13:34:26.424379 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:34:38 crc kubenswrapper[4846]: I0202 13:34:38.424397 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:34:38 crc kubenswrapper[4846]: E0202 13:34:38.425162 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:34:50 crc kubenswrapper[4846]: I0202 13:34:50.424115 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:34:50 crc kubenswrapper[4846]: E0202 13:34:50.425051 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:35:01 crc kubenswrapper[4846]: I0202 13:35:01.424211 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:35:01 crc kubenswrapper[4846]: E0202 13:35:01.425055 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:35:12 crc kubenswrapper[4846]: I0202 13:35:12.424148 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:35:12 crc kubenswrapper[4846]: E0202 13:35:12.425582 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:35:24 crc kubenswrapper[4846]: I0202 13:35:24.423924 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:35:24 crc kubenswrapper[4846]: E0202 13:35:24.424817 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:35:37 crc kubenswrapper[4846]: I0202 13:35:37.423918 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:35:37 crc kubenswrapper[4846]: E0202 13:35:37.424681 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:35:37 crc kubenswrapper[4846]: I0202 13:35:37.935464 4846 scope.go:117] "RemoveContainer" containerID="d3dfe31162f6806d3acc550e0d36502902440b56801f3a2f311ce6d90f93062c" Feb 02 13:35:48 crc kubenswrapper[4846]: I0202 13:35:48.423592 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:35:48 crc kubenswrapper[4846]: E0202 13:35:48.424524 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:36:03 crc kubenswrapper[4846]: I0202 13:36:03.423681 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:36:03 crc kubenswrapper[4846]: E0202 13:36:03.424437 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.423425 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:36:18 crc kubenswrapper[4846]: E0202 13:36:18.424531 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.559365 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mklqc"] Feb 02 13:36:18 crc kubenswrapper[4846]: E0202 13:36:18.559913 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a9498fd-a304-4c6b-85a5-d9c02350b75b" containerName="mariadb-client" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.559944 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a9498fd-a304-4c6b-85a5-d9c02350b75b" containerName="mariadb-client" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.560180 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a9498fd-a304-4c6b-85a5-d9c02350b75b" containerName="mariadb-client" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.561900 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.581868 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mklqc"] Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.699578 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-catalog-content\") pod \"certified-operators-mklqc\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.699960 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-utilities\") pod \"certified-operators-mklqc\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.700130 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms9hl\" (UniqueName: \"kubernetes.io/projected/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-kube-api-access-ms9hl\") pod \"certified-operators-mklqc\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.801048 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-utilities\") pod \"certified-operators-mklqc\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.801148 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms9hl\" (UniqueName: \"kubernetes.io/projected/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-kube-api-access-ms9hl\") pod \"certified-operators-mklqc\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.801210 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-catalog-content\") pod \"certified-operators-mklqc\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.801540 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-utilities\") pod \"certified-operators-mklqc\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.801584 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-catalog-content\") pod \"certified-operators-mklqc\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.825124 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms9hl\" (UniqueName: \"kubernetes.io/projected/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-kube-api-access-ms9hl\") pod \"certified-operators-mklqc\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:18 crc kubenswrapper[4846]: I0202 13:36:18.892926 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:19 crc kubenswrapper[4846]: I0202 13:36:19.392236 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mklqc"] Feb 02 13:36:19 crc kubenswrapper[4846]: E0202 13:36:19.700123 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f332f37_12ce_4e3f_bfc3_bec6bc76d0b7.slice/crio-6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801.scope\": RecentStats: unable to find data in memory cache]" Feb 02 13:36:20 crc kubenswrapper[4846]: I0202 13:36:20.188469 4846 generic.go:334] "Generic (PLEG): container finished" podID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerID="6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801" exitCode=0 Feb 02 13:36:20 crc kubenswrapper[4846]: I0202 13:36:20.188539 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mklqc" event={"ID":"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7","Type":"ContainerDied","Data":"6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801"} Feb 02 13:36:20 crc kubenswrapper[4846]: I0202 13:36:20.188596 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mklqc" event={"ID":"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7","Type":"ContainerStarted","Data":"0e30479f48e8c60460f280eef282e56a2f9295d6f1353c6ad07fd2b1a40f1f87"} Feb 02 13:36:22 crc kubenswrapper[4846]: I0202 13:36:22.205174 4846 generic.go:334] "Generic (PLEG): container finished" podID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerID="348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae" exitCode=0 Feb 02 13:36:22 crc kubenswrapper[4846]: I0202 13:36:22.205255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mklqc" event={"ID":"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7","Type":"ContainerDied","Data":"348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae"} Feb 02 13:36:23 crc kubenswrapper[4846]: I0202 13:36:23.214756 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mklqc" event={"ID":"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7","Type":"ContainerStarted","Data":"0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a"} Feb 02 13:36:23 crc kubenswrapper[4846]: I0202 13:36:23.238884 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mklqc" podStartSLOduration=2.700767169 podStartE2EDuration="5.238865559s" podCreationTimestamp="2026-02-02 13:36:18 +0000 UTC" firstStartedPulling="2026-02-02 13:36:20.191105642 +0000 UTC m=+5211.419692515" lastFinishedPulling="2026-02-02 13:36:22.729204042 +0000 UTC m=+5213.957790905" observedRunningTime="2026-02-02 13:36:23.231967362 +0000 UTC m=+5214.460554225" watchObservedRunningTime="2026-02-02 13:36:23.238865559 +0000 UTC m=+5214.467452422" Feb 02 13:36:28 crc kubenswrapper[4846]: I0202 13:36:28.893786 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:28 crc kubenswrapper[4846]: I0202 13:36:28.894859 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:28 crc kubenswrapper[4846]: I0202 13:36:28.946242 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:29 crc kubenswrapper[4846]: I0202 13:36:29.309650 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:29 crc kubenswrapper[4846]: I0202 13:36:29.355161 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mklqc"] Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.283752 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mklqc" podUID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerName="registry-server" containerID="cri-o://0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a" gracePeriod=2 Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.679024 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.792325 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-catalog-content\") pod \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.792387 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms9hl\" (UniqueName: \"kubernetes.io/projected/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-kube-api-access-ms9hl\") pod \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.792411 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-utilities\") pod \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\" (UID: \"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7\") " Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.794068 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-utilities" (OuterVolumeSpecName: "utilities") pod "3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" (UID: "3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.808138 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-kube-api-access-ms9hl" (OuterVolumeSpecName: "kube-api-access-ms9hl") pod "3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" (UID: "3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7"). InnerVolumeSpecName "kube-api-access-ms9hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.859920 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" (UID: "3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.893826 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.893873 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms9hl\" (UniqueName: \"kubernetes.io/projected/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-kube-api-access-ms9hl\") on node \"crc\" DevicePath \"\"" Feb 02 13:36:31 crc kubenswrapper[4846]: I0202 13:36:31.893923 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.293844 4846 generic.go:334] "Generic (PLEG): container finished" podID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerID="0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a" exitCode=0 Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.293886 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mklqc" event={"ID":"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7","Type":"ContainerDied","Data":"0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a"} Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.293914 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mklqc" event={"ID":"3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7","Type":"ContainerDied","Data":"0e30479f48e8c60460f280eef282e56a2f9295d6f1353c6ad07fd2b1a40f1f87"} Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.293932 4846 scope.go:117] "RemoveContainer" containerID="0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.293971 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mklqc" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.313004 4846 scope.go:117] "RemoveContainer" containerID="348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.334030 4846 scope.go:117] "RemoveContainer" containerID="6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.334087 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mklqc"] Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.340959 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mklqc"] Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.361608 4846 scope.go:117] "RemoveContainer" containerID="0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a" Feb 02 13:36:32 crc kubenswrapper[4846]: E0202 13:36:32.362327 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a\": container with ID starting with 0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a not found: ID does not exist" containerID="0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.362380 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a"} err="failed to get container status \"0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a\": rpc error: code = NotFound desc = could not find container \"0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a\": container with ID starting with 0121678a965a2a74a662dfa3cb75943d61ddf8eaefce62c382bbb4662d42600a not found: ID does not exist" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.362412 4846 scope.go:117] "RemoveContainer" containerID="348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae" Feb 02 13:36:32 crc kubenswrapper[4846]: E0202 13:36:32.363548 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae\": container with ID starting with 348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae not found: ID does not exist" containerID="348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.363602 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae"} err="failed to get container status \"348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae\": rpc error: code = NotFound desc = could not find container \"348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae\": container with ID starting with 348eeb724a60d2b511db71c529975ace80a0d0669e1047d2b963c9d93408f3ae not found: ID does not exist" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.363656 4846 scope.go:117] "RemoveContainer" containerID="6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801" Feb 02 13:36:32 crc kubenswrapper[4846]: E0202 13:36:32.364039 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801\": container with ID starting with 6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801 not found: ID does not exist" containerID="6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.364074 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801"} err="failed to get container status \"6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801\": rpc error: code = NotFound desc = could not find container \"6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801\": container with ID starting with 6e2983898daed6acec326f445e537748593ca80f95fee62278f509f3721eb801 not found: ID does not exist" Feb 02 13:36:32 crc kubenswrapper[4846]: I0202 13:36:32.424195 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:36:32 crc kubenswrapper[4846]: E0202 13:36:32.424392 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:36:33 crc kubenswrapper[4846]: I0202 13:36:33.433820 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" path="/var/lib/kubelet/pods/3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7/volumes" Feb 02 13:36:46 crc kubenswrapper[4846]: I0202 13:36:46.422915 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:36:46 crc kubenswrapper[4846]: E0202 13:36:46.423687 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:36:57 crc kubenswrapper[4846]: I0202 13:36:57.423037 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:36:57 crc kubenswrapper[4846]: E0202 13:36:57.423904 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:37:12 crc kubenswrapper[4846]: I0202 13:37:12.424379 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:37:13 crc kubenswrapper[4846]: I0202 13:37:13.592352 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"cc41b07cc0804d15d8ee13c8c57c6dfb00243755829e268a57694312efd7b440"} Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.562768 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Feb 02 13:37:32 crc kubenswrapper[4846]: E0202 13:37:32.564312 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerName="extract-utilities" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.564338 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerName="extract-utilities" Feb 02 13:37:32 crc kubenswrapper[4846]: E0202 13:37:32.564350 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerName="extract-content" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.564358 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerName="extract-content" Feb 02 13:37:32 crc kubenswrapper[4846]: E0202 13:37:32.564395 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerName="registry-server" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.564406 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerName="registry-server" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.564583 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f332f37-12ce-4e3f-bfc3-bec6bc76d0b7" containerName="registry-server" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.565461 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.574101 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5jfpr" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.582724 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.740172 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\") pod \"mariadb-copy-data\" (UID: \"a5799c70-3207-43d8-a2d6-495aa207445c\") " pod="openstack/mariadb-copy-data" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.740302 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phz4d\" (UniqueName: \"kubernetes.io/projected/a5799c70-3207-43d8-a2d6-495aa207445c-kube-api-access-phz4d\") pod \"mariadb-copy-data\" (UID: \"a5799c70-3207-43d8-a2d6-495aa207445c\") " pod="openstack/mariadb-copy-data" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.841466 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\") pod \"mariadb-copy-data\" (UID: \"a5799c70-3207-43d8-a2d6-495aa207445c\") " pod="openstack/mariadb-copy-data" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.841585 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phz4d\" (UniqueName: \"kubernetes.io/projected/a5799c70-3207-43d8-a2d6-495aa207445c-kube-api-access-phz4d\") pod \"mariadb-copy-data\" (UID: \"a5799c70-3207-43d8-a2d6-495aa207445c\") " pod="openstack/mariadb-copy-data" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.844316 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.844354 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\") pod \"mariadb-copy-data\" (UID: \"a5799c70-3207-43d8-a2d6-495aa207445c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/91c18bb5954371d59db90ba5b7f2d04a0fda00530f2c76a22ea03dd9477711f8/globalmount\"" pod="openstack/mariadb-copy-data" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.870780 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phz4d\" (UniqueName: \"kubernetes.io/projected/a5799c70-3207-43d8-a2d6-495aa207445c-kube-api-access-phz4d\") pod \"mariadb-copy-data\" (UID: \"a5799c70-3207-43d8-a2d6-495aa207445c\") " pod="openstack/mariadb-copy-data" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.872296 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\") pod \"mariadb-copy-data\" (UID: \"a5799c70-3207-43d8-a2d6-495aa207445c\") " pod="openstack/mariadb-copy-data" Feb 02 13:37:32 crc kubenswrapper[4846]: I0202 13:37:32.892734 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Feb 02 13:37:33 crc kubenswrapper[4846]: I0202 13:37:33.972949 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Feb 02 13:37:33 crc kubenswrapper[4846]: W0202 13:37:33.977560 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5799c70_3207_43d8_a2d6_495aa207445c.slice/crio-62d327e126d7cc804be6a8d3f410f5dcca719f1bd24600110532d97bf69c6e32 WatchSource:0}: Error finding container 62d327e126d7cc804be6a8d3f410f5dcca719f1bd24600110532d97bf69c6e32: Status 404 returned error can't find the container with id 62d327e126d7cc804be6a8d3f410f5dcca719f1bd24600110532d97bf69c6e32 Feb 02 13:37:34 crc kubenswrapper[4846]: I0202 13:37:34.737201 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"a5799c70-3207-43d8-a2d6-495aa207445c","Type":"ContainerStarted","Data":"48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e"} Feb 02 13:37:34 crc kubenswrapper[4846]: I0202 13:37:34.737555 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"a5799c70-3207-43d8-a2d6-495aa207445c","Type":"ContainerStarted","Data":"62d327e126d7cc804be6a8d3f410f5dcca719f1bd24600110532d97bf69c6e32"} Feb 02 13:37:37 crc kubenswrapper[4846]: I0202 13:37:37.491246 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=6.491223262 podStartE2EDuration="6.491223262s" podCreationTimestamp="2026-02-02 13:37:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:37:34.757120402 +0000 UTC m=+5285.985707285" watchObservedRunningTime="2026-02-02 13:37:37.491223262 +0000 UTC m=+5288.719810125" Feb 02 13:37:37 crc kubenswrapper[4846]: I0202 13:37:37.492167 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Feb 02 13:37:37 crc kubenswrapper[4846]: I0202 13:37:37.493851 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:37:37 crc kubenswrapper[4846]: I0202 13:37:37.499034 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:37:37 crc kubenswrapper[4846]: I0202 13:37:37.615429 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mznlx\" (UniqueName: \"kubernetes.io/projected/1a8e81b4-d55f-4d63-9338-60f8e6ec84a0-kube-api-access-mznlx\") pod \"mariadb-client\" (UID: \"1a8e81b4-d55f-4d63-9338-60f8e6ec84a0\") " pod="openstack/mariadb-client" Feb 02 13:37:37 crc kubenswrapper[4846]: I0202 13:37:37.717340 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mznlx\" (UniqueName: \"kubernetes.io/projected/1a8e81b4-d55f-4d63-9338-60f8e6ec84a0-kube-api-access-mznlx\") pod \"mariadb-client\" (UID: \"1a8e81b4-d55f-4d63-9338-60f8e6ec84a0\") " pod="openstack/mariadb-client" Feb 02 13:37:37 crc kubenswrapper[4846]: I0202 13:37:37.738069 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mznlx\" (UniqueName: \"kubernetes.io/projected/1a8e81b4-d55f-4d63-9338-60f8e6ec84a0-kube-api-access-mznlx\") pod \"mariadb-client\" (UID: \"1a8e81b4-d55f-4d63-9338-60f8e6ec84a0\") " pod="openstack/mariadb-client" Feb 02 13:37:37 crc kubenswrapper[4846]: I0202 13:37:37.817488 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:37:38 crc kubenswrapper[4846]: I0202 13:37:38.020701 4846 scope.go:117] "RemoveContainer" containerID="81e7147869ec2463b15a4897692271a4ef96bfb852d09ba51098763cd66b2d24" Feb 02 13:37:38 crc kubenswrapper[4846]: I0202 13:37:38.041017 4846 scope.go:117] "RemoveContainer" containerID="1df810e40e64a52836ac0e8e3b00831c733ffeb575ce07c24efa8dd191aff9ab" Feb 02 13:37:38 crc kubenswrapper[4846]: I0202 13:37:38.245215 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:37:38 crc kubenswrapper[4846]: W0202 13:37:38.247220 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a8e81b4_d55f_4d63_9338_60f8e6ec84a0.slice/crio-21ab44a046cf1a43bdfbffbe690167c8535ed3b985b9bb751b9fecd06378387c WatchSource:0}: Error finding container 21ab44a046cf1a43bdfbffbe690167c8535ed3b985b9bb751b9fecd06378387c: Status 404 returned error can't find the container with id 21ab44a046cf1a43bdfbffbe690167c8535ed3b985b9bb751b9fecd06378387c Feb 02 13:37:38 crc kubenswrapper[4846]: I0202 13:37:38.765387 4846 generic.go:334] "Generic (PLEG): container finished" podID="1a8e81b4-d55f-4d63-9338-60f8e6ec84a0" containerID="90d90ae85de5986e2fa7d084c94a411a460ca65f71721f007924a4127ed2274c" exitCode=0 Feb 02 13:37:38 crc kubenswrapper[4846]: I0202 13:37:38.765444 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"1a8e81b4-d55f-4d63-9338-60f8e6ec84a0","Type":"ContainerDied","Data":"90d90ae85de5986e2fa7d084c94a411a460ca65f71721f007924a4127ed2274c"} Feb 02 13:37:38 crc kubenswrapper[4846]: I0202 13:37:38.765477 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"1a8e81b4-d55f-4d63-9338-60f8e6ec84a0","Type":"ContainerStarted","Data":"21ab44a046cf1a43bdfbffbe690167c8535ed3b985b9bb751b9fecd06378387c"} Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.099028 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.126806 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_1a8e81b4-d55f-4d63-9338-60f8e6ec84a0/mariadb-client/0.log" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.155152 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.160819 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.256247 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mznlx\" (UniqueName: \"kubernetes.io/projected/1a8e81b4-d55f-4d63-9338-60f8e6ec84a0-kube-api-access-mznlx\") pod \"1a8e81b4-d55f-4d63-9338-60f8e6ec84a0\" (UID: \"1a8e81b4-d55f-4d63-9338-60f8e6ec84a0\") " Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.272910 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a8e81b4-d55f-4d63-9338-60f8e6ec84a0-kube-api-access-mznlx" (OuterVolumeSpecName: "kube-api-access-mznlx") pod "1a8e81b4-d55f-4d63-9338-60f8e6ec84a0" (UID: "1a8e81b4-d55f-4d63-9338-60f8e6ec84a0"). InnerVolumeSpecName "kube-api-access-mznlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.281853 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Feb 02 13:37:40 crc kubenswrapper[4846]: E0202 13:37:40.282179 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a8e81b4-d55f-4d63-9338-60f8e6ec84a0" containerName="mariadb-client" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.282206 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a8e81b4-d55f-4d63-9338-60f8e6ec84a0" containerName="mariadb-client" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.282383 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a8e81b4-d55f-4d63-9338-60f8e6ec84a0" containerName="mariadb-client" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.282892 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.299564 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.359515 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxfhw\" (UniqueName: \"kubernetes.io/projected/0c2b8853-e439-4a04-a39d-500772139657-kube-api-access-bxfhw\") pod \"mariadb-client\" (UID: \"0c2b8853-e439-4a04-a39d-500772139657\") " pod="openstack/mariadb-client" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.360485 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mznlx\" (UniqueName: \"kubernetes.io/projected/1a8e81b4-d55f-4d63-9338-60f8e6ec84a0-kube-api-access-mznlx\") on node \"crc\" DevicePath \"\"" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.461998 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxfhw\" (UniqueName: \"kubernetes.io/projected/0c2b8853-e439-4a04-a39d-500772139657-kube-api-access-bxfhw\") pod \"mariadb-client\" (UID: \"0c2b8853-e439-4a04-a39d-500772139657\") " pod="openstack/mariadb-client" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.484601 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxfhw\" (UniqueName: \"kubernetes.io/projected/0c2b8853-e439-4a04-a39d-500772139657-kube-api-access-bxfhw\") pod \"mariadb-client\" (UID: \"0c2b8853-e439-4a04-a39d-500772139657\") " pod="openstack/mariadb-client" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.623744 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.779564 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21ab44a046cf1a43bdfbffbe690167c8535ed3b985b9bb751b9fecd06378387c" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.779646 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:37:40 crc kubenswrapper[4846]: I0202 13:37:40.800401 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="1a8e81b4-d55f-4d63-9338-60f8e6ec84a0" podUID="0c2b8853-e439-4a04-a39d-500772139657" Feb 02 13:37:41 crc kubenswrapper[4846]: W0202 13:37:41.033367 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c2b8853_e439_4a04_a39d_500772139657.slice/crio-ffae76f085dc92b6defde786cfefd6a5701328076275baae5085ce682bc941c5 WatchSource:0}: Error finding container ffae76f085dc92b6defde786cfefd6a5701328076275baae5085ce682bc941c5: Status 404 returned error can't find the container with id ffae76f085dc92b6defde786cfefd6a5701328076275baae5085ce682bc941c5 Feb 02 13:37:41 crc kubenswrapper[4846]: I0202 13:37:41.033736 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:37:41 crc kubenswrapper[4846]: I0202 13:37:41.433801 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a8e81b4-d55f-4d63-9338-60f8e6ec84a0" path="/var/lib/kubelet/pods/1a8e81b4-d55f-4d63-9338-60f8e6ec84a0/volumes" Feb 02 13:37:41 crc kubenswrapper[4846]: I0202 13:37:41.788793 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c2b8853-e439-4a04-a39d-500772139657" containerID="f39ad7e438ed67302e002ed927484dbb86b13083f654b58d350e15c83a2302e9" exitCode=0 Feb 02 13:37:41 crc kubenswrapper[4846]: I0202 13:37:41.788908 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"0c2b8853-e439-4a04-a39d-500772139657","Type":"ContainerDied","Data":"f39ad7e438ed67302e002ed927484dbb86b13083f654b58d350e15c83a2302e9"} Feb 02 13:37:41 crc kubenswrapper[4846]: I0202 13:37:41.789662 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"0c2b8853-e439-4a04-a39d-500772139657","Type":"ContainerStarted","Data":"ffae76f085dc92b6defde786cfefd6a5701328076275baae5085ce682bc941c5"} Feb 02 13:37:43 crc kubenswrapper[4846]: I0202 13:37:43.152426 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:37:43 crc kubenswrapper[4846]: I0202 13:37:43.171941 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_0c2b8853-e439-4a04-a39d-500772139657/mariadb-client/0.log" Feb 02 13:37:43 crc kubenswrapper[4846]: I0202 13:37:43.202951 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:37:43 crc kubenswrapper[4846]: I0202 13:37:43.210045 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Feb 02 13:37:43 crc kubenswrapper[4846]: I0202 13:37:43.329912 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxfhw\" (UniqueName: \"kubernetes.io/projected/0c2b8853-e439-4a04-a39d-500772139657-kube-api-access-bxfhw\") pod \"0c2b8853-e439-4a04-a39d-500772139657\" (UID: \"0c2b8853-e439-4a04-a39d-500772139657\") " Feb 02 13:37:43 crc kubenswrapper[4846]: I0202 13:37:43.339828 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c2b8853-e439-4a04-a39d-500772139657-kube-api-access-bxfhw" (OuterVolumeSpecName: "kube-api-access-bxfhw") pod "0c2b8853-e439-4a04-a39d-500772139657" (UID: "0c2b8853-e439-4a04-a39d-500772139657"). InnerVolumeSpecName "kube-api-access-bxfhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:37:43 crc kubenswrapper[4846]: I0202 13:37:43.432312 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxfhw\" (UniqueName: \"kubernetes.io/projected/0c2b8853-e439-4a04-a39d-500772139657-kube-api-access-bxfhw\") on node \"crc\" DevicePath \"\"" Feb 02 13:37:43 crc kubenswrapper[4846]: I0202 13:37:43.434915 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c2b8853-e439-4a04-a39d-500772139657" path="/var/lib/kubelet/pods/0c2b8853-e439-4a04-a39d-500772139657/volumes" Feb 02 13:37:43 crc kubenswrapper[4846]: I0202 13:37:43.806315 4846 scope.go:117] "RemoveContainer" containerID="f39ad7e438ed67302e002ed927484dbb86b13083f654b58d350e15c83a2302e9" Feb 02 13:37:43 crc kubenswrapper[4846]: I0202 13:37:43.806404 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 02 13:38:11 crc kubenswrapper[4846]: I0202 13:38:11.925717 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j7kpg"] Feb 02 13:38:11 crc kubenswrapper[4846]: E0202 13:38:11.926907 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c2b8853-e439-4a04-a39d-500772139657" containerName="mariadb-client" Feb 02 13:38:11 crc kubenswrapper[4846]: I0202 13:38:11.926926 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c2b8853-e439-4a04-a39d-500772139657" containerName="mariadb-client" Feb 02 13:38:11 crc kubenswrapper[4846]: I0202 13:38:11.927122 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c2b8853-e439-4a04-a39d-500772139657" containerName="mariadb-client" Feb 02 13:38:11 crc kubenswrapper[4846]: I0202 13:38:11.941956 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:11 crc kubenswrapper[4846]: I0202 13:38:11.962511 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j7kpg"] Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.115136 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-utilities\") pod \"community-operators-j7kpg\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.115235 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-catalog-content\") pod \"community-operators-j7kpg\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.115297 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlcz2\" (UniqueName: \"kubernetes.io/projected/0c803fb9-798b-42ee-8330-7bea4144955b-kube-api-access-hlcz2\") pod \"community-operators-j7kpg\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.216761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-utilities\") pod \"community-operators-j7kpg\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.216888 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-catalog-content\") pod \"community-operators-j7kpg\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.216955 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlcz2\" (UniqueName: \"kubernetes.io/projected/0c803fb9-798b-42ee-8330-7bea4144955b-kube-api-access-hlcz2\") pod \"community-operators-j7kpg\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.217441 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-catalog-content\") pod \"community-operators-j7kpg\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.217611 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-utilities\") pod \"community-operators-j7kpg\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.242226 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlcz2\" (UniqueName: \"kubernetes.io/projected/0c803fb9-798b-42ee-8330-7bea4144955b-kube-api-access-hlcz2\") pod \"community-operators-j7kpg\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.273170 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:12 crc kubenswrapper[4846]: I0202 13:38:12.848914 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j7kpg"] Feb 02 13:38:13 crc kubenswrapper[4846]: I0202 13:38:13.031802 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7kpg" event={"ID":"0c803fb9-798b-42ee-8330-7bea4144955b","Type":"ContainerStarted","Data":"41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e"} Feb 02 13:38:13 crc kubenswrapper[4846]: I0202 13:38:13.032058 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7kpg" event={"ID":"0c803fb9-798b-42ee-8330-7bea4144955b","Type":"ContainerStarted","Data":"a203c277d50dbffb8359e50136e9f4809a09b481bfdd781bc8cce80d2d74c53a"} Feb 02 13:38:14 crc kubenswrapper[4846]: I0202 13:38:14.046583 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c803fb9-798b-42ee-8330-7bea4144955b" containerID="41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e" exitCode=0 Feb 02 13:38:14 crc kubenswrapper[4846]: I0202 13:38:14.046653 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7kpg" event={"ID":"0c803fb9-798b-42ee-8330-7bea4144955b","Type":"ContainerDied","Data":"41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e"} Feb 02 13:38:15 crc kubenswrapper[4846]: I0202 13:38:15.055463 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7kpg" event={"ID":"0c803fb9-798b-42ee-8330-7bea4144955b","Type":"ContainerStarted","Data":"96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5"} Feb 02 13:38:16 crc kubenswrapper[4846]: I0202 13:38:16.063989 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c803fb9-798b-42ee-8330-7bea4144955b" containerID="96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5" exitCode=0 Feb 02 13:38:16 crc kubenswrapper[4846]: I0202 13:38:16.064029 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7kpg" event={"ID":"0c803fb9-798b-42ee-8330-7bea4144955b","Type":"ContainerDied","Data":"96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5"} Feb 02 13:38:17 crc kubenswrapper[4846]: I0202 13:38:17.072116 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7kpg" event={"ID":"0c803fb9-798b-42ee-8330-7bea4144955b","Type":"ContainerStarted","Data":"f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74"} Feb 02 13:38:17 crc kubenswrapper[4846]: I0202 13:38:17.093900 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j7kpg" podStartSLOduration=3.651625718 podStartE2EDuration="6.093880679s" podCreationTimestamp="2026-02-02 13:38:11 +0000 UTC" firstStartedPulling="2026-02-02 13:38:14.048224914 +0000 UTC m=+5325.276811787" lastFinishedPulling="2026-02-02 13:38:16.490479885 +0000 UTC m=+5327.719066748" observedRunningTime="2026-02-02 13:38:17.087500314 +0000 UTC m=+5328.316087197" watchObservedRunningTime="2026-02-02 13:38:17.093880679 +0000 UTC m=+5328.322467542" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.506941 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.508899 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.510902 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.520025 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.520849 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.520848 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.521237 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.521439 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-jcjgp" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.529256 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.541937 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.601899 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.604221 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.626131 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.638075 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.669470 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.669583 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/396188ac-2f00-4ae5-8a3a-0273a704e60f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.669668 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.669709 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc6zs\" (UniqueName: \"kubernetes.io/projected/396188ac-2f00-4ae5-8a3a-0273a704e60f-kube-api-access-qc6zs\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.669773 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b2e951a4-613c-4c5a-9ac8-fa89e89f63da\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2e951a4-613c-4c5a-9ac8-fa89e89f63da\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.669859 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f7ebd127-0b83-4f29-878b-d3d3ea675f43\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f7ebd127-0b83-4f29-878b-d3d3ea675f43\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.669913 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-config\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.669935 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/396188ac-2f00-4ae5-8a3a-0273a704e60f-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.669950 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/396188ac-2f00-4ae5-8a3a-0273a704e60f-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.670017 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.670041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.670074 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/396188ac-2f00-4ae5-8a3a-0273a704e60f-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.670093 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/396188ac-2f00-4ae5-8a3a-0273a704e60f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.670146 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw5gv\" (UniqueName: \"kubernetes.io/projected/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-kube-api-access-rw5gv\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.670176 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.670210 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/396188ac-2f00-4ae5-8a3a-0273a704e60f-config\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.771993 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/396188ac-2f00-4ae5-8a3a-0273a704e60f-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772052 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/396188ac-2f00-4ae5-8a3a-0273a704e60f-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772084 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb3364f-23b6-4875-a37d-a16af9832978-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772114 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772150 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772177 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/396188ac-2f00-4ae5-8a3a-0273a704e60f-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772201 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/396188ac-2f00-4ae5-8a3a-0273a704e60f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772247 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb3364f-23b6-4875-a37d-a16af9832978-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772283 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw5gv\" (UniqueName: \"kubernetes.io/projected/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-kube-api-access-rw5gv\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772322 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772349 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb3364f-23b6-4875-a37d-a16af9832978-config\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772384 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/396188ac-2f00-4ae5-8a3a-0273a704e60f-config\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772411 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772446 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9e53ad48-6c77-4ff8-9777-e45152af3e5c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9e53ad48-6c77-4ff8-9777-e45152af3e5c\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772476 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/396188ac-2f00-4ae5-8a3a-0273a704e60f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772501 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772532 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt89n\" (UniqueName: \"kubernetes.io/projected/fdb3364f-23b6-4875-a37d-a16af9832978-kube-api-access-kt89n\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772563 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb3364f-23b6-4875-a37d-a16af9832978-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772594 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc6zs\" (UniqueName: \"kubernetes.io/projected/396188ac-2f00-4ae5-8a3a-0273a704e60f-kube-api-access-qc6zs\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772644 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb3364f-23b6-4875-a37d-a16af9832978-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772686 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b2e951a4-613c-4c5a-9ac8-fa89e89f63da\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2e951a4-613c-4c5a-9ac8-fa89e89f63da\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772709 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fdb3364f-23b6-4875-a37d-a16af9832978-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772745 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f7ebd127-0b83-4f29-878b-d3d3ea675f43\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f7ebd127-0b83-4f29-878b-d3d3ea675f43\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772773 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-config\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.772803 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/396188ac-2f00-4ae5-8a3a-0273a704e60f-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.773396 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.773537 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-config\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.773967 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.774220 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/396188ac-2f00-4ae5-8a3a-0273a704e60f-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.774537 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/396188ac-2f00-4ae5-8a3a-0273a704e60f-config\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.776023 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.776062 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f7ebd127-0b83-4f29-878b-d3d3ea675f43\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f7ebd127-0b83-4f29-878b-d3d3ea675f43\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/543bf47d9b3324fd2a123bb7b663e0ed35da37553614ba38bf087ac038b021d0/globalmount\"" pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.776028 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.776148 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b2e951a4-613c-4c5a-9ac8-fa89e89f63da\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2e951a4-613c-4c5a-9ac8-fa89e89f63da\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/77b39bab998744272bfe390ba4f7701943a38ef6fa6adabaaa3cfa4d416bdf5e/globalmount\"" pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.779388 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/396188ac-2f00-4ae5-8a3a-0273a704e60f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.781037 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/396188ac-2f00-4ae5-8a3a-0273a704e60f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.781327 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.782827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.783083 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.785326 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/396188ac-2f00-4ae5-8a3a-0273a704e60f-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.789942 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc6zs\" (UniqueName: \"kubernetes.io/projected/396188ac-2f00-4ae5-8a3a-0273a704e60f-kube-api-access-qc6zs\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.790046 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw5gv\" (UniqueName: \"kubernetes.io/projected/692cf6ee-1eb8-45e5-be12-44bb709fd6ca-kube-api-access-rw5gv\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.814970 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f7ebd127-0b83-4f29-878b-d3d3ea675f43\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f7ebd127-0b83-4f29-878b-d3d3ea675f43\") pod \"ovsdbserver-nb-0\" (UID: \"692cf6ee-1eb8-45e5-be12-44bb709fd6ca\") " pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.821957 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b2e951a4-613c-4c5a-9ac8-fa89e89f63da\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b2e951a4-613c-4c5a-9ac8-fa89e89f63da\") pod \"ovsdbserver-nb-2\" (UID: \"396188ac-2f00-4ae5-8a3a-0273a704e60f\") " pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.850057 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.873870 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.874599 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fdb3364f-23b6-4875-a37d-a16af9832978-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.874690 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb3364f-23b6-4875-a37d-a16af9832978-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.874747 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb3364f-23b6-4875-a37d-a16af9832978-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.874794 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb3364f-23b6-4875-a37d-a16af9832978-config\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.874844 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9e53ad48-6c77-4ff8-9777-e45152af3e5c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9e53ad48-6c77-4ff8-9777-e45152af3e5c\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.874876 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt89n\" (UniqueName: \"kubernetes.io/projected/fdb3364f-23b6-4875-a37d-a16af9832978-kube-api-access-kt89n\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.874906 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb3364f-23b6-4875-a37d-a16af9832978-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.874941 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb3364f-23b6-4875-a37d-a16af9832978-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.875179 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fdb3364f-23b6-4875-a37d-a16af9832978-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.876113 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdb3364f-23b6-4875-a37d-a16af9832978-config\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.877478 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb3364f-23b6-4875-a37d-a16af9832978-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.877964 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.878012 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9e53ad48-6c77-4ff8-9777-e45152af3e5c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9e53ad48-6c77-4ff8-9777-e45152af3e5c\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bc88b421c90d63aa7a05f1248ed9f9cb8a865c0a23f5aaf7f8c52cc3b7029ba0/globalmount\"" pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.879767 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb3364f-23b6-4875-a37d-a16af9832978-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.880135 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fdb3364f-23b6-4875-a37d-a16af9832978-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.880863 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb3364f-23b6-4875-a37d-a16af9832978-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.893935 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt89n\" (UniqueName: \"kubernetes.io/projected/fdb3364f-23b6-4875-a37d-a16af9832978-kube-api-access-kt89n\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.921787 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9e53ad48-6c77-4ff8-9777-e45152af3e5c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9e53ad48-6c77-4ff8-9777-e45152af3e5c\") pod \"ovsdbserver-nb-1\" (UID: \"fdb3364f-23b6-4875-a37d-a16af9832978\") " pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:21 crc kubenswrapper[4846]: I0202 13:38:21.932756 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:22 crc kubenswrapper[4846]: I0202 13:38:22.228458 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 02 13:38:22 crc kubenswrapper[4846]: I0202 13:38:22.274004 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:22 crc kubenswrapper[4846]: I0202 13:38:22.274070 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:22 crc kubenswrapper[4846]: I0202 13:38:22.331906 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:22 crc kubenswrapper[4846]: I0202 13:38:22.551987 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Feb 02 13:38:23 crc kubenswrapper[4846]: I0202 13:38:23.119843 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"692cf6ee-1eb8-45e5-be12-44bb709fd6ca","Type":"ContainerStarted","Data":"6ccf111e62a0bb170af7f798e1b246266e14746894fa45c4911714fea5137dab"} Feb 02 13:38:23 crc kubenswrapper[4846]: I0202 13:38:23.120155 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"692cf6ee-1eb8-45e5-be12-44bb709fd6ca","Type":"ContainerStarted","Data":"c3ca92db0f3cfea980f004855f18993ed6f66ad348fd922bea0141f18772b702"} Feb 02 13:38:23 crc kubenswrapper[4846]: I0202 13:38:23.120167 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"692cf6ee-1eb8-45e5-be12-44bb709fd6ca","Type":"ContainerStarted","Data":"3d24a94bb15518d4310fca7e1a8c3432370ae7940a534c6829d4b742ac60dc0b"} Feb 02 13:38:23 crc kubenswrapper[4846]: I0202 13:38:23.122034 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"396188ac-2f00-4ae5-8a3a-0273a704e60f","Type":"ContainerStarted","Data":"5a57a0b6f96353be54bc224544d72cb3a0420eb0604dfeefc244e5a4b0686d19"} Feb 02 13:38:23 crc kubenswrapper[4846]: I0202 13:38:23.122065 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"396188ac-2f00-4ae5-8a3a-0273a704e60f","Type":"ContainerStarted","Data":"1d5870d649b29d7056479430f3c438f4fb7cb66213eba41965992f3b128d08b2"} Feb 02 13:38:23 crc kubenswrapper[4846]: I0202 13:38:23.157138 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.157112178 podStartE2EDuration="3.157112178s" podCreationTimestamp="2026-02-02 13:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:23.155810767 +0000 UTC m=+5334.384397630" watchObservedRunningTime="2026-02-02 13:38:23.157112178 +0000 UTC m=+5334.385699031" Feb 02 13:38:23 crc kubenswrapper[4846]: I0202 13:38:23.201935 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Feb 02 13:38:23 crc kubenswrapper[4846]: W0202 13:38:23.203846 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdb3364f_23b6_4875_a37d_a16af9832978.slice/crio-f2a7611f59996ec0e9c545556e86d31f7bcec27e5e20dbefbfc158c85e5c26eb WatchSource:0}: Error finding container f2a7611f59996ec0e9c545556e86d31f7bcec27e5e20dbefbfc158c85e5c26eb: Status 404 returned error can't find the container with id f2a7611f59996ec0e9c545556e86d31f7bcec27e5e20dbefbfc158c85e5c26eb Feb 02 13:38:23 crc kubenswrapper[4846]: I0202 13:38:23.204605 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:23 crc kubenswrapper[4846]: I0202 13:38:23.266239 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j7kpg"] Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.022852 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.024352 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.027658 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.028016 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vlkxl" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.028331 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.028574 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.052521 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.080787 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.088083 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.097535 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.105205 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.107493 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.114607 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.116644 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/106a77c0-d654-4a19-a691-cf7bf2efd9c4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.116706 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106a77c0-d654-4a19-a691-cf7bf2efd9c4-config\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.116733 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/106a77c0-d654-4a19-a691-cf7bf2efd9c4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.116756 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/106a77c0-d654-4a19-a691-cf7bf2efd9c4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.116782 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/106a77c0-d654-4a19-a691-cf7bf2efd9c4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.116821 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c7afa487-24dc-4d51-9df2-9c7236741a85\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c7afa487-24dc-4d51-9df2-9c7236741a85\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.116881 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nxff\" (UniqueName: \"kubernetes.io/projected/106a77c0-d654-4a19-a691-cf7bf2efd9c4-kube-api-access-5nxff\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.116930 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/106a77c0-d654-4a19-a691-cf7bf2efd9c4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.135402 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"396188ac-2f00-4ae5-8a3a-0273a704e60f","Type":"ContainerStarted","Data":"d30a09254a4d069fd886266a5bb2226731914f92fe409cb05edc5078ec5323ca"} Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.137868 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"fdb3364f-23b6-4875-a37d-a16af9832978","Type":"ContainerStarted","Data":"6de4034ef198b18b8e8a6bf6c06f806bd2617bef89dbf3c3d9c760e47a8d847b"} Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.137949 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"fdb3364f-23b6-4875-a37d-a16af9832978","Type":"ContainerStarted","Data":"24db570b8cff71511fbca03c0805dfb366599055c25b19ac36302a84567e9a23"} Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.137971 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"fdb3364f-23b6-4875-a37d-a16af9832978","Type":"ContainerStarted","Data":"f2a7611f59996ec0e9c545556e86d31f7bcec27e5e20dbefbfc158c85e5c26eb"} Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.154664 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=4.154614094 podStartE2EDuration="4.154614094s" podCreationTimestamp="2026-02-02 13:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:24.151751954 +0000 UTC m=+5335.380338827" watchObservedRunningTime="2026-02-02 13:38:24.154614094 +0000 UTC m=+5335.383200957" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.178133 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.178112604 podStartE2EDuration="4.178112604s" podCreationTimestamp="2026-02-02 13:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:24.170460498 +0000 UTC m=+5335.399047361" watchObservedRunningTime="2026-02-02 13:38:24.178112604 +0000 UTC m=+5335.406699467" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.218256 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c7afa487-24dc-4d51-9df2-9c7236741a85\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c7afa487-24dc-4d51-9df2-9c7236741a85\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.218310 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-83be50bf-608e-4aa3-8f71-aedc782217b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83be50bf-608e-4aa3-8f71-aedc782217b9\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.218345 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e326ede9-6300-47da-b9b8-48e8ade170a4-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.218385 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5088d2b9-57d7-4471-bc6f-d986d5c09106\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5088d2b9-57d7-4471-bc6f-d986d5c09106\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.218421 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e326ede9-6300-47da-b9b8-48e8ade170a4-config\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.218450 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nxff\" (UniqueName: \"kubernetes.io/projected/106a77c0-d654-4a19-a691-cf7bf2efd9c4-kube-api-access-5nxff\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.218480 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/230ca7a4-fc29-44a0-8986-e25db00d4ebc-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.218502 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/230ca7a4-fc29-44a0-8986-e25db00d4ebc-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.219816 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e326ede9-6300-47da-b9b8-48e8ade170a4-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.219974 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/106a77c0-d654-4a19-a691-cf7bf2efd9c4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220025 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/230ca7a4-fc29-44a0-8986-e25db00d4ebc-config\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220063 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e326ede9-6300-47da-b9b8-48e8ade170a4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220115 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e326ede9-6300-47da-b9b8-48e8ade170a4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220159 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/230ca7a4-fc29-44a0-8986-e25db00d4ebc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220220 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e326ede9-6300-47da-b9b8-48e8ade170a4-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220314 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/230ca7a4-fc29-44a0-8986-e25db00d4ebc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220497 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hp77\" (UniqueName: \"kubernetes.io/projected/230ca7a4-fc29-44a0-8986-e25db00d4ebc-kube-api-access-2hp77\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220665 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/106a77c0-d654-4a19-a691-cf7bf2efd9c4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220689 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/106a77c0-d654-4a19-a691-cf7bf2efd9c4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230ca7a4-fc29-44a0-8986-e25db00d4ebc-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220832 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kx7z\" (UniqueName: \"kubernetes.io/projected/e326ede9-6300-47da-b9b8-48e8ade170a4-kube-api-access-4kx7z\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.220952 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106a77c0-d654-4a19-a691-cf7bf2efd9c4-config\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.221019 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/106a77c0-d654-4a19-a691-cf7bf2efd9c4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.221076 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/106a77c0-d654-4a19-a691-cf7bf2efd9c4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.221104 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/106a77c0-d654-4a19-a691-cf7bf2efd9c4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.222849 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/106a77c0-d654-4a19-a691-cf7bf2efd9c4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.222974 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106a77c0-d654-4a19-a691-cf7bf2efd9c4-config\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.223669 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.223730 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c7afa487-24dc-4d51-9df2-9c7236741a85\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c7afa487-24dc-4d51-9df2-9c7236741a85\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0254c7c11bf68c4efcaa256fa47fb2a91f99de084e19093fc1c469aa87974124/globalmount\"" pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.227100 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/106a77c0-d654-4a19-a691-cf7bf2efd9c4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.227200 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/106a77c0-d654-4a19-a691-cf7bf2efd9c4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.234615 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/106a77c0-d654-4a19-a691-cf7bf2efd9c4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.240890 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nxff\" (UniqueName: \"kubernetes.io/projected/106a77c0-d654-4a19-a691-cf7bf2efd9c4-kube-api-access-5nxff\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.270869 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c7afa487-24dc-4d51-9df2-9c7236741a85\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c7afa487-24dc-4d51-9df2-9c7236741a85\") pod \"ovsdbserver-sb-0\" (UID: \"106a77c0-d654-4a19-a691-cf7bf2efd9c4\") " pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e326ede9-6300-47da-b9b8-48e8ade170a4-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323527 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5088d2b9-57d7-4471-bc6f-d986d5c09106\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5088d2b9-57d7-4471-bc6f-d986d5c09106\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323555 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e326ede9-6300-47da-b9b8-48e8ade170a4-config\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/230ca7a4-fc29-44a0-8986-e25db00d4ebc-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323684 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/230ca7a4-fc29-44a0-8986-e25db00d4ebc-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323712 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e326ede9-6300-47da-b9b8-48e8ade170a4-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323743 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/230ca7a4-fc29-44a0-8986-e25db00d4ebc-config\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323769 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e326ede9-6300-47da-b9b8-48e8ade170a4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323791 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e326ede9-6300-47da-b9b8-48e8ade170a4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323814 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/230ca7a4-fc29-44a0-8986-e25db00d4ebc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323834 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e326ede9-6300-47da-b9b8-48e8ade170a4-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323859 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/230ca7a4-fc29-44a0-8986-e25db00d4ebc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323890 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hp77\" (UniqueName: \"kubernetes.io/projected/230ca7a4-fc29-44a0-8986-e25db00d4ebc-kube-api-access-2hp77\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323932 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230ca7a4-fc29-44a0-8986-e25db00d4ebc-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.323956 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kx7z\" (UniqueName: \"kubernetes.io/projected/e326ede9-6300-47da-b9b8-48e8ade170a4-kube-api-access-4kx7z\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.324015 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-83be50bf-608e-4aa3-8f71-aedc782217b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83be50bf-608e-4aa3-8f71-aedc782217b9\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.325394 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e326ede9-6300-47da-b9b8-48e8ade170a4-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.325735 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e326ede9-6300-47da-b9b8-48e8ade170a4-config\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.326889 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e326ede9-6300-47da-b9b8-48e8ade170a4-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.327264 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/230ca7a4-fc29-44a0-8986-e25db00d4ebc-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.327967 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.327975 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/230ca7a4-fc29-44a0-8986-e25db00d4ebc-config\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.328002 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5088d2b9-57d7-4471-bc6f-d986d5c09106\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5088d2b9-57d7-4471-bc6f-d986d5c09106\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5dd7f4a64856cf88613eb0a2fb92c15b04914a4d6fa5a7984c692c9486fef5b6/globalmount\"" pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.328197 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/230ca7a4-fc29-44a0-8986-e25db00d4ebc-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.328243 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.328285 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-83be50bf-608e-4aa3-8f71-aedc782217b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83be50bf-608e-4aa3-8f71-aedc782217b9\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e688360666d1cc001835874daaf4a7ca2b33d3daffe0f498786e8117ecf069bc/globalmount\"" pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.330356 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e326ede9-6300-47da-b9b8-48e8ade170a4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.331377 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/230ca7a4-fc29-44a0-8986-e25db00d4ebc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.331733 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e326ede9-6300-47da-b9b8-48e8ade170a4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.331962 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/230ca7a4-fc29-44a0-8986-e25db00d4ebc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.336837 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e326ede9-6300-47da-b9b8-48e8ade170a4-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.339269 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230ca7a4-fc29-44a0-8986-e25db00d4ebc-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.345237 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hp77\" (UniqueName: \"kubernetes.io/projected/230ca7a4-fc29-44a0-8986-e25db00d4ebc-kube-api-access-2hp77\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.348320 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kx7z\" (UniqueName: \"kubernetes.io/projected/e326ede9-6300-47da-b9b8-48e8ade170a4-kube-api-access-4kx7z\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.364201 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5088d2b9-57d7-4471-bc6f-d986d5c09106\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5088d2b9-57d7-4471-bc6f-d986d5c09106\") pod \"ovsdbserver-sb-1\" (UID: \"230ca7a4-fc29-44a0-8986-e25db00d4ebc\") " pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.368113 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-83be50bf-608e-4aa3-8f71-aedc782217b9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-83be50bf-608e-4aa3-8f71-aedc782217b9\") pod \"ovsdbserver-sb-2\" (UID: \"e326ede9-6300-47da-b9b8-48e8ade170a4\") " pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.372305 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.412977 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.427817 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.851005 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.874211 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.907866 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 02 13:38:24 crc kubenswrapper[4846]: W0202 13:38:24.916017 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod106a77c0_d654_4a19_a691_cf7bf2efd9c4.slice/crio-028bed3841ea3f4f005578e77f3e5b2a18005e982475e06d7f6f8294cefd93e3 WatchSource:0}: Error finding container 028bed3841ea3f4f005578e77f3e5b2a18005e982475e06d7f6f8294cefd93e3: Status 404 returned error can't find the container with id 028bed3841ea3f4f005578e77f3e5b2a18005e982475e06d7f6f8294cefd93e3 Feb 02 13:38:24 crc kubenswrapper[4846]: I0202 13:38:24.933370 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.024032 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.096956 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Feb 02 13:38:25 crc kubenswrapper[4846]: W0202 13:38:25.109256 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod230ca7a4_fc29_44a0_8986_e25db00d4ebc.slice/crio-752e76ed6e52ef2ac7f8f9d3728bd0d344d28579dba860f6ffb543998cf846a5 WatchSource:0}: Error finding container 752e76ed6e52ef2ac7f8f9d3728bd0d344d28579dba860f6ffb543998cf846a5: Status 404 returned error can't find the container with id 752e76ed6e52ef2ac7f8f9d3728bd0d344d28579dba860f6ffb543998cf846a5 Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.155126 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"106a77c0-d654-4a19-a691-cf7bf2efd9c4","Type":"ContainerStarted","Data":"028bed3841ea3f4f005578e77f3e5b2a18005e982475e06d7f6f8294cefd93e3"} Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.156555 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"e326ede9-6300-47da-b9b8-48e8ade170a4","Type":"ContainerStarted","Data":"167286fe87d1d1fdfa5fa95dcef7a7af6bc877420a92569513b51884a9c4d663"} Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.157954 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"230ca7a4-fc29-44a0-8986-e25db00d4ebc","Type":"ContainerStarted","Data":"752e76ed6e52ef2ac7f8f9d3728bd0d344d28579dba860f6ffb543998cf846a5"} Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.158059 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j7kpg" podUID="0c803fb9-798b-42ee-8330-7bea4144955b" containerName="registry-server" containerID="cri-o://f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74" gracePeriod=2 Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.517820 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.648862 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-catalog-content\") pod \"0c803fb9-798b-42ee-8330-7bea4144955b\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.649126 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-utilities\") pod \"0c803fb9-798b-42ee-8330-7bea4144955b\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.649177 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlcz2\" (UniqueName: \"kubernetes.io/projected/0c803fb9-798b-42ee-8330-7bea4144955b-kube-api-access-hlcz2\") pod \"0c803fb9-798b-42ee-8330-7bea4144955b\" (UID: \"0c803fb9-798b-42ee-8330-7bea4144955b\") " Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.650541 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-utilities" (OuterVolumeSpecName: "utilities") pod "0c803fb9-798b-42ee-8330-7bea4144955b" (UID: "0c803fb9-798b-42ee-8330-7bea4144955b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.655464 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c803fb9-798b-42ee-8330-7bea4144955b-kube-api-access-hlcz2" (OuterVolumeSpecName: "kube-api-access-hlcz2") pod "0c803fb9-798b-42ee-8330-7bea4144955b" (UID: "0c803fb9-798b-42ee-8330-7bea4144955b"). InnerVolumeSpecName "kube-api-access-hlcz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.715600 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c803fb9-798b-42ee-8330-7bea4144955b" (UID: "0c803fb9-798b-42ee-8330-7bea4144955b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.750904 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.750935 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c803fb9-798b-42ee-8330-7bea4144955b-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:25 crc kubenswrapper[4846]: I0202 13:38:25.750944 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlcz2\" (UniqueName: \"kubernetes.io/projected/0c803fb9-798b-42ee-8330-7bea4144955b-kube-api-access-hlcz2\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.165245 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"230ca7a4-fc29-44a0-8986-e25db00d4ebc","Type":"ContainerStarted","Data":"bfe19a15fe08940b189bec352dc6b4a9a586624e8659cd1b03a679c2260c4090"} Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.165300 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"230ca7a4-fc29-44a0-8986-e25db00d4ebc","Type":"ContainerStarted","Data":"dbf7df036b8d006fbc5a75d9a3ebf95daa886b9a187b8b4a600efa9572b08b69"} Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.167128 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c803fb9-798b-42ee-8330-7bea4144955b" containerID="f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74" exitCode=0 Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.167194 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j7kpg" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.167208 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7kpg" event={"ID":"0c803fb9-798b-42ee-8330-7bea4144955b","Type":"ContainerDied","Data":"f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74"} Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.167252 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j7kpg" event={"ID":"0c803fb9-798b-42ee-8330-7bea4144955b","Type":"ContainerDied","Data":"a203c277d50dbffb8359e50136e9f4809a09b481bfdd781bc8cce80d2d74c53a"} Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.167272 4846 scope.go:117] "RemoveContainer" containerID="f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.168921 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"106a77c0-d654-4a19-a691-cf7bf2efd9c4","Type":"ContainerStarted","Data":"1f339ea26d50d25a1ef9992072915ff0d438f9cbabe24aa8ada9bed5a9d82cbf"} Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.168969 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"106a77c0-d654-4a19-a691-cf7bf2efd9c4","Type":"ContainerStarted","Data":"2c89a1cd2155b2a68eb86d5a7b6489780161b0bc2dbfc6f53419f18550f77da7"} Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.173070 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"e326ede9-6300-47da-b9b8-48e8ade170a4","Type":"ContainerStarted","Data":"5464bc3020b205be01229f28fb57e370cfac596a2563d67fd9f895b0d30918c4"} Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.173189 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"e326ede9-6300-47da-b9b8-48e8ade170a4","Type":"ContainerStarted","Data":"f9a73e950b4d68dc863e45e03af82e31067325c5633119927d4071f34f1afc28"} Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.190999 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.190971628 podStartE2EDuration="3.190971628s" podCreationTimestamp="2026-02-02 13:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:26.187977315 +0000 UTC m=+5337.416564198" watchObservedRunningTime="2026-02-02 13:38:26.190971628 +0000 UTC m=+5337.419558491" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.195267 4846 scope.go:117] "RemoveContainer" containerID="96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.217080 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.217061282 podStartE2EDuration="3.217061282s" podCreationTimestamp="2026-02-02 13:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:26.21407215 +0000 UTC m=+5337.442659033" watchObservedRunningTime="2026-02-02 13:38:26.217061282 +0000 UTC m=+5337.445648155" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.232388 4846 scope.go:117] "RemoveContainer" containerID="41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.242702 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.242680244 podStartE2EDuration="4.242680244s" podCreationTimestamp="2026-02-02 13:38:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:26.236245627 +0000 UTC m=+5337.464832640" watchObservedRunningTime="2026-02-02 13:38:26.242680244 +0000 UTC m=+5337.471267107" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.269318 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j7kpg"] Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.273303 4846 scope.go:117] "RemoveContainer" containerID="f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74" Feb 02 13:38:26 crc kubenswrapper[4846]: E0202 13:38:26.274039 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74\": container with ID starting with f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74 not found: ID does not exist" containerID="f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.274093 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74"} err="failed to get container status \"f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74\": rpc error: code = NotFound desc = could not find container \"f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74\": container with ID starting with f9a1b562505f7cfa1dac28835dc400f5940183e0292e9997d2f7f7db55bb3f74 not found: ID does not exist" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.274113 4846 scope.go:117] "RemoveContainer" containerID="96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5" Feb 02 13:38:26 crc kubenswrapper[4846]: E0202 13:38:26.274615 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5\": container with ID starting with 96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5 not found: ID does not exist" containerID="96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.274680 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5"} err="failed to get container status \"96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5\": rpc error: code = NotFound desc = could not find container \"96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5\": container with ID starting with 96667f160ea0ab5d57991d5dedee11e4ea50937e77c359328e2d241aeadb27e5 not found: ID does not exist" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.274707 4846 scope.go:117] "RemoveContainer" containerID="41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e" Feb 02 13:38:26 crc kubenswrapper[4846]: E0202 13:38:26.275080 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e\": container with ID starting with 41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e not found: ID does not exist" containerID="41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.275138 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e"} err="failed to get container status \"41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e\": rpc error: code = NotFound desc = could not find container \"41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e\": container with ID starting with 41afdf9040173f5e0ac99057d9fd75a9c8a3076e1d151dc731c07789b87fac5e not found: ID does not exist" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.278190 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j7kpg"] Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.850725 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.874327 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:26 crc kubenswrapper[4846]: I0202 13:38:26.933360 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:27 crc kubenswrapper[4846]: I0202 13:38:27.372504 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:27 crc kubenswrapper[4846]: I0202 13:38:27.413778 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:27 crc kubenswrapper[4846]: I0202 13:38:27.436723 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c803fb9-798b-42ee-8330-7bea4144955b" path="/var/lib/kubelet/pods/0c803fb9-798b-42ee-8330-7bea4144955b/volumes" Feb 02 13:38:27 crc kubenswrapper[4846]: I0202 13:38:27.437775 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:27 crc kubenswrapper[4846]: I0202 13:38:27.891434 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:27 crc kubenswrapper[4846]: I0202 13:38:27.922067 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:27 crc kubenswrapper[4846]: I0202 13:38:27.934165 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 02 13:38:27 crc kubenswrapper[4846]: I0202 13:38:27.981494 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.261061 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-697b8d7675-cxfhc"] Feb 02 13:38:28 crc kubenswrapper[4846]: E0202 13:38:28.261469 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c803fb9-798b-42ee-8330-7bea4144955b" containerName="extract-content" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.261494 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c803fb9-798b-42ee-8330-7bea4144955b" containerName="extract-content" Feb 02 13:38:28 crc kubenswrapper[4846]: E0202 13:38:28.261519 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c803fb9-798b-42ee-8330-7bea4144955b" containerName="extract-utilities" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.261528 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c803fb9-798b-42ee-8330-7bea4144955b" containerName="extract-utilities" Feb 02 13:38:28 crc kubenswrapper[4846]: E0202 13:38:28.261552 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c803fb9-798b-42ee-8330-7bea4144955b" containerName="registry-server" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.261560 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c803fb9-798b-42ee-8330-7bea4144955b" containerName="registry-server" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.261810 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c803fb9-798b-42ee-8330-7bea4144955b" containerName="registry-server" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.262736 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.263988 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.268358 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.270394 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-697b8d7675-cxfhc"] Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.278267 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.419748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-dns-svc\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.419864 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-config\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.420110 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-ovsdbserver-nb\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.420319 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9pzq\" (UniqueName: \"kubernetes.io/projected/aa61dea3-202c-45e9-a5d5-647a9948e006-kube-api-access-x9pzq\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.522809 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9pzq\" (UniqueName: \"kubernetes.io/projected/aa61dea3-202c-45e9-a5d5-647a9948e006-kube-api-access-x9pzq\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.522942 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-dns-svc\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.522996 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-config\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.523090 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-ovsdbserver-nb\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.523952 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-config\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.524383 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-ovsdbserver-nb\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.524611 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-dns-svc\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.549036 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9pzq\" (UniqueName: \"kubernetes.io/projected/aa61dea3-202c-45e9-a5d5-647a9948e006-kube-api-access-x9pzq\") pod \"dnsmasq-dns-697b8d7675-cxfhc\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:28 crc kubenswrapper[4846]: I0202 13:38:28.613783 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:29 crc kubenswrapper[4846]: I0202 13:38:29.085642 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-697b8d7675-cxfhc"] Feb 02 13:38:29 crc kubenswrapper[4846]: W0202 13:38:29.091209 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa61dea3_202c_45e9_a5d5_647a9948e006.slice/crio-930242ef569cecd9c25315921c8f62246fc5c2f17a1258e050d720ed96f9b002 WatchSource:0}: Error finding container 930242ef569cecd9c25315921c8f62246fc5c2f17a1258e050d720ed96f9b002: Status 404 returned error can't find the container with id 930242ef569cecd9c25315921c8f62246fc5c2f17a1258e050d720ed96f9b002 Feb 02 13:38:29 crc kubenswrapper[4846]: I0202 13:38:29.205796 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" event={"ID":"aa61dea3-202c-45e9-a5d5-647a9948e006","Type":"ContainerStarted","Data":"930242ef569cecd9c25315921c8f62246fc5c2f17a1258e050d720ed96f9b002"} Feb 02 13:38:29 crc kubenswrapper[4846]: I0202 13:38:29.373189 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:29 crc kubenswrapper[4846]: I0202 13:38:29.413877 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:29 crc kubenswrapper[4846]: I0202 13:38:29.437360 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.218077 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa61dea3-202c-45e9-a5d5-647a9948e006" containerID="b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0" exitCode=0 Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.218140 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" event={"ID":"aa61dea3-202c-45e9-a5d5-647a9948e006","Type":"ContainerDied","Data":"b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0"} Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.428570 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.460231 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.486550 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.494390 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.520114 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.532983 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.804722 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-697b8d7675-cxfhc"] Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.833120 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55cb6fc89-4lctn"] Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.834970 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.838013 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.854761 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55cb6fc89-4lctn"] Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.968403 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-sb\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.968497 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-config\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.968525 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpjvd\" (UniqueName: \"kubernetes.io/projected/9a717af0-77ea-4be4-ba01-8390c53da81c-kube-api-access-gpjvd\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.968930 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-dns-svc\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:30 crc kubenswrapper[4846]: I0202 13:38:30.969099 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-nb\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.071031 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-nb\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.071393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-sb\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.071425 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-config\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.071449 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpjvd\" (UniqueName: \"kubernetes.io/projected/9a717af0-77ea-4be4-ba01-8390c53da81c-kube-api-access-gpjvd\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.071535 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-dns-svc\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.071985 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-nb\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.072207 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-sb\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.072271 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-dns-svc\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.072581 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-config\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.090775 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpjvd\" (UniqueName: \"kubernetes.io/projected/9a717af0-77ea-4be4-ba01-8390c53da81c-kube-api-access-gpjvd\") pod \"dnsmasq-dns-55cb6fc89-4lctn\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.157680 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.229532 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" event={"ID":"aa61dea3-202c-45e9-a5d5-647a9948e006","Type":"ContainerStarted","Data":"b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f"} Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.230171 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.588096 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" podStartSLOduration=3.588073181 podStartE2EDuration="3.588073181s" podCreationTimestamp="2026-02-02 13:38:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:31.252725186 +0000 UTC m=+5342.481312059" watchObservedRunningTime="2026-02-02 13:38:31.588073181 +0000 UTC m=+5342.816660044" Feb 02 13:38:31 crc kubenswrapper[4846]: I0202 13:38:31.594609 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55cb6fc89-4lctn"] Feb 02 13:38:31 crc kubenswrapper[4846]: W0202 13:38:31.603463 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a717af0_77ea_4be4_ba01_8390c53da81c.slice/crio-671eb996cb8edd21d55614ca9f6fe8bf7fe7abbdfc5fa7b84e94269bb553b795 WatchSource:0}: Error finding container 671eb996cb8edd21d55614ca9f6fe8bf7fe7abbdfc5fa7b84e94269bb553b795: Status 404 returned error can't find the container with id 671eb996cb8edd21d55614ca9f6fe8bf7fe7abbdfc5fa7b84e94269bb553b795 Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.239007 4846 generic.go:334] "Generic (PLEG): container finished" podID="9a717af0-77ea-4be4-ba01-8390c53da81c" containerID="772335beca7d1b5e1d0b58c203fa5b7d7d0ef58f1904c8998b8a7ce8a4b930e2" exitCode=0 Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.239055 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" event={"ID":"9a717af0-77ea-4be4-ba01-8390c53da81c","Type":"ContainerDied","Data":"772335beca7d1b5e1d0b58c203fa5b7d7d0ef58f1904c8998b8a7ce8a4b930e2"} Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.239421 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" event={"ID":"9a717af0-77ea-4be4-ba01-8390c53da81c","Type":"ContainerStarted","Data":"671eb996cb8edd21d55614ca9f6fe8bf7fe7abbdfc5fa7b84e94269bb553b795"} Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.239532 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" podUID="aa61dea3-202c-45e9-a5d5-647a9948e006" containerName="dnsmasq-dns" containerID="cri-o://b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f" gracePeriod=10 Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.627329 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.701979 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-config\") pod \"aa61dea3-202c-45e9-a5d5-647a9948e006\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.702168 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-dns-svc\") pod \"aa61dea3-202c-45e9-a5d5-647a9948e006\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.702304 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-ovsdbserver-nb\") pod \"aa61dea3-202c-45e9-a5d5-647a9948e006\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.702341 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9pzq\" (UniqueName: \"kubernetes.io/projected/aa61dea3-202c-45e9-a5d5-647a9948e006-kube-api-access-x9pzq\") pod \"aa61dea3-202c-45e9-a5d5-647a9948e006\" (UID: \"aa61dea3-202c-45e9-a5d5-647a9948e006\") " Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.708148 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa61dea3-202c-45e9-a5d5-647a9948e006-kube-api-access-x9pzq" (OuterVolumeSpecName: "kube-api-access-x9pzq") pod "aa61dea3-202c-45e9-a5d5-647a9948e006" (UID: "aa61dea3-202c-45e9-a5d5-647a9948e006"). InnerVolumeSpecName "kube-api-access-x9pzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.742704 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa61dea3-202c-45e9-a5d5-647a9948e006" (UID: "aa61dea3-202c-45e9-a5d5-647a9948e006"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.742751 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-config" (OuterVolumeSpecName: "config") pod "aa61dea3-202c-45e9-a5d5-647a9948e006" (UID: "aa61dea3-202c-45e9-a5d5-647a9948e006"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.743288 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa61dea3-202c-45e9-a5d5-647a9948e006" (UID: "aa61dea3-202c-45e9-a5d5-647a9948e006"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.804870 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.804923 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.804946 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9pzq\" (UniqueName: \"kubernetes.io/projected/aa61dea3-202c-45e9-a5d5-647a9948e006-kube-api-access-x9pzq\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:32 crc kubenswrapper[4846]: I0202 13:38:32.804992 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa61dea3-202c-45e9-a5d5-647a9948e006-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.055673 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Feb 02 13:38:33 crc kubenswrapper[4846]: E0202 13:38:33.056065 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa61dea3-202c-45e9-a5d5-647a9948e006" containerName="dnsmasq-dns" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.056081 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa61dea3-202c-45e9-a5d5-647a9948e006" containerName="dnsmasq-dns" Feb 02 13:38:33 crc kubenswrapper[4846]: E0202 13:38:33.056120 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa61dea3-202c-45e9-a5d5-647a9948e006" containerName="init" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.056130 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa61dea3-202c-45e9-a5d5-647a9948e006" containerName="init" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.056327 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa61dea3-202c-45e9-a5d5-647a9948e006" containerName="dnsmasq-dns" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.056976 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.061572 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.067751 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.210224 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/9f811429-9750-4803-b9e6-931cac4216aa-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.210316 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\") pod \"ovn-copy-data\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.210453 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcpmf\" (UniqueName: \"kubernetes.io/projected/9f811429-9750-4803-b9e6-931cac4216aa-kube-api-access-lcpmf\") pod \"ovn-copy-data\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.249556 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa61dea3-202c-45e9-a5d5-647a9948e006" containerID="b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f" exitCode=0 Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.249671 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" event={"ID":"aa61dea3-202c-45e9-a5d5-647a9948e006","Type":"ContainerDied","Data":"b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f"} Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.249695 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.249721 4846 scope.go:117] "RemoveContainer" containerID="b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.249707 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697b8d7675-cxfhc" event={"ID":"aa61dea3-202c-45e9-a5d5-647a9948e006","Type":"ContainerDied","Data":"930242ef569cecd9c25315921c8f62246fc5c2f17a1258e050d720ed96f9b002"} Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.252300 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" event={"ID":"9a717af0-77ea-4be4-ba01-8390c53da81c","Type":"ContainerStarted","Data":"3769afb30f73489bf34fbf983573ff8deab91b1d95e5418d2ef191f747eb9b14"} Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.252469 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.270556 4846 scope.go:117] "RemoveContainer" containerID="b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.274534 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" podStartSLOduration=3.274515957 podStartE2EDuration="3.274515957s" podCreationTimestamp="2026-02-02 13:38:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:33.273288787 +0000 UTC m=+5344.501875660" watchObservedRunningTime="2026-02-02 13:38:33.274515957 +0000 UTC m=+5344.503102820" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.292850 4846 scope.go:117] "RemoveContainer" containerID="b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.301825 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-697b8d7675-cxfhc"] Feb 02 13:38:33 crc kubenswrapper[4846]: E0202 13:38:33.302438 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f\": container with ID starting with b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f not found: ID does not exist" containerID="b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.302532 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f"} err="failed to get container status \"b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f\": rpc error: code = NotFound desc = could not find container \"b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f\": container with ID starting with b072c411a39f503f1f7016ed7212be8434e75361284b6d6c5668b0878c1a1c2f not found: ID does not exist" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.302642 4846 scope.go:117] "RemoveContainer" containerID="b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0" Feb 02 13:38:33 crc kubenswrapper[4846]: E0202 13:38:33.303171 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0\": container with ID starting with b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0 not found: ID does not exist" containerID="b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.303213 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0"} err="failed to get container status \"b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0\": rpc error: code = NotFound desc = could not find container \"b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0\": container with ID starting with b6c34706250ef712b65cdcdc8a8c9fdcbc3bee118dacc2640ebe890e6995cdc0 not found: ID does not exist" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.309676 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-697b8d7675-cxfhc"] Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.312402 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\") pod \"ovn-copy-data\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.312453 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcpmf\" (UniqueName: \"kubernetes.io/projected/9f811429-9750-4803-b9e6-931cac4216aa-kube-api-access-lcpmf\") pod \"ovn-copy-data\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.312547 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/9f811429-9750-4803-b9e6-931cac4216aa-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.316101 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.316142 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\") pod \"ovn-copy-data\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/acca076fdeb77578cab860eb46ae011491066b6ac97aa8ae64be3be83d63f9bf/globalmount\"" pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.318019 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/9f811429-9750-4803-b9e6-931cac4216aa-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.333416 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcpmf\" (UniqueName: \"kubernetes.io/projected/9f811429-9750-4803-b9e6-931cac4216aa-kube-api-access-lcpmf\") pod \"ovn-copy-data\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.368071 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\") pod \"ovn-copy-data\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.378742 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.432929 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa61dea3-202c-45e9-a5d5-647a9948e006" path="/var/lib/kubelet/pods/aa61dea3-202c-45e9-a5d5-647a9948e006/volumes" Feb 02 13:38:33 crc kubenswrapper[4846]: I0202 13:38:33.855501 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Feb 02 13:38:33 crc kubenswrapper[4846]: W0202 13:38:33.859130 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f811429_9750_4803_b9e6_931cac4216aa.slice/crio-9b01284e644d1c116e5e7cb8ecb4fc3a4ea449f93092dadd648c3dbc2f6b49ab WatchSource:0}: Error finding container 9b01284e644d1c116e5e7cb8ecb4fc3a4ea449f93092dadd648c3dbc2f6b49ab: Status 404 returned error can't find the container with id 9b01284e644d1c116e5e7cb8ecb4fc3a4ea449f93092dadd648c3dbc2f6b49ab Feb 02 13:38:34 crc kubenswrapper[4846]: I0202 13:38:34.262243 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"9f811429-9750-4803-b9e6-931cac4216aa","Type":"ContainerStarted","Data":"9b01284e644d1c116e5e7cb8ecb4fc3a4ea449f93092dadd648c3dbc2f6b49ab"} Feb 02 13:38:35 crc kubenswrapper[4846]: I0202 13:38:35.274918 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"9f811429-9750-4803-b9e6-931cac4216aa","Type":"ContainerStarted","Data":"a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4"} Feb 02 13:38:35 crc kubenswrapper[4846]: I0202 13:38:35.296993 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=2.649843087 podStartE2EDuration="3.296972214s" podCreationTimestamp="2026-02-02 13:38:32 +0000 UTC" firstStartedPulling="2026-02-02 13:38:33.861545953 +0000 UTC m=+5345.090132816" lastFinishedPulling="2026-02-02 13:38:34.50867508 +0000 UTC m=+5345.737261943" observedRunningTime="2026-02-02 13:38:35.293272104 +0000 UTC m=+5346.521858977" watchObservedRunningTime="2026-02-02 13:38:35.296972214 +0000 UTC m=+5346.525559077" Feb 02 13:38:38 crc kubenswrapper[4846]: I0202 13:38:38.110951 4846 scope.go:117] "RemoveContainer" containerID="c8b08e4c031e4ebef7e1a74092d58616de6c71c9dc6920c55aca4c82f85f0e6e" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.255889 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.257441 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.259657 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.260364 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.260554 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-zgkff" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.260775 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.278846 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.395970 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9c4be3d-913c-42f8-9847-db39fcc49afd-scripts\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.396156 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e9c4be3d-913c-42f8-9847-db39fcc49afd-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.399733 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c4be3d-913c-42f8-9847-db39fcc49afd-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.400028 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9c4be3d-913c-42f8-9847-db39fcc49afd-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.400158 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c4be3d-913c-42f8-9847-db39fcc49afd-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.400292 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c4be3d-913c-42f8-9847-db39fcc49afd-config\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.400329 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24sgk\" (UniqueName: \"kubernetes.io/projected/e9c4be3d-913c-42f8-9847-db39fcc49afd-kube-api-access-24sgk\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.502048 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c4be3d-913c-42f8-9847-db39fcc49afd-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.502152 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9c4be3d-913c-42f8-9847-db39fcc49afd-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.502187 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c4be3d-913c-42f8-9847-db39fcc49afd-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.502218 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c4be3d-913c-42f8-9847-db39fcc49afd-config\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.502237 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24sgk\" (UniqueName: \"kubernetes.io/projected/e9c4be3d-913c-42f8-9847-db39fcc49afd-kube-api-access-24sgk\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.502274 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9c4be3d-913c-42f8-9847-db39fcc49afd-scripts\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.502299 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e9c4be3d-913c-42f8-9847-db39fcc49afd-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.503605 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e9c4be3d-913c-42f8-9847-db39fcc49afd-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.503978 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c4be3d-913c-42f8-9847-db39fcc49afd-config\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.504053 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9c4be3d-913c-42f8-9847-db39fcc49afd-scripts\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.508066 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c4be3d-913c-42f8-9847-db39fcc49afd-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.508376 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9c4be3d-913c-42f8-9847-db39fcc49afd-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.515385 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c4be3d-913c-42f8-9847-db39fcc49afd-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.523520 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24sgk\" (UniqueName: \"kubernetes.io/projected/e9c4be3d-913c-42f8-9847-db39fcc49afd-kube-api-access-24sgk\") pod \"ovn-northd-0\" (UID: \"e9c4be3d-913c-42f8-9847-db39fcc49afd\") " pod="openstack/ovn-northd-0" Feb 02 13:38:40 crc kubenswrapper[4846]: I0202 13:38:40.591586 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.022064 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.159807 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.231023 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54dc9c94cc-jwf4j"] Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.231668 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" podUID="02fe5a7e-e42b-48f0-997c-13465a788818" containerName="dnsmasq-dns" containerID="cri-o://9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833" gracePeriod=10 Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.327779 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e9c4be3d-913c-42f8-9847-db39fcc49afd","Type":"ContainerStarted","Data":"3490f56c8314fd9a380c14b555f02e07b468237db063564cc6f689170dd48037"} Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.327841 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e9c4be3d-913c-42f8-9847-db39fcc49afd","Type":"ContainerStarted","Data":"84083013f2635fb237f2ced687c60d5bd75e21d731dd994be43841eef8012b07"} Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.758847 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.922292 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-dns-svc\") pod \"02fe5a7e-e42b-48f0-997c-13465a788818\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.923504 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-config\") pod \"02fe5a7e-e42b-48f0-997c-13465a788818\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.923711 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84xtg\" (UniqueName: \"kubernetes.io/projected/02fe5a7e-e42b-48f0-997c-13465a788818-kube-api-access-84xtg\") pod \"02fe5a7e-e42b-48f0-997c-13465a788818\" (UID: \"02fe5a7e-e42b-48f0-997c-13465a788818\") " Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.931728 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02fe5a7e-e42b-48f0-997c-13465a788818-kube-api-access-84xtg" (OuterVolumeSpecName: "kube-api-access-84xtg") pod "02fe5a7e-e42b-48f0-997c-13465a788818" (UID: "02fe5a7e-e42b-48f0-997c-13465a788818"). InnerVolumeSpecName "kube-api-access-84xtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.970140 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "02fe5a7e-e42b-48f0-997c-13465a788818" (UID: "02fe5a7e-e42b-48f0-997c-13465a788818"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:38:41 crc kubenswrapper[4846]: I0202 13:38:41.973695 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-config" (OuterVolumeSpecName: "config") pod "02fe5a7e-e42b-48f0-997c-13465a788818" (UID: "02fe5a7e-e42b-48f0-997c-13465a788818"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.025708 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84xtg\" (UniqueName: \"kubernetes.io/projected/02fe5a7e-e42b-48f0-997c-13465a788818-kube-api-access-84xtg\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.025749 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.025761 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02fe5a7e-e42b-48f0-997c-13465a788818-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.335732 4846 generic.go:334] "Generic (PLEG): container finished" podID="02fe5a7e-e42b-48f0-997c-13465a788818" containerID="9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833" exitCode=0 Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.335774 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" event={"ID":"02fe5a7e-e42b-48f0-997c-13465a788818","Type":"ContainerDied","Data":"9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833"} Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.335812 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.335835 4846 scope.go:117] "RemoveContainer" containerID="9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.335822 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54dc9c94cc-jwf4j" event={"ID":"02fe5a7e-e42b-48f0-997c-13465a788818","Type":"ContainerDied","Data":"b15c550733f4cad63befdd938a9d46a018f3fe30abba396d09200c03265ff980"} Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.338797 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"e9c4be3d-913c-42f8-9847-db39fcc49afd","Type":"ContainerStarted","Data":"597af7de51bcf74184cd3fbce67b48510fa038955dcfcdf0cb4abb8f60d45e47"} Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.338957 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.353273 4846 scope.go:117] "RemoveContainer" containerID="7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.376373 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.376354361 podStartE2EDuration="2.376354361s" podCreationTimestamp="2026-02-02 13:38:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:42.359214045 +0000 UTC m=+5353.587800928" watchObservedRunningTime="2026-02-02 13:38:42.376354361 +0000 UTC m=+5353.604941224" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.387779 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54dc9c94cc-jwf4j"] Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.393564 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54dc9c94cc-jwf4j"] Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.395717 4846 scope.go:117] "RemoveContainer" containerID="9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833" Feb 02 13:38:42 crc kubenswrapper[4846]: E0202 13:38:42.396730 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833\": container with ID starting with 9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833 not found: ID does not exist" containerID="9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.396782 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833"} err="failed to get container status \"9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833\": rpc error: code = NotFound desc = could not find container \"9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833\": container with ID starting with 9a0b3d6e228ef4c7463789229dc91a6c314e1c86b9ea8e3055ad8d47b49e6833 not found: ID does not exist" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.396806 4846 scope.go:117] "RemoveContainer" containerID="7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054" Feb 02 13:38:42 crc kubenswrapper[4846]: E0202 13:38:42.398921 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054\": container with ID starting with 7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054 not found: ID does not exist" containerID="7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054" Feb 02 13:38:42 crc kubenswrapper[4846]: I0202 13:38:42.398964 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054"} err="failed to get container status \"7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054\": rpc error: code = NotFound desc = could not find container \"7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054\": container with ID starting with 7a599e7aeebd7124e89f9eb9431baf40d9a18c1be778fbb564b337488696f054 not found: ID does not exist" Feb 02 13:38:43 crc kubenswrapper[4846]: I0202 13:38:43.433611 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02fe5a7e-e42b-48f0-997c-13465a788818" path="/var/lib/kubelet/pods/02fe5a7e-e42b-48f0-997c-13465a788818/volumes" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.467859 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-wbckr"] Feb 02 13:38:45 crc kubenswrapper[4846]: E0202 13:38:45.468445 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fe5a7e-e42b-48f0-997c-13465a788818" containerName="init" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.468466 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fe5a7e-e42b-48f0-997c-13465a788818" containerName="init" Feb 02 13:38:45 crc kubenswrapper[4846]: E0202 13:38:45.468484 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fe5a7e-e42b-48f0-997c-13465a788818" containerName="dnsmasq-dns" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.468490 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fe5a7e-e42b-48f0-997c-13465a788818" containerName="dnsmasq-dns" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.468674 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="02fe5a7e-e42b-48f0-997c-13465a788818" containerName="dnsmasq-dns" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.469565 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wbckr" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.488492 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wbckr"] Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.578841 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7a3d-account-create-update-h2wdw"] Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.580295 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7a3d-account-create-update-h2wdw" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.583601 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d318950f-d4db-427f-845d-b4b265d36587-operator-scripts\") pod \"keystone-db-create-wbckr\" (UID: \"d318950f-d4db-427f-845d-b4b265d36587\") " pod="openstack/keystone-db-create-wbckr" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.583788 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glhsm\" (UniqueName: \"kubernetes.io/projected/d318950f-d4db-427f-845d-b4b265d36587-kube-api-access-glhsm\") pod \"keystone-db-create-wbckr\" (UID: \"d318950f-d4db-427f-845d-b4b265d36587\") " pod="openstack/keystone-db-create-wbckr" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.585005 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.598101 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7a3d-account-create-update-h2wdw"] Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.685107 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glhsm\" (UniqueName: \"kubernetes.io/projected/d318950f-d4db-427f-845d-b4b265d36587-kube-api-access-glhsm\") pod \"keystone-db-create-wbckr\" (UID: \"d318950f-d4db-427f-845d-b4b265d36587\") " pod="openstack/keystone-db-create-wbckr" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.685983 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qz77\" (UniqueName: \"kubernetes.io/projected/e90b7676-c930-4301-8a8c-14d273d1112b-kube-api-access-8qz77\") pod \"keystone-7a3d-account-create-update-h2wdw\" (UID: \"e90b7676-c930-4301-8a8c-14d273d1112b\") " pod="openstack/keystone-7a3d-account-create-update-h2wdw" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.686297 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e90b7676-c930-4301-8a8c-14d273d1112b-operator-scripts\") pod \"keystone-7a3d-account-create-update-h2wdw\" (UID: \"e90b7676-c930-4301-8a8c-14d273d1112b\") " pod="openstack/keystone-7a3d-account-create-update-h2wdw" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.686827 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d318950f-d4db-427f-845d-b4b265d36587-operator-scripts\") pod \"keystone-db-create-wbckr\" (UID: \"d318950f-d4db-427f-845d-b4b265d36587\") " pod="openstack/keystone-db-create-wbckr" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.687597 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d318950f-d4db-427f-845d-b4b265d36587-operator-scripts\") pod \"keystone-db-create-wbckr\" (UID: \"d318950f-d4db-427f-845d-b4b265d36587\") " pod="openstack/keystone-db-create-wbckr" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.709439 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glhsm\" (UniqueName: \"kubernetes.io/projected/d318950f-d4db-427f-845d-b4b265d36587-kube-api-access-glhsm\") pod \"keystone-db-create-wbckr\" (UID: \"d318950f-d4db-427f-845d-b4b265d36587\") " pod="openstack/keystone-db-create-wbckr" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.791715 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qz77\" (UniqueName: \"kubernetes.io/projected/e90b7676-c930-4301-8a8c-14d273d1112b-kube-api-access-8qz77\") pod \"keystone-7a3d-account-create-update-h2wdw\" (UID: \"e90b7676-c930-4301-8a8c-14d273d1112b\") " pod="openstack/keystone-7a3d-account-create-update-h2wdw" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.791827 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e90b7676-c930-4301-8a8c-14d273d1112b-operator-scripts\") pod \"keystone-7a3d-account-create-update-h2wdw\" (UID: \"e90b7676-c930-4301-8a8c-14d273d1112b\") " pod="openstack/keystone-7a3d-account-create-update-h2wdw" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.792788 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e90b7676-c930-4301-8a8c-14d273d1112b-operator-scripts\") pod \"keystone-7a3d-account-create-update-h2wdw\" (UID: \"e90b7676-c930-4301-8a8c-14d273d1112b\") " pod="openstack/keystone-7a3d-account-create-update-h2wdw" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.800036 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wbckr" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.816240 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qz77\" (UniqueName: \"kubernetes.io/projected/e90b7676-c930-4301-8a8c-14d273d1112b-kube-api-access-8qz77\") pod \"keystone-7a3d-account-create-update-h2wdw\" (UID: \"e90b7676-c930-4301-8a8c-14d273d1112b\") " pod="openstack/keystone-7a3d-account-create-update-h2wdw" Feb 02 13:38:45 crc kubenswrapper[4846]: I0202 13:38:45.897483 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7a3d-account-create-update-h2wdw" Feb 02 13:38:46 crc kubenswrapper[4846]: I0202 13:38:46.354631 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wbckr"] Feb 02 13:38:46 crc kubenswrapper[4846]: I0202 13:38:46.370653 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wbckr" event={"ID":"d318950f-d4db-427f-845d-b4b265d36587","Type":"ContainerStarted","Data":"730a88a9c2588d92317d68e6df7414bd1a896ca3923e1c71051871e04e606761"} Feb 02 13:38:46 crc kubenswrapper[4846]: I0202 13:38:46.457092 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7a3d-account-create-update-h2wdw"] Feb 02 13:38:46 crc kubenswrapper[4846]: W0202 13:38:46.465856 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode90b7676_c930_4301_8a8c_14d273d1112b.slice/crio-d9a43e0429b236bf1a99dbcd592738f899eb4581bb0a67c2fd6e3f74a39cddb5 WatchSource:0}: Error finding container d9a43e0429b236bf1a99dbcd592738f899eb4581bb0a67c2fd6e3f74a39cddb5: Status 404 returned error can't find the container with id d9a43e0429b236bf1a99dbcd592738f899eb4581bb0a67c2fd6e3f74a39cddb5 Feb 02 13:38:47 crc kubenswrapper[4846]: I0202 13:38:47.379289 4846 generic.go:334] "Generic (PLEG): container finished" podID="d318950f-d4db-427f-845d-b4b265d36587" containerID="9e08b753754ca20b9dd23cd131c22e0726555c25a46b244d99d1a89c470c94b3" exitCode=0 Feb 02 13:38:47 crc kubenswrapper[4846]: I0202 13:38:47.379344 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wbckr" event={"ID":"d318950f-d4db-427f-845d-b4b265d36587","Type":"ContainerDied","Data":"9e08b753754ca20b9dd23cd131c22e0726555c25a46b244d99d1a89c470c94b3"} Feb 02 13:38:47 crc kubenswrapper[4846]: I0202 13:38:47.380930 4846 generic.go:334] "Generic (PLEG): container finished" podID="e90b7676-c930-4301-8a8c-14d273d1112b" containerID="05c09344383f9c8e5d51704d8fa580cf13550a53818d06fb1d07da6df906a258" exitCode=0 Feb 02 13:38:47 crc kubenswrapper[4846]: I0202 13:38:47.380977 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7a3d-account-create-update-h2wdw" event={"ID":"e90b7676-c930-4301-8a8c-14d273d1112b","Type":"ContainerDied","Data":"05c09344383f9c8e5d51704d8fa580cf13550a53818d06fb1d07da6df906a258"} Feb 02 13:38:47 crc kubenswrapper[4846]: I0202 13:38:47.381006 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7a3d-account-create-update-h2wdw" event={"ID":"e90b7676-c930-4301-8a8c-14d273d1112b","Type":"ContainerStarted","Data":"d9a43e0429b236bf1a99dbcd592738f899eb4581bb0a67c2fd6e3f74a39cddb5"} Feb 02 13:38:48 crc kubenswrapper[4846]: I0202 13:38:48.829495 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wbckr" Feb 02 13:38:48 crc kubenswrapper[4846]: I0202 13:38:48.838524 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7a3d-account-create-update-h2wdw" Feb 02 13:38:48 crc kubenswrapper[4846]: I0202 13:38:48.956695 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d318950f-d4db-427f-845d-b4b265d36587-operator-scripts\") pod \"d318950f-d4db-427f-845d-b4b265d36587\" (UID: \"d318950f-d4db-427f-845d-b4b265d36587\") " Feb 02 13:38:48 crc kubenswrapper[4846]: I0202 13:38:48.956834 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glhsm\" (UniqueName: \"kubernetes.io/projected/d318950f-d4db-427f-845d-b4b265d36587-kube-api-access-glhsm\") pod \"d318950f-d4db-427f-845d-b4b265d36587\" (UID: \"d318950f-d4db-427f-845d-b4b265d36587\") " Feb 02 13:38:48 crc kubenswrapper[4846]: I0202 13:38:48.957076 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qz77\" (UniqueName: \"kubernetes.io/projected/e90b7676-c930-4301-8a8c-14d273d1112b-kube-api-access-8qz77\") pod \"e90b7676-c930-4301-8a8c-14d273d1112b\" (UID: \"e90b7676-c930-4301-8a8c-14d273d1112b\") " Feb 02 13:38:48 crc kubenswrapper[4846]: I0202 13:38:48.957130 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e90b7676-c930-4301-8a8c-14d273d1112b-operator-scripts\") pod \"e90b7676-c930-4301-8a8c-14d273d1112b\" (UID: \"e90b7676-c930-4301-8a8c-14d273d1112b\") " Feb 02 13:38:48 crc kubenswrapper[4846]: I0202 13:38:48.958770 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e90b7676-c930-4301-8a8c-14d273d1112b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e90b7676-c930-4301-8a8c-14d273d1112b" (UID: "e90b7676-c930-4301-8a8c-14d273d1112b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:38:48 crc kubenswrapper[4846]: I0202 13:38:48.959198 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d318950f-d4db-427f-845d-b4b265d36587-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d318950f-d4db-427f-845d-b4b265d36587" (UID: "d318950f-d4db-427f-845d-b4b265d36587"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:38:48 crc kubenswrapper[4846]: I0202 13:38:48.966762 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e90b7676-c930-4301-8a8c-14d273d1112b-kube-api-access-8qz77" (OuterVolumeSpecName: "kube-api-access-8qz77") pod "e90b7676-c930-4301-8a8c-14d273d1112b" (UID: "e90b7676-c930-4301-8a8c-14d273d1112b"). InnerVolumeSpecName "kube-api-access-8qz77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:38:48 crc kubenswrapper[4846]: I0202 13:38:48.968073 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d318950f-d4db-427f-845d-b4b265d36587-kube-api-access-glhsm" (OuterVolumeSpecName: "kube-api-access-glhsm") pod "d318950f-d4db-427f-845d-b4b265d36587" (UID: "d318950f-d4db-427f-845d-b4b265d36587"). InnerVolumeSpecName "kube-api-access-glhsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:38:49 crc kubenswrapper[4846]: I0202 13:38:49.059572 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qz77\" (UniqueName: \"kubernetes.io/projected/e90b7676-c930-4301-8a8c-14d273d1112b-kube-api-access-8qz77\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:49 crc kubenswrapper[4846]: I0202 13:38:49.059629 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e90b7676-c930-4301-8a8c-14d273d1112b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:49 crc kubenswrapper[4846]: I0202 13:38:49.059643 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d318950f-d4db-427f-845d-b4b265d36587-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:49 crc kubenswrapper[4846]: I0202 13:38:49.059652 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glhsm\" (UniqueName: \"kubernetes.io/projected/d318950f-d4db-427f-845d-b4b265d36587-kube-api-access-glhsm\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:49 crc kubenswrapper[4846]: I0202 13:38:49.396226 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7a3d-account-create-update-h2wdw" event={"ID":"e90b7676-c930-4301-8a8c-14d273d1112b","Type":"ContainerDied","Data":"d9a43e0429b236bf1a99dbcd592738f899eb4581bb0a67c2fd6e3f74a39cddb5"} Feb 02 13:38:49 crc kubenswrapper[4846]: I0202 13:38:49.396281 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9a43e0429b236bf1a99dbcd592738f899eb4581bb0a67c2fd6e3f74a39cddb5" Feb 02 13:38:49 crc kubenswrapper[4846]: I0202 13:38:49.396217 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7a3d-account-create-update-h2wdw" Feb 02 13:38:49 crc kubenswrapper[4846]: I0202 13:38:49.400237 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wbckr" event={"ID":"d318950f-d4db-427f-845d-b4b265d36587","Type":"ContainerDied","Data":"730a88a9c2588d92317d68e6df7414bd1a896ca3923e1c71051871e04e606761"} Feb 02 13:38:49 crc kubenswrapper[4846]: I0202 13:38:49.400281 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="730a88a9c2588d92317d68e6df7414bd1a896ca3923e1c71051871e04e606761" Feb 02 13:38:49 crc kubenswrapper[4846]: I0202 13:38:49.400332 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wbckr" Feb 02 13:38:50 crc kubenswrapper[4846]: I0202 13:38:50.990577 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-h47dj"] Feb 02 13:38:50 crc kubenswrapper[4846]: E0202 13:38:50.991274 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d318950f-d4db-427f-845d-b4b265d36587" containerName="mariadb-database-create" Feb 02 13:38:50 crc kubenswrapper[4846]: I0202 13:38:50.991291 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d318950f-d4db-427f-845d-b4b265d36587" containerName="mariadb-database-create" Feb 02 13:38:50 crc kubenswrapper[4846]: E0202 13:38:50.991327 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e90b7676-c930-4301-8a8c-14d273d1112b" containerName="mariadb-account-create-update" Feb 02 13:38:50 crc kubenswrapper[4846]: I0202 13:38:50.991335 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e90b7676-c930-4301-8a8c-14d273d1112b" containerName="mariadb-account-create-update" Feb 02 13:38:50 crc kubenswrapper[4846]: I0202 13:38:50.991511 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e90b7676-c930-4301-8a8c-14d273d1112b" containerName="mariadb-account-create-update" Feb 02 13:38:50 crc kubenswrapper[4846]: I0202 13:38:50.991526 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d318950f-d4db-427f-845d-b4b265d36587" containerName="mariadb-database-create" Feb 02 13:38:50 crc kubenswrapper[4846]: I0202 13:38:50.992086 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:50 crc kubenswrapper[4846]: I0202 13:38:50.994518 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 13:38:50 crc kubenswrapper[4846]: I0202 13:38:50.995127 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dxvw2" Feb 02 13:38:50 crc kubenswrapper[4846]: I0202 13:38:50.995271 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 13:38:50 crc kubenswrapper[4846]: I0202 13:38:50.996410 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.014967 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-h47dj"] Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.089354 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-combined-ca-bundle\") pod \"keystone-db-sync-h47dj\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.089707 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8klv\" (UniqueName: \"kubernetes.io/projected/2dd78cf8-d095-4c69-849d-750d826aba63-kube-api-access-d8klv\") pod \"keystone-db-sync-h47dj\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.089948 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-config-data\") pod \"keystone-db-sync-h47dj\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.191949 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-combined-ca-bundle\") pod \"keystone-db-sync-h47dj\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.192075 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8klv\" (UniqueName: \"kubernetes.io/projected/2dd78cf8-d095-4c69-849d-750d826aba63-kube-api-access-d8klv\") pod \"keystone-db-sync-h47dj\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.192143 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-config-data\") pod \"keystone-db-sync-h47dj\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.196516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-combined-ca-bundle\") pod \"keystone-db-sync-h47dj\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.196541 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-config-data\") pod \"keystone-db-sync-h47dj\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.210070 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8klv\" (UniqueName: \"kubernetes.io/projected/2dd78cf8-d095-4c69-849d-750d826aba63-kube-api-access-d8klv\") pod \"keystone-db-sync-h47dj\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.313052 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:51 crc kubenswrapper[4846]: I0202 13:38:51.765562 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-h47dj"] Feb 02 13:38:51 crc kubenswrapper[4846]: W0202 13:38:51.768252 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2dd78cf8_d095_4c69_849d_750d826aba63.slice/crio-eee36aab8e9ed9006a66e9e79ff86bb0c0f960dfa47b0e7ca0969e0f97eab4d4 WatchSource:0}: Error finding container eee36aab8e9ed9006a66e9e79ff86bb0c0f960dfa47b0e7ca0969e0f97eab4d4: Status 404 returned error can't find the container with id eee36aab8e9ed9006a66e9e79ff86bb0c0f960dfa47b0e7ca0969e0f97eab4d4 Feb 02 13:38:52 crc kubenswrapper[4846]: I0202 13:38:52.429802 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-h47dj" event={"ID":"2dd78cf8-d095-4c69-849d-750d826aba63","Type":"ContainerStarted","Data":"6d5280498824192bfb8f8dddfa98520726b5ea653d620ec0ff92e2a970352078"} Feb 02 13:38:52 crc kubenswrapper[4846]: I0202 13:38:52.430105 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-h47dj" event={"ID":"2dd78cf8-d095-4c69-849d-750d826aba63","Type":"ContainerStarted","Data":"eee36aab8e9ed9006a66e9e79ff86bb0c0f960dfa47b0e7ca0969e0f97eab4d4"} Feb 02 13:38:52 crc kubenswrapper[4846]: I0202 13:38:52.449206 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-h47dj" podStartSLOduration=2.449186897 podStartE2EDuration="2.449186897s" podCreationTimestamp="2026-02-02 13:38:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:52.443312764 +0000 UTC m=+5363.671899647" watchObservedRunningTime="2026-02-02 13:38:52.449186897 +0000 UTC m=+5363.677773750" Feb 02 13:38:54 crc kubenswrapper[4846]: I0202 13:38:54.445433 4846 generic.go:334] "Generic (PLEG): container finished" podID="2dd78cf8-d095-4c69-849d-750d826aba63" containerID="6d5280498824192bfb8f8dddfa98520726b5ea653d620ec0ff92e2a970352078" exitCode=0 Feb 02 13:38:54 crc kubenswrapper[4846]: I0202 13:38:54.445512 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-h47dj" event={"ID":"2dd78cf8-d095-4c69-849d-750d826aba63","Type":"ContainerDied","Data":"6d5280498824192bfb8f8dddfa98520726b5ea653d620ec0ff92e2a970352078"} Feb 02 13:38:55 crc kubenswrapper[4846]: I0202 13:38:55.857776 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.027041 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-config-data\") pod \"2dd78cf8-d095-4c69-849d-750d826aba63\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.027148 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-combined-ca-bundle\") pod \"2dd78cf8-d095-4c69-849d-750d826aba63\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.027537 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8klv\" (UniqueName: \"kubernetes.io/projected/2dd78cf8-d095-4c69-849d-750d826aba63-kube-api-access-d8klv\") pod \"2dd78cf8-d095-4c69-849d-750d826aba63\" (UID: \"2dd78cf8-d095-4c69-849d-750d826aba63\") " Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.035470 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dd78cf8-d095-4c69-849d-750d826aba63-kube-api-access-d8klv" (OuterVolumeSpecName: "kube-api-access-d8klv") pod "2dd78cf8-d095-4c69-849d-750d826aba63" (UID: "2dd78cf8-d095-4c69-849d-750d826aba63"). InnerVolumeSpecName "kube-api-access-d8klv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.068423 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2dd78cf8-d095-4c69-849d-750d826aba63" (UID: "2dd78cf8-d095-4c69-849d-750d826aba63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.088389 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-config-data" (OuterVolumeSpecName: "config-data") pod "2dd78cf8-d095-4c69-849d-750d826aba63" (UID: "2dd78cf8-d095-4c69-849d-750d826aba63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.130806 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8klv\" (UniqueName: \"kubernetes.io/projected/2dd78cf8-d095-4c69-849d-750d826aba63-kube-api-access-d8klv\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.130868 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.130889 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dd78cf8-d095-4c69-849d-750d826aba63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.466497 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-h47dj" event={"ID":"2dd78cf8-d095-4c69-849d-750d826aba63","Type":"ContainerDied","Data":"eee36aab8e9ed9006a66e9e79ff86bb0c0f960dfa47b0e7ca0969e0f97eab4d4"} Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.466795 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eee36aab8e9ed9006a66e9e79ff86bb0c0f960dfa47b0e7ca0969e0f97eab4d4" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.466601 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-h47dj" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.723040 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5757586b9-6tq5q"] Feb 02 13:38:56 crc kubenswrapper[4846]: E0202 13:38:56.723728 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dd78cf8-d095-4c69-849d-750d826aba63" containerName="keystone-db-sync" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.723747 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dd78cf8-d095-4c69-849d-750d826aba63" containerName="keystone-db-sync" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.723990 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dd78cf8-d095-4c69-849d-750d826aba63" containerName="keystone-db-sync" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.724819 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.746668 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5757586b9-6tq5q"] Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.768125 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rv7wf"] Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.769862 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.774441 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.774727 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.774970 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.775116 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dxvw2" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.775224 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.789017 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rv7wf"] Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.845682 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62rwj\" (UniqueName: \"kubernetes.io/projected/b036df9b-ce4c-49f2-8cbc-47bacccabd05-kube-api-access-62rwj\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.845864 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-dns-svc\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.845904 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-nb\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.846161 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-config\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.846209 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-sb\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.947512 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-config-data\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.947604 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-credential-keys\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.947723 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-config\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.947842 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-scripts\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.947909 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-sb\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.947951 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtlbc\" (UniqueName: \"kubernetes.io/projected/89671f4a-50dc-4c74-b6ec-3c7e20a17353-kube-api-access-jtlbc\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.947990 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-fernet-keys\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.948132 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62rwj\" (UniqueName: \"kubernetes.io/projected/b036df9b-ce4c-49f2-8cbc-47bacccabd05-kube-api-access-62rwj\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.948164 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-combined-ca-bundle\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.948190 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-dns-svc\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.948205 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-nb\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.948705 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-config\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.948875 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-nb\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.948987 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-sb\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.949485 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-dns-svc\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:56 crc kubenswrapper[4846]: I0202 13:38:56.971363 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62rwj\" (UniqueName: \"kubernetes.io/projected/b036df9b-ce4c-49f2-8cbc-47bacccabd05-kube-api-access-62rwj\") pod \"dnsmasq-dns-5757586b9-6tq5q\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.049435 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-combined-ca-bundle\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.049507 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-config-data\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.049554 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-credential-keys\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.049579 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-scripts\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.049641 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtlbc\" (UniqueName: \"kubernetes.io/projected/89671f4a-50dc-4c74-b6ec-3c7e20a17353-kube-api-access-jtlbc\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.049666 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-fernet-keys\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.054231 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-scripts\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.054400 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-combined-ca-bundle\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.054462 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-fernet-keys\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.054955 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-config-data\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.055122 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-credential-keys\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.071126 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtlbc\" (UniqueName: \"kubernetes.io/projected/89671f4a-50dc-4c74-b6ec-3c7e20a17353-kube-api-access-jtlbc\") pod \"keystone-bootstrap-rv7wf\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.111301 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.126735 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.619541 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5757586b9-6tq5q"] Feb 02 13:38:57 crc kubenswrapper[4846]: I0202 13:38:57.706115 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rv7wf"] Feb 02 13:38:58 crc kubenswrapper[4846]: I0202 13:38:58.494951 4846 generic.go:334] "Generic (PLEG): container finished" podID="b036df9b-ce4c-49f2-8cbc-47bacccabd05" containerID="423c31df98a8da5193b64f08771d8d787dcb037074778f4c06f71d3f3a2d8086" exitCode=0 Feb 02 13:38:58 crc kubenswrapper[4846]: I0202 13:38:58.495071 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" event={"ID":"b036df9b-ce4c-49f2-8cbc-47bacccabd05","Type":"ContainerDied","Data":"423c31df98a8da5193b64f08771d8d787dcb037074778f4c06f71d3f3a2d8086"} Feb 02 13:38:58 crc kubenswrapper[4846]: I0202 13:38:58.495604 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" event={"ID":"b036df9b-ce4c-49f2-8cbc-47bacccabd05","Type":"ContainerStarted","Data":"5985716d5b3b663a6fbe641f671194347395dadd23dd5b13a13311fa5c45dd21"} Feb 02 13:38:58 crc kubenswrapper[4846]: I0202 13:38:58.499911 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv7wf" event={"ID":"89671f4a-50dc-4c74-b6ec-3c7e20a17353","Type":"ContainerStarted","Data":"d666ebbafb8bb66a9e900c734c5679cad2fefe8dca5e81ae52b505726fe93e18"} Feb 02 13:38:58 crc kubenswrapper[4846]: I0202 13:38:58.499992 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv7wf" event={"ID":"89671f4a-50dc-4c74-b6ec-3c7e20a17353","Type":"ContainerStarted","Data":"1f107a36c2a6b71f8befe6e95795cbd7a190101ce49ebb7d60dec741f6fa64b0"} Feb 02 13:38:58 crc kubenswrapper[4846]: I0202 13:38:58.552609 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rv7wf" podStartSLOduration=2.552590231 podStartE2EDuration="2.552590231s" podCreationTimestamp="2026-02-02 13:38:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:58.552398346 +0000 UTC m=+5369.780985229" watchObservedRunningTime="2026-02-02 13:38:58.552590231 +0000 UTC m=+5369.781177094" Feb 02 13:38:59 crc kubenswrapper[4846]: I0202 13:38:59.511404 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" event={"ID":"b036df9b-ce4c-49f2-8cbc-47bacccabd05","Type":"ContainerStarted","Data":"5201dae633bb7e7f75cbf443d2c265c3a01e7387321f251ae2d1408f0d13bd4e"} Feb 02 13:38:59 crc kubenswrapper[4846]: I0202 13:38:59.545583 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" podStartSLOduration=3.545566516 podStartE2EDuration="3.545566516s" podCreationTimestamp="2026-02-02 13:38:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:38:59.541937548 +0000 UTC m=+5370.770524421" watchObservedRunningTime="2026-02-02 13:38:59.545566516 +0000 UTC m=+5370.774153379" Feb 02 13:39:00 crc kubenswrapper[4846]: I0202 13:39:00.519097 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:39:00 crc kubenswrapper[4846]: I0202 13:39:00.648917 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 02 13:39:02 crc kubenswrapper[4846]: I0202 13:39:02.549201 4846 generic.go:334] "Generic (PLEG): container finished" podID="89671f4a-50dc-4c74-b6ec-3c7e20a17353" containerID="d666ebbafb8bb66a9e900c734c5679cad2fefe8dca5e81ae52b505726fe93e18" exitCode=0 Feb 02 13:39:02 crc kubenswrapper[4846]: I0202 13:39:02.549253 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv7wf" event={"ID":"89671f4a-50dc-4c74-b6ec-3c7e20a17353","Type":"ContainerDied","Data":"d666ebbafb8bb66a9e900c734c5679cad2fefe8dca5e81ae52b505726fe93e18"} Feb 02 13:39:03 crc kubenswrapper[4846]: I0202 13:39:03.910653 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.093913 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-scripts\") pod \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.093981 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-config-data\") pod \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.094005 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtlbc\" (UniqueName: \"kubernetes.io/projected/89671f4a-50dc-4c74-b6ec-3c7e20a17353-kube-api-access-jtlbc\") pod \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.094069 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-credential-keys\") pod \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.094116 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-combined-ca-bundle\") pod \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.094153 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-fernet-keys\") pod \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\" (UID: \"89671f4a-50dc-4c74-b6ec-3c7e20a17353\") " Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.100574 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "89671f4a-50dc-4c74-b6ec-3c7e20a17353" (UID: "89671f4a-50dc-4c74-b6ec-3c7e20a17353"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.100670 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89671f4a-50dc-4c74-b6ec-3c7e20a17353-kube-api-access-jtlbc" (OuterVolumeSpecName: "kube-api-access-jtlbc") pod "89671f4a-50dc-4c74-b6ec-3c7e20a17353" (UID: "89671f4a-50dc-4c74-b6ec-3c7e20a17353"). InnerVolumeSpecName "kube-api-access-jtlbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.106839 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "89671f4a-50dc-4c74-b6ec-3c7e20a17353" (UID: "89671f4a-50dc-4c74-b6ec-3c7e20a17353"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.106948 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-scripts" (OuterVolumeSpecName: "scripts") pod "89671f4a-50dc-4c74-b6ec-3c7e20a17353" (UID: "89671f4a-50dc-4c74-b6ec-3c7e20a17353"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.120532 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89671f4a-50dc-4c74-b6ec-3c7e20a17353" (UID: "89671f4a-50dc-4c74-b6ec-3c7e20a17353"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.131803 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-config-data" (OuterVolumeSpecName: "config-data") pod "89671f4a-50dc-4c74-b6ec-3c7e20a17353" (UID: "89671f4a-50dc-4c74-b6ec-3c7e20a17353"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.196678 4846 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.196717 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.196727 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.196738 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.196747 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89671f4a-50dc-4c74-b6ec-3c7e20a17353-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.196759 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtlbc\" (UniqueName: \"kubernetes.io/projected/89671f4a-50dc-4c74-b6ec-3c7e20a17353-kube-api-access-jtlbc\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.568865 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv7wf" event={"ID":"89671f4a-50dc-4c74-b6ec-3c7e20a17353","Type":"ContainerDied","Data":"1f107a36c2a6b71f8befe6e95795cbd7a190101ce49ebb7d60dec741f6fa64b0"} Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.568911 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f107a36c2a6b71f8befe6e95795cbd7a190101ce49ebb7d60dec741f6fa64b0" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.568962 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv7wf" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.696328 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rv7wf"] Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.703234 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rv7wf"] Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.776206 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2qq96"] Feb 02 13:39:04 crc kubenswrapper[4846]: E0202 13:39:04.776853 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89671f4a-50dc-4c74-b6ec-3c7e20a17353" containerName="keystone-bootstrap" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.776873 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="89671f4a-50dc-4c74-b6ec-3c7e20a17353" containerName="keystone-bootstrap" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.777039 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="89671f4a-50dc-4c74-b6ec-3c7e20a17353" containerName="keystone-bootstrap" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.777587 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.783351 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.787513 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.787737 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.787850 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.787963 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dxvw2" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.793541 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2qq96"] Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.919738 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-fernet-keys\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.919814 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-credential-keys\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.919853 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtndw\" (UniqueName: \"kubernetes.io/projected/b03f2dde-e1dd-4f93-a8b4-215163728bbd-kube-api-access-xtndw\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.919929 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-config-data\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.920031 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-combined-ca-bundle\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:04 crc kubenswrapper[4846]: I0202 13:39:04.920100 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-scripts\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.021408 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-combined-ca-bundle\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.021481 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-scripts\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.021523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-fernet-keys\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.021552 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-credential-keys\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.021576 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtndw\" (UniqueName: \"kubernetes.io/projected/b03f2dde-e1dd-4f93-a8b4-215163728bbd-kube-api-access-xtndw\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.021607 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-config-data\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.026252 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-credential-keys\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.026368 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-combined-ca-bundle\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.026578 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-config-data\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.030702 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-fernet-keys\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.034159 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-scripts\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.040066 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtndw\" (UniqueName: \"kubernetes.io/projected/b03f2dde-e1dd-4f93-a8b4-215163728bbd-kube-api-access-xtndw\") pod \"keystone-bootstrap-2qq96\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.095876 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.434600 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89671f4a-50dc-4c74-b6ec-3c7e20a17353" path="/var/lib/kubelet/pods/89671f4a-50dc-4c74-b6ec-3c7e20a17353/volumes" Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.527561 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2qq96"] Feb 02 13:39:05 crc kubenswrapper[4846]: I0202 13:39:05.579577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2qq96" event={"ID":"b03f2dde-e1dd-4f93-a8b4-215163728bbd","Type":"ContainerStarted","Data":"670e76972c0fc0910667c0ad1a5d8809b8a46788aafbd9fec26d65fd87b0ce77"} Feb 02 13:39:06 crc kubenswrapper[4846]: I0202 13:39:06.590062 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2qq96" event={"ID":"b03f2dde-e1dd-4f93-a8b4-215163728bbd","Type":"ContainerStarted","Data":"23cf60ad08070e0a533b416658baa04fbb571b1b7ce7abd32d62d60c6ee62746"} Feb 02 13:39:06 crc kubenswrapper[4846]: I0202 13:39:06.611013 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2qq96" podStartSLOduration=2.610995096 podStartE2EDuration="2.610995096s" podCreationTimestamp="2026-02-02 13:39:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:39:06.607004189 +0000 UTC m=+5377.835591062" watchObservedRunningTime="2026-02-02 13:39:06.610995096 +0000 UTC m=+5377.839581959" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.112879 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.180487 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55cb6fc89-4lctn"] Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.180840 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" podUID="9a717af0-77ea-4be4-ba01-8390c53da81c" containerName="dnsmasq-dns" containerID="cri-o://3769afb30f73489bf34fbf983573ff8deab91b1d95e5418d2ef191f747eb9b14" gracePeriod=10 Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.602439 4846 generic.go:334] "Generic (PLEG): container finished" podID="9a717af0-77ea-4be4-ba01-8390c53da81c" containerID="3769afb30f73489bf34fbf983573ff8deab91b1d95e5418d2ef191f747eb9b14" exitCode=0 Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.603836 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" event={"ID":"9a717af0-77ea-4be4-ba01-8390c53da81c","Type":"ContainerDied","Data":"3769afb30f73489bf34fbf983573ff8deab91b1d95e5418d2ef191f747eb9b14"} Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.603873 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" event={"ID":"9a717af0-77ea-4be4-ba01-8390c53da81c","Type":"ContainerDied","Data":"671eb996cb8edd21d55614ca9f6fe8bf7fe7abbdfc5fa7b84e94269bb553b795"} Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.603891 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="671eb996cb8edd21d55614ca9f6fe8bf7fe7abbdfc5fa7b84e94269bb553b795" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.675176 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.869402 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-sb\") pod \"9a717af0-77ea-4be4-ba01-8390c53da81c\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.869453 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-dns-svc\") pod \"9a717af0-77ea-4be4-ba01-8390c53da81c\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.869500 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-config\") pod \"9a717af0-77ea-4be4-ba01-8390c53da81c\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.869523 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-nb\") pod \"9a717af0-77ea-4be4-ba01-8390c53da81c\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.869578 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpjvd\" (UniqueName: \"kubernetes.io/projected/9a717af0-77ea-4be4-ba01-8390c53da81c-kube-api-access-gpjvd\") pod \"9a717af0-77ea-4be4-ba01-8390c53da81c\" (UID: \"9a717af0-77ea-4be4-ba01-8390c53da81c\") " Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.874590 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a717af0-77ea-4be4-ba01-8390c53da81c-kube-api-access-gpjvd" (OuterVolumeSpecName: "kube-api-access-gpjvd") pod "9a717af0-77ea-4be4-ba01-8390c53da81c" (UID: "9a717af0-77ea-4be4-ba01-8390c53da81c"). InnerVolumeSpecName "kube-api-access-gpjvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.908760 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a717af0-77ea-4be4-ba01-8390c53da81c" (UID: "9a717af0-77ea-4be4-ba01-8390c53da81c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.911937 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9a717af0-77ea-4be4-ba01-8390c53da81c" (UID: "9a717af0-77ea-4be4-ba01-8390c53da81c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.916138 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9a717af0-77ea-4be4-ba01-8390c53da81c" (UID: "9a717af0-77ea-4be4-ba01-8390c53da81c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.918576 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-config" (OuterVolumeSpecName: "config") pod "9a717af0-77ea-4be4-ba01-8390c53da81c" (UID: "9a717af0-77ea-4be4-ba01-8390c53da81c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.971332 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.971383 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.971392 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.971400 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a717af0-77ea-4be4-ba01-8390c53da81c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:07 crc kubenswrapper[4846]: I0202 13:39:07.971410 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpjvd\" (UniqueName: \"kubernetes.io/projected/9a717af0-77ea-4be4-ba01-8390c53da81c-kube-api-access-gpjvd\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:08 crc kubenswrapper[4846]: I0202 13:39:08.612187 4846 generic.go:334] "Generic (PLEG): container finished" podID="b03f2dde-e1dd-4f93-a8b4-215163728bbd" containerID="23cf60ad08070e0a533b416658baa04fbb571b1b7ce7abd32d62d60c6ee62746" exitCode=0 Feb 02 13:39:08 crc kubenswrapper[4846]: I0202 13:39:08.612236 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2qq96" event={"ID":"b03f2dde-e1dd-4f93-a8b4-215163728bbd","Type":"ContainerDied","Data":"23cf60ad08070e0a533b416658baa04fbb571b1b7ce7abd32d62d60c6ee62746"} Feb 02 13:39:08 crc kubenswrapper[4846]: I0202 13:39:08.612410 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cb6fc89-4lctn" Feb 02 13:39:08 crc kubenswrapper[4846]: I0202 13:39:08.654936 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55cb6fc89-4lctn"] Feb 02 13:39:08 crc kubenswrapper[4846]: I0202 13:39:08.662867 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55cb6fc89-4lctn"] Feb 02 13:39:09 crc kubenswrapper[4846]: I0202 13:39:09.435826 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a717af0-77ea-4be4-ba01-8390c53da81c" path="/var/lib/kubelet/pods/9a717af0-77ea-4be4-ba01-8390c53da81c/volumes" Feb 02 13:39:09 crc kubenswrapper[4846]: I0202 13:39:09.936575 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.105804 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-credential-keys\") pod \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.105851 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-fernet-keys\") pod \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.105900 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-config-data\") pod \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.105995 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-combined-ca-bundle\") pod \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.106078 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-scripts\") pod \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.106102 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtndw\" (UniqueName: \"kubernetes.io/projected/b03f2dde-e1dd-4f93-a8b4-215163728bbd-kube-api-access-xtndw\") pod \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\" (UID: \"b03f2dde-e1dd-4f93-a8b4-215163728bbd\") " Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.110789 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-scripts" (OuterVolumeSpecName: "scripts") pod "b03f2dde-e1dd-4f93-a8b4-215163728bbd" (UID: "b03f2dde-e1dd-4f93-a8b4-215163728bbd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.110945 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03f2dde-e1dd-4f93-a8b4-215163728bbd-kube-api-access-xtndw" (OuterVolumeSpecName: "kube-api-access-xtndw") pod "b03f2dde-e1dd-4f93-a8b4-215163728bbd" (UID: "b03f2dde-e1dd-4f93-a8b4-215163728bbd"). InnerVolumeSpecName "kube-api-access-xtndw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.111165 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b03f2dde-e1dd-4f93-a8b4-215163728bbd" (UID: "b03f2dde-e1dd-4f93-a8b4-215163728bbd"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.112296 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b03f2dde-e1dd-4f93-a8b4-215163728bbd" (UID: "b03f2dde-e1dd-4f93-a8b4-215163728bbd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.134980 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b03f2dde-e1dd-4f93-a8b4-215163728bbd" (UID: "b03f2dde-e1dd-4f93-a8b4-215163728bbd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.138864 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-config-data" (OuterVolumeSpecName: "config-data") pod "b03f2dde-e1dd-4f93-a8b4-215163728bbd" (UID: "b03f2dde-e1dd-4f93-a8b4-215163728bbd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.207723 4846 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.207963 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.207974 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.207985 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.207994 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03f2dde-e1dd-4f93-a8b4-215163728bbd-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.208003 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtndw\" (UniqueName: \"kubernetes.io/projected/b03f2dde-e1dd-4f93-a8b4-215163728bbd-kube-api-access-xtndw\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.631662 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2qq96" event={"ID":"b03f2dde-e1dd-4f93-a8b4-215163728bbd","Type":"ContainerDied","Data":"670e76972c0fc0910667c0ad1a5d8809b8a46788aafbd9fec26d65fd87b0ce77"} Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.631704 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="670e76972c0fc0910667c0ad1a5d8809b8a46788aafbd9fec26d65fd87b0ce77" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.631756 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2qq96" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.833801 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-964849ccd-5gkkw"] Feb 02 13:39:10 crc kubenswrapper[4846]: E0202 13:39:10.834247 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03f2dde-e1dd-4f93-a8b4-215163728bbd" containerName="keystone-bootstrap" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.834279 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03f2dde-e1dd-4f93-a8b4-215163728bbd" containerName="keystone-bootstrap" Feb 02 13:39:10 crc kubenswrapper[4846]: E0202 13:39:10.834302 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a717af0-77ea-4be4-ba01-8390c53da81c" containerName="init" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.834311 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a717af0-77ea-4be4-ba01-8390c53da81c" containerName="init" Feb 02 13:39:10 crc kubenswrapper[4846]: E0202 13:39:10.834334 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a717af0-77ea-4be4-ba01-8390c53da81c" containerName="dnsmasq-dns" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.834345 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a717af0-77ea-4be4-ba01-8390c53da81c" containerName="dnsmasq-dns" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.834564 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03f2dde-e1dd-4f93-a8b4-215163728bbd" containerName="keystone-bootstrap" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.834590 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a717af0-77ea-4be4-ba01-8390c53da81c" containerName="dnsmasq-dns" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.835252 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.838431 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.838431 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.841358 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.841359 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.844297 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.846031 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dxvw2" Feb 02 13:39:10 crc kubenswrapper[4846]: I0202 13:39:10.857143 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-964849ccd-5gkkw"] Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.023956 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-combined-ca-bundle\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.024859 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-fernet-keys\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.024923 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-internal-tls-certs\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.024978 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-public-tls-certs\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.025040 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-config-data\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.025425 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-credential-keys\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.025758 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-scripts\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.025846 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f9cc\" (UniqueName: \"kubernetes.io/projected/900a578d-a369-4470-a522-ee6b4cf5019d-kube-api-access-2f9cc\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.127372 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-scripts\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.127444 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f9cc\" (UniqueName: \"kubernetes.io/projected/900a578d-a369-4470-a522-ee6b4cf5019d-kube-api-access-2f9cc\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.127511 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-combined-ca-bundle\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.127539 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-fernet-keys\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.127559 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-internal-tls-certs\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.127573 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-public-tls-certs\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.127593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-config-data\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.127617 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-credential-keys\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.132150 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-credential-keys\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.132170 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-internal-tls-certs\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.133066 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-scripts\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.133179 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-combined-ca-bundle\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.133753 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-config-data\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.134567 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-fernet-keys\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.143310 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f9cc\" (UniqueName: \"kubernetes.io/projected/900a578d-a369-4470-a522-ee6b4cf5019d-kube-api-access-2f9cc\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.146119 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/900a578d-a369-4470-a522-ee6b4cf5019d-public-tls-certs\") pod \"keystone-964849ccd-5gkkw\" (UID: \"900a578d-a369-4470-a522-ee6b4cf5019d\") " pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.152715 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:11 crc kubenswrapper[4846]: I0202 13:39:11.654132 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-964849ccd-5gkkw"] Feb 02 13:39:11 crc kubenswrapper[4846]: W0202 13:39:11.659805 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod900a578d_a369_4470_a522_ee6b4cf5019d.slice/crio-b5b1910ac4dbc167bc03e3c096eae8d63b42f06043db71039a1fe6ace0fab53d WatchSource:0}: Error finding container b5b1910ac4dbc167bc03e3c096eae8d63b42f06043db71039a1fe6ace0fab53d: Status 404 returned error can't find the container with id b5b1910ac4dbc167bc03e3c096eae8d63b42f06043db71039a1fe6ace0fab53d Feb 02 13:39:12 crc kubenswrapper[4846]: I0202 13:39:12.651003 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-964849ccd-5gkkw" event={"ID":"900a578d-a369-4470-a522-ee6b4cf5019d","Type":"ContainerStarted","Data":"30f30714d2e5cd4363680045bf044bc71dda4a8a3c98a1537304a00c0cd829d3"} Feb 02 13:39:12 crc kubenswrapper[4846]: I0202 13:39:12.651371 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-964849ccd-5gkkw" event={"ID":"900a578d-a369-4470-a522-ee6b4cf5019d","Type":"ContainerStarted","Data":"b5b1910ac4dbc167bc03e3c096eae8d63b42f06043db71039a1fe6ace0fab53d"} Feb 02 13:39:12 crc kubenswrapper[4846]: I0202 13:39:12.651390 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:12 crc kubenswrapper[4846]: I0202 13:39:12.678107 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-964849ccd-5gkkw" podStartSLOduration=2.678081559 podStartE2EDuration="2.678081559s" podCreationTimestamp="2026-02-02 13:39:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:39:12.668808133 +0000 UTC m=+5383.897395006" watchObservedRunningTime="2026-02-02 13:39:12.678081559 +0000 UTC m=+5383.906668422" Feb 02 13:39:30 crc kubenswrapper[4846]: I0202 13:39:30.478640 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:39:30 crc kubenswrapper[4846]: I0202 13:39:30.479202 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.239135 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j6r4n"] Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.242014 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.251248 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6r4n"] Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.340338 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-utilities\") pod \"redhat-marketplace-j6r4n\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.340414 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-catalog-content\") pod \"redhat-marketplace-j6r4n\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.340716 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvml2\" (UniqueName: \"kubernetes.io/projected/f6039dd7-16d2-4d2c-87b0-460f275e68a3-kube-api-access-kvml2\") pod \"redhat-marketplace-j6r4n\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.442186 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvml2\" (UniqueName: \"kubernetes.io/projected/f6039dd7-16d2-4d2c-87b0-460f275e68a3-kube-api-access-kvml2\") pod \"redhat-marketplace-j6r4n\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.442338 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-utilities\") pod \"redhat-marketplace-j6r4n\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.442385 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-catalog-content\") pod \"redhat-marketplace-j6r4n\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.442915 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-utilities\") pod \"redhat-marketplace-j6r4n\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.443012 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-catalog-content\") pod \"redhat-marketplace-j6r4n\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.464371 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvml2\" (UniqueName: \"kubernetes.io/projected/f6039dd7-16d2-4d2c-87b0-460f275e68a3-kube-api-access-kvml2\") pod \"redhat-marketplace-j6r4n\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:35 crc kubenswrapper[4846]: I0202 13:39:35.564970 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:36 crc kubenswrapper[4846]: I0202 13:39:36.223935 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6r4n"] Feb 02 13:39:36 crc kubenswrapper[4846]: I0202 13:39:36.841465 4846 generic.go:334] "Generic (PLEG): container finished" podID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerID="087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c" exitCode=0 Feb 02 13:39:36 crc kubenswrapper[4846]: I0202 13:39:36.841530 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6r4n" event={"ID":"f6039dd7-16d2-4d2c-87b0-460f275e68a3","Type":"ContainerDied","Data":"087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c"} Feb 02 13:39:36 crc kubenswrapper[4846]: I0202 13:39:36.841789 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6r4n" event={"ID":"f6039dd7-16d2-4d2c-87b0-460f275e68a3","Type":"ContainerStarted","Data":"c65197855c0872c4709489018d492a48f984368834045f41b4d2c8c6f883729b"} Feb 02 13:39:36 crc kubenswrapper[4846]: I0202 13:39:36.843428 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 13:39:37 crc kubenswrapper[4846]: I0202 13:39:37.850913 4846 generic.go:334] "Generic (PLEG): container finished" podID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerID="02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d" exitCode=0 Feb 02 13:39:37 crc kubenswrapper[4846]: I0202 13:39:37.850982 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6r4n" event={"ID":"f6039dd7-16d2-4d2c-87b0-460f275e68a3","Type":"ContainerDied","Data":"02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d"} Feb 02 13:39:38 crc kubenswrapper[4846]: I0202 13:39:38.862614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6r4n" event={"ID":"f6039dd7-16d2-4d2c-87b0-460f275e68a3","Type":"ContainerStarted","Data":"f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75"} Feb 02 13:39:42 crc kubenswrapper[4846]: I0202 13:39:42.741142 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-964849ccd-5gkkw" Feb 02 13:39:42 crc kubenswrapper[4846]: I0202 13:39:42.763913 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j6r4n" podStartSLOduration=6.241302214 podStartE2EDuration="7.76389267s" podCreationTimestamp="2026-02-02 13:39:35 +0000 UTC" firstStartedPulling="2026-02-02 13:39:36.84307488 +0000 UTC m=+5408.071661753" lastFinishedPulling="2026-02-02 13:39:38.365665346 +0000 UTC m=+5409.594252209" observedRunningTime="2026-02-02 13:39:38.88833868 +0000 UTC m=+5410.116925563" watchObservedRunningTime="2026-02-02 13:39:42.76389267 +0000 UTC m=+5413.992479533" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.489667 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.491480 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.496048 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-c6px7" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.496094 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.496056 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.507259 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.565163 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.565509 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.618261 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.624321 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-openstack-config-secret\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.624443 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-combined-ca-bundle\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.624518 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/143a8135-2e86-46cf-a748-b7429af615df-openstack-config\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.624571 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scw2m\" (UniqueName: \"kubernetes.io/projected/143a8135-2e86-46cf-a748-b7429af615df-kube-api-access-scw2m\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.725942 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-openstack-config-secret\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.726296 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-combined-ca-bundle\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.726447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/143a8135-2e86-46cf-a748-b7429af615df-openstack-config\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.726576 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scw2m\" (UniqueName: \"kubernetes.io/projected/143a8135-2e86-46cf-a748-b7429af615df-kube-api-access-scw2m\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.727477 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/143a8135-2e86-46cf-a748-b7429af615df-openstack-config\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.732591 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-openstack-config-secret\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.732765 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-combined-ca-bundle\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.746247 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scw2m\" (UniqueName: \"kubernetes.io/projected/143a8135-2e86-46cf-a748-b7429af615df-kube-api-access-scw2m\") pod \"openstackclient\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.810227 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 13:39:45 crc kubenswrapper[4846]: I0202 13:39:45.992608 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:46 crc kubenswrapper[4846]: I0202 13:39:46.047803 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6r4n"] Feb 02 13:39:46 crc kubenswrapper[4846]: I0202 13:39:46.248782 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 13:39:46 crc kubenswrapper[4846]: I0202 13:39:46.955382 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"143a8135-2e86-46cf-a748-b7429af615df","Type":"ContainerStarted","Data":"495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd"} Feb 02 13:39:46 crc kubenswrapper[4846]: I0202 13:39:46.955720 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"143a8135-2e86-46cf-a748-b7429af615df","Type":"ContainerStarted","Data":"556943b684dcfb39cafb25bae19dea5b4a64753f6ec2f763277b9dea2556156f"} Feb 02 13:39:46 crc kubenswrapper[4846]: I0202 13:39:46.971658 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.971598008 podStartE2EDuration="1.971598008s" podCreationTimestamp="2026-02-02 13:39:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:39:46.970434279 +0000 UTC m=+5418.199021162" watchObservedRunningTime="2026-02-02 13:39:46.971598008 +0000 UTC m=+5418.200184871" Feb 02 13:39:47 crc kubenswrapper[4846]: I0202 13:39:47.963897 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j6r4n" podUID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerName="registry-server" containerID="cri-o://f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75" gracePeriod=2 Feb 02 13:39:48 crc kubenswrapper[4846]: I0202 13:39:48.954574 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:48 crc kubenswrapper[4846]: I0202 13:39:48.980106 4846 generic.go:334] "Generic (PLEG): container finished" podID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerID="f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75" exitCode=0 Feb 02 13:39:48 crc kubenswrapper[4846]: I0202 13:39:48.980145 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6r4n" event={"ID":"f6039dd7-16d2-4d2c-87b0-460f275e68a3","Type":"ContainerDied","Data":"f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75"} Feb 02 13:39:48 crc kubenswrapper[4846]: I0202 13:39:48.980175 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6r4n" event={"ID":"f6039dd7-16d2-4d2c-87b0-460f275e68a3","Type":"ContainerDied","Data":"c65197855c0872c4709489018d492a48f984368834045f41b4d2c8c6f883729b"} Feb 02 13:39:48 crc kubenswrapper[4846]: I0202 13:39:48.980180 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6r4n" Feb 02 13:39:48 crc kubenswrapper[4846]: I0202 13:39:48.980210 4846 scope.go:117] "RemoveContainer" containerID="f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.001411 4846 scope.go:117] "RemoveContainer" containerID="02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.021003 4846 scope.go:117] "RemoveContainer" containerID="087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.069006 4846 scope.go:117] "RemoveContainer" containerID="f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75" Feb 02 13:39:49 crc kubenswrapper[4846]: E0202 13:39:49.069859 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75\": container with ID starting with f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75 not found: ID does not exist" containerID="f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.070004 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75"} err="failed to get container status \"f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75\": rpc error: code = NotFound desc = could not find container \"f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75\": container with ID starting with f351bb48854fdd6a29d2e98e77ba31d321b6fc1fd13c2531f2f35387c0d36a75 not found: ID does not exist" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.070120 4846 scope.go:117] "RemoveContainer" containerID="02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d" Feb 02 13:39:49 crc kubenswrapper[4846]: E0202 13:39:49.070910 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d\": container with ID starting with 02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d not found: ID does not exist" containerID="02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.070942 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d"} err="failed to get container status \"02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d\": rpc error: code = NotFound desc = could not find container \"02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d\": container with ID starting with 02f17a823225ca0049496b83316e91aa6c68df398dd5f99f396c6aae404ec54d not found: ID does not exist" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.070964 4846 scope.go:117] "RemoveContainer" containerID="087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c" Feb 02 13:39:49 crc kubenswrapper[4846]: E0202 13:39:49.071153 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c\": container with ID starting with 087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c not found: ID does not exist" containerID="087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.071173 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c"} err="failed to get container status \"087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c\": rpc error: code = NotFound desc = could not find container \"087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c\": container with ID starting with 087af29320b7f3768edeacac2f1c289443c273e6fa4d038e6033756ecc68fb1c not found: ID does not exist" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.084452 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-catalog-content\") pod \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.084649 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvml2\" (UniqueName: \"kubernetes.io/projected/f6039dd7-16d2-4d2c-87b0-460f275e68a3-kube-api-access-kvml2\") pod \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.084710 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-utilities\") pod \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\" (UID: \"f6039dd7-16d2-4d2c-87b0-460f275e68a3\") " Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.086540 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-utilities" (OuterVolumeSpecName: "utilities") pod "f6039dd7-16d2-4d2c-87b0-460f275e68a3" (UID: "f6039dd7-16d2-4d2c-87b0-460f275e68a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.090676 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6039dd7-16d2-4d2c-87b0-460f275e68a3-kube-api-access-kvml2" (OuterVolumeSpecName: "kube-api-access-kvml2") pod "f6039dd7-16d2-4d2c-87b0-460f275e68a3" (UID: "f6039dd7-16d2-4d2c-87b0-460f275e68a3"). InnerVolumeSpecName "kube-api-access-kvml2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.107848 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6039dd7-16d2-4d2c-87b0-460f275e68a3" (UID: "f6039dd7-16d2-4d2c-87b0-460f275e68a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.186576 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvml2\" (UniqueName: \"kubernetes.io/projected/f6039dd7-16d2-4d2c-87b0-460f275e68a3-kube-api-access-kvml2\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.186613 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.186648 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6039dd7-16d2-4d2c-87b0-460f275e68a3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.312558 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6r4n"] Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.321639 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6r4n"] Feb 02 13:39:49 crc kubenswrapper[4846]: I0202 13:39:49.432724 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" path="/var/lib/kubelet/pods/f6039dd7-16d2-4d2c-87b0-460f275e68a3/volumes" Feb 02 13:40:00 crc kubenswrapper[4846]: I0202 13:40:00.479551 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:40:00 crc kubenswrapper[4846]: I0202 13:40:00.480183 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:40:30 crc kubenswrapper[4846]: I0202 13:40:30.479332 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:40:30 crc kubenswrapper[4846]: I0202 13:40:30.480002 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:40:30 crc kubenswrapper[4846]: I0202 13:40:30.480060 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 13:40:30 crc kubenswrapper[4846]: I0202 13:40:30.480897 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cc41b07cc0804d15d8ee13c8c57c6dfb00243755829e268a57694312efd7b440"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 13:40:30 crc kubenswrapper[4846]: I0202 13:40:30.480954 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://cc41b07cc0804d15d8ee13c8c57c6dfb00243755829e268a57694312efd7b440" gracePeriod=600 Feb 02 13:40:31 crc kubenswrapper[4846]: I0202 13:40:31.302387 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="cc41b07cc0804d15d8ee13c8c57c6dfb00243755829e268a57694312efd7b440" exitCode=0 Feb 02 13:40:31 crc kubenswrapper[4846]: I0202 13:40:31.302459 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"cc41b07cc0804d15d8ee13c8c57c6dfb00243755829e268a57694312efd7b440"} Feb 02 13:40:31 crc kubenswrapper[4846]: I0202 13:40:31.302880 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4"} Feb 02 13:40:31 crc kubenswrapper[4846]: I0202 13:40:31.302913 4846 scope.go:117] "RemoveContainer" containerID="25e943a44a43e77f74c9573160473b8d05f63cbc6621dc66dd0fba0389fdf1f5" Feb 02 13:40:40 crc kubenswrapper[4846]: I0202 13:40:40.863348 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-79rbt"] Feb 02 13:40:40 crc kubenswrapper[4846]: E0202 13:40:40.872969 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerName="extract-content" Feb 02 13:40:40 crc kubenswrapper[4846]: I0202 13:40:40.873006 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerName="extract-content" Feb 02 13:40:40 crc kubenswrapper[4846]: E0202 13:40:40.873056 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerName="extract-utilities" Feb 02 13:40:40 crc kubenswrapper[4846]: I0202 13:40:40.873065 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerName="extract-utilities" Feb 02 13:40:40 crc kubenswrapper[4846]: E0202 13:40:40.873081 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerName="registry-server" Feb 02 13:40:40 crc kubenswrapper[4846]: I0202 13:40:40.873090 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerName="registry-server" Feb 02 13:40:40 crc kubenswrapper[4846]: I0202 13:40:40.873462 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6039dd7-16d2-4d2c-87b0-460f275e68a3" containerName="registry-server" Feb 02 13:40:40 crc kubenswrapper[4846]: I0202 13:40:40.879059 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:40 crc kubenswrapper[4846]: I0202 13:40:40.882531 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-79rbt"] Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.005785 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-utilities\") pod \"redhat-operators-79rbt\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.005862 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-catalog-content\") pod \"redhat-operators-79rbt\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.006052 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m7q5\" (UniqueName: \"kubernetes.io/projected/786c7b50-d4ba-4abb-9073-f630d3d67a00-kube-api-access-6m7q5\") pod \"redhat-operators-79rbt\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.107593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-catalog-content\") pod \"redhat-operators-79rbt\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.108070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m7q5\" (UniqueName: \"kubernetes.io/projected/786c7b50-d4ba-4abb-9073-f630d3d67a00-kube-api-access-6m7q5\") pod \"redhat-operators-79rbt\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.108129 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-utilities\") pod \"redhat-operators-79rbt\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.108135 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-catalog-content\") pod \"redhat-operators-79rbt\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.108730 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-utilities\") pod \"redhat-operators-79rbt\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.129992 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m7q5\" (UniqueName: \"kubernetes.io/projected/786c7b50-d4ba-4abb-9073-f630d3d67a00-kube-api-access-6m7q5\") pod \"redhat-operators-79rbt\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.198669 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:41 crc kubenswrapper[4846]: I0202 13:40:41.672039 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-79rbt"] Feb 02 13:40:42 crc kubenswrapper[4846]: I0202 13:40:42.405984 4846 generic.go:334] "Generic (PLEG): container finished" podID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerID="7dae938dcc962a9b1cd1d7222b563c8731d08e5ccf25936fe79ed54bd03e6008" exitCode=0 Feb 02 13:40:42 crc kubenswrapper[4846]: I0202 13:40:42.406267 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79rbt" event={"ID":"786c7b50-d4ba-4abb-9073-f630d3d67a00","Type":"ContainerDied","Data":"7dae938dcc962a9b1cd1d7222b563c8731d08e5ccf25936fe79ed54bd03e6008"} Feb 02 13:40:42 crc kubenswrapper[4846]: I0202 13:40:42.406336 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79rbt" event={"ID":"786c7b50-d4ba-4abb-9073-f630d3d67a00","Type":"ContainerStarted","Data":"d5e085b76632ebaf86bdd6285b283a06995313f0f54a6bea7e714a8646eb9355"} Feb 02 13:40:43 crc kubenswrapper[4846]: I0202 13:40:43.418320 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79rbt" event={"ID":"786c7b50-d4ba-4abb-9073-f630d3d67a00","Type":"ContainerStarted","Data":"45e96db01cad0dd0296aec20adbcdb058e45e4c997208a8083e96f519b173adf"} Feb 02 13:40:44 crc kubenswrapper[4846]: I0202 13:40:44.428860 4846 generic.go:334] "Generic (PLEG): container finished" podID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerID="45e96db01cad0dd0296aec20adbcdb058e45e4c997208a8083e96f519b173adf" exitCode=0 Feb 02 13:40:44 crc kubenswrapper[4846]: I0202 13:40:44.428934 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79rbt" event={"ID":"786c7b50-d4ba-4abb-9073-f630d3d67a00","Type":"ContainerDied","Data":"45e96db01cad0dd0296aec20adbcdb058e45e4c997208a8083e96f519b173adf"} Feb 02 13:40:45 crc kubenswrapper[4846]: I0202 13:40:45.439116 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79rbt" event={"ID":"786c7b50-d4ba-4abb-9073-f630d3d67a00","Type":"ContainerStarted","Data":"8d0a1ba5019c134f939fb29ee531db7c053c4da2018d94c2063c48762d6c64d7"} Feb 02 13:40:45 crc kubenswrapper[4846]: I0202 13:40:45.460418 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-79rbt" podStartSLOduration=3.027880353 podStartE2EDuration="5.460392278s" podCreationTimestamp="2026-02-02 13:40:40 +0000 UTC" firstStartedPulling="2026-02-02 13:40:42.407836955 +0000 UTC m=+5473.636423818" lastFinishedPulling="2026-02-02 13:40:44.84034888 +0000 UTC m=+5476.068935743" observedRunningTime="2026-02-02 13:40:45.459144608 +0000 UTC m=+5476.687731491" watchObservedRunningTime="2026-02-02 13:40:45.460392278 +0000 UTC m=+5476.688979161" Feb 02 13:40:51 crc kubenswrapper[4846]: I0202 13:40:51.198919 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:51 crc kubenswrapper[4846]: I0202 13:40:51.199499 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:51 crc kubenswrapper[4846]: I0202 13:40:51.241936 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:51 crc kubenswrapper[4846]: I0202 13:40:51.529903 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:51 crc kubenswrapper[4846]: I0202 13:40:51.587477 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-79rbt"] Feb 02 13:40:53 crc kubenswrapper[4846]: I0202 13:40:53.506618 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-79rbt" podUID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerName="registry-server" containerID="cri-o://8d0a1ba5019c134f939fb29ee531db7c053c4da2018d94c2063c48762d6c64d7" gracePeriod=2 Feb 02 13:40:55 crc kubenswrapper[4846]: I0202 13:40:55.540479 4846 generic.go:334] "Generic (PLEG): container finished" podID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerID="8d0a1ba5019c134f939fb29ee531db7c053c4da2018d94c2063c48762d6c64d7" exitCode=0 Feb 02 13:40:55 crc kubenswrapper[4846]: I0202 13:40:55.540607 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79rbt" event={"ID":"786c7b50-d4ba-4abb-9073-f630d3d67a00","Type":"ContainerDied","Data":"8d0a1ba5019c134f939fb29ee531db7c053c4da2018d94c2063c48762d6c64d7"} Feb 02 13:40:55 crc kubenswrapper[4846]: I0202 13:40:55.982118 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.074985 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-utilities\") pod \"786c7b50-d4ba-4abb-9073-f630d3d67a00\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.075095 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-catalog-content\") pod \"786c7b50-d4ba-4abb-9073-f630d3d67a00\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.075136 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m7q5\" (UniqueName: \"kubernetes.io/projected/786c7b50-d4ba-4abb-9073-f630d3d67a00-kube-api-access-6m7q5\") pod \"786c7b50-d4ba-4abb-9073-f630d3d67a00\" (UID: \"786c7b50-d4ba-4abb-9073-f630d3d67a00\") " Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.077742 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-utilities" (OuterVolumeSpecName: "utilities") pod "786c7b50-d4ba-4abb-9073-f630d3d67a00" (UID: "786c7b50-d4ba-4abb-9073-f630d3d67a00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.081764 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/786c7b50-d4ba-4abb-9073-f630d3d67a00-kube-api-access-6m7q5" (OuterVolumeSpecName: "kube-api-access-6m7q5") pod "786c7b50-d4ba-4abb-9073-f630d3d67a00" (UID: "786c7b50-d4ba-4abb-9073-f630d3d67a00"). InnerVolumeSpecName "kube-api-access-6m7q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.178958 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.179023 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m7q5\" (UniqueName: \"kubernetes.io/projected/786c7b50-d4ba-4abb-9073-f630d3d67a00-kube-api-access-6m7q5\") on node \"crc\" DevicePath \"\"" Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.197104 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "786c7b50-d4ba-4abb-9073-f630d3d67a00" (UID: "786c7b50-d4ba-4abb-9073-f630d3d67a00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.280762 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/786c7b50-d4ba-4abb-9073-f630d3d67a00-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.554375 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79rbt" event={"ID":"786c7b50-d4ba-4abb-9073-f630d3d67a00","Type":"ContainerDied","Data":"d5e085b76632ebaf86bdd6285b283a06995313f0f54a6bea7e714a8646eb9355"} Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.554449 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-79rbt" Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.554486 4846 scope.go:117] "RemoveContainer" containerID="8d0a1ba5019c134f939fb29ee531db7c053c4da2018d94c2063c48762d6c64d7" Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.576834 4846 scope.go:117] "RemoveContainer" containerID="45e96db01cad0dd0296aec20adbcdb058e45e4c997208a8083e96f519b173adf" Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.593313 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-79rbt"] Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.600421 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-79rbt"] Feb 02 13:40:56 crc kubenswrapper[4846]: I0202 13:40:56.618253 4846 scope.go:117] "RemoveContainer" containerID="7dae938dcc962a9b1cd1d7222b563c8731d08e5ccf25936fe79ed54bd03e6008" Feb 02 13:40:57 crc kubenswrapper[4846]: I0202 13:40:57.442586 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="786c7b50-d4ba-4abb-9073-f630d3d67a00" path="/var/lib/kubelet/pods/786c7b50-d4ba-4abb-9073-f630d3d67a00/volumes" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.611665 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-sxfkf"] Feb 02 13:41:30 crc kubenswrapper[4846]: E0202 13:41:30.612516 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerName="registry-server" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.612530 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerName="registry-server" Feb 02 13:41:30 crc kubenswrapper[4846]: E0202 13:41:30.612544 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerName="extract-utilities" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.612552 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerName="extract-utilities" Feb 02 13:41:30 crc kubenswrapper[4846]: E0202 13:41:30.612583 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerName="extract-content" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.612588 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerName="extract-content" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.612764 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="786c7b50-d4ba-4abb-9073-f630d3d67a00" containerName="registry-server" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.613383 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sxfkf" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.631342 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-sxfkf"] Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.705008 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf7v2\" (UniqueName: \"kubernetes.io/projected/9de90124-4479-4ae4-a6c2-e12c41df67b4-kube-api-access-kf7v2\") pod \"barbican-db-create-sxfkf\" (UID: \"9de90124-4479-4ae4-a6c2-e12c41df67b4\") " pod="openstack/barbican-db-create-sxfkf" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.705190 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9de90124-4479-4ae4-a6c2-e12c41df67b4-operator-scripts\") pod \"barbican-db-create-sxfkf\" (UID: \"9de90124-4479-4ae4-a6c2-e12c41df67b4\") " pod="openstack/barbican-db-create-sxfkf" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.712448 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-da7f-account-create-update-tbvtt"] Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.713540 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-da7f-account-create-update-tbvtt" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.715463 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.726088 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-da7f-account-create-update-tbvtt"] Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.806566 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff66eedf-5da6-4676-9a52-fd6e5d1df411-operator-scripts\") pod \"barbican-da7f-account-create-update-tbvtt\" (UID: \"ff66eedf-5da6-4676-9a52-fd6e5d1df411\") " pod="openstack/barbican-da7f-account-create-update-tbvtt" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.806791 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9de90124-4479-4ae4-a6c2-e12c41df67b4-operator-scripts\") pod \"barbican-db-create-sxfkf\" (UID: \"9de90124-4479-4ae4-a6c2-e12c41df67b4\") " pod="openstack/barbican-db-create-sxfkf" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.806909 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf7v2\" (UniqueName: \"kubernetes.io/projected/9de90124-4479-4ae4-a6c2-e12c41df67b4-kube-api-access-kf7v2\") pod \"barbican-db-create-sxfkf\" (UID: \"9de90124-4479-4ae4-a6c2-e12c41df67b4\") " pod="openstack/barbican-db-create-sxfkf" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.807118 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wgcg\" (UniqueName: \"kubernetes.io/projected/ff66eedf-5da6-4676-9a52-fd6e5d1df411-kube-api-access-2wgcg\") pod \"barbican-da7f-account-create-update-tbvtt\" (UID: \"ff66eedf-5da6-4676-9a52-fd6e5d1df411\") " pod="openstack/barbican-da7f-account-create-update-tbvtt" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.807904 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9de90124-4479-4ae4-a6c2-e12c41df67b4-operator-scripts\") pod \"barbican-db-create-sxfkf\" (UID: \"9de90124-4479-4ae4-a6c2-e12c41df67b4\") " pod="openstack/barbican-db-create-sxfkf" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.829186 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf7v2\" (UniqueName: \"kubernetes.io/projected/9de90124-4479-4ae4-a6c2-e12c41df67b4-kube-api-access-kf7v2\") pod \"barbican-db-create-sxfkf\" (UID: \"9de90124-4479-4ae4-a6c2-e12c41df67b4\") " pod="openstack/barbican-db-create-sxfkf" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.908862 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff66eedf-5da6-4676-9a52-fd6e5d1df411-operator-scripts\") pod \"barbican-da7f-account-create-update-tbvtt\" (UID: \"ff66eedf-5da6-4676-9a52-fd6e5d1df411\") " pod="openstack/barbican-da7f-account-create-update-tbvtt" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.909293 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wgcg\" (UniqueName: \"kubernetes.io/projected/ff66eedf-5da6-4676-9a52-fd6e5d1df411-kube-api-access-2wgcg\") pod \"barbican-da7f-account-create-update-tbvtt\" (UID: \"ff66eedf-5da6-4676-9a52-fd6e5d1df411\") " pod="openstack/barbican-da7f-account-create-update-tbvtt" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.909702 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff66eedf-5da6-4676-9a52-fd6e5d1df411-operator-scripts\") pod \"barbican-da7f-account-create-update-tbvtt\" (UID: \"ff66eedf-5da6-4676-9a52-fd6e5d1df411\") " pod="openstack/barbican-da7f-account-create-update-tbvtt" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.929350 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wgcg\" (UniqueName: \"kubernetes.io/projected/ff66eedf-5da6-4676-9a52-fd6e5d1df411-kube-api-access-2wgcg\") pod \"barbican-da7f-account-create-update-tbvtt\" (UID: \"ff66eedf-5da6-4676-9a52-fd6e5d1df411\") " pod="openstack/barbican-da7f-account-create-update-tbvtt" Feb 02 13:41:30 crc kubenswrapper[4846]: I0202 13:41:30.942758 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sxfkf" Feb 02 13:41:31 crc kubenswrapper[4846]: I0202 13:41:31.033372 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-da7f-account-create-update-tbvtt" Feb 02 13:41:31 crc kubenswrapper[4846]: I0202 13:41:31.275253 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-sxfkf"] Feb 02 13:41:31 crc kubenswrapper[4846]: W0202 13:41:31.569342 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff66eedf_5da6_4676_9a52_fd6e5d1df411.slice/crio-a3cc759cc3a1a4bfe06b48df7241cbd1461f4cd5e5ad61ae97da39b148f80126 WatchSource:0}: Error finding container a3cc759cc3a1a4bfe06b48df7241cbd1461f4cd5e5ad61ae97da39b148f80126: Status 404 returned error can't find the container with id a3cc759cc3a1a4bfe06b48df7241cbd1461f4cd5e5ad61ae97da39b148f80126 Feb 02 13:41:31 crc kubenswrapper[4846]: I0202 13:41:31.569358 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-da7f-account-create-update-tbvtt"] Feb 02 13:41:31 crc kubenswrapper[4846]: I0202 13:41:31.875830 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-da7f-account-create-update-tbvtt" event={"ID":"ff66eedf-5da6-4676-9a52-fd6e5d1df411","Type":"ContainerStarted","Data":"6fff3bdb8343bc093cc9c6be08d6845a546b8a2009d749868f3475a9f27c7ee3"} Feb 02 13:41:31 crc kubenswrapper[4846]: I0202 13:41:31.875874 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-da7f-account-create-update-tbvtt" event={"ID":"ff66eedf-5da6-4676-9a52-fd6e5d1df411","Type":"ContainerStarted","Data":"a3cc759cc3a1a4bfe06b48df7241cbd1461f4cd5e5ad61ae97da39b148f80126"} Feb 02 13:41:31 crc kubenswrapper[4846]: I0202 13:41:31.877584 4846 generic.go:334] "Generic (PLEG): container finished" podID="9de90124-4479-4ae4-a6c2-e12c41df67b4" containerID="4175c9c531fb6923c23164ac327144d04119c8099e366ea3af132f5ea1af4d14" exitCode=0 Feb 02 13:41:31 crc kubenswrapper[4846]: I0202 13:41:31.877638 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-sxfkf" event={"ID":"9de90124-4479-4ae4-a6c2-e12c41df67b4","Type":"ContainerDied","Data":"4175c9c531fb6923c23164ac327144d04119c8099e366ea3af132f5ea1af4d14"} Feb 02 13:41:31 crc kubenswrapper[4846]: I0202 13:41:31.877660 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-sxfkf" event={"ID":"9de90124-4479-4ae4-a6c2-e12c41df67b4","Type":"ContainerStarted","Data":"91f78cdd3faf79b6f9f7da2edcd793da186c8a079d14a52492caadbf46e3ad4f"} Feb 02 13:41:31 crc kubenswrapper[4846]: I0202 13:41:31.900036 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-da7f-account-create-update-tbvtt" podStartSLOduration=1.9000113459999999 podStartE2EDuration="1.900011346s" podCreationTimestamp="2026-02-02 13:41:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:41:31.891288045 +0000 UTC m=+5523.119874948" watchObservedRunningTime="2026-02-02 13:41:31.900011346 +0000 UTC m=+5523.128598209" Feb 02 13:41:32 crc kubenswrapper[4846]: I0202 13:41:32.887105 4846 generic.go:334] "Generic (PLEG): container finished" podID="ff66eedf-5da6-4676-9a52-fd6e5d1df411" containerID="6fff3bdb8343bc093cc9c6be08d6845a546b8a2009d749868f3475a9f27c7ee3" exitCode=0 Feb 02 13:41:32 crc kubenswrapper[4846]: I0202 13:41:32.888199 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-da7f-account-create-update-tbvtt" event={"ID":"ff66eedf-5da6-4676-9a52-fd6e5d1df411","Type":"ContainerDied","Data":"6fff3bdb8343bc093cc9c6be08d6845a546b8a2009d749868f3475a9f27c7ee3"} Feb 02 13:41:33 crc kubenswrapper[4846]: I0202 13:41:33.218760 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sxfkf" Feb 02 13:41:33 crc kubenswrapper[4846]: I0202 13:41:33.358930 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9de90124-4479-4ae4-a6c2-e12c41df67b4-operator-scripts\") pod \"9de90124-4479-4ae4-a6c2-e12c41df67b4\" (UID: \"9de90124-4479-4ae4-a6c2-e12c41df67b4\") " Feb 02 13:41:33 crc kubenswrapper[4846]: I0202 13:41:33.359136 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf7v2\" (UniqueName: \"kubernetes.io/projected/9de90124-4479-4ae4-a6c2-e12c41df67b4-kube-api-access-kf7v2\") pod \"9de90124-4479-4ae4-a6c2-e12c41df67b4\" (UID: \"9de90124-4479-4ae4-a6c2-e12c41df67b4\") " Feb 02 13:41:33 crc kubenswrapper[4846]: I0202 13:41:33.359601 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9de90124-4479-4ae4-a6c2-e12c41df67b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9de90124-4479-4ae4-a6c2-e12c41df67b4" (UID: "9de90124-4479-4ae4-a6c2-e12c41df67b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:41:33 crc kubenswrapper[4846]: I0202 13:41:33.364913 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9de90124-4479-4ae4-a6c2-e12c41df67b4-kube-api-access-kf7v2" (OuterVolumeSpecName: "kube-api-access-kf7v2") pod "9de90124-4479-4ae4-a6c2-e12c41df67b4" (UID: "9de90124-4479-4ae4-a6c2-e12c41df67b4"). InnerVolumeSpecName "kube-api-access-kf7v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:41:33 crc kubenswrapper[4846]: I0202 13:41:33.460538 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9de90124-4479-4ae4-a6c2-e12c41df67b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:33 crc kubenswrapper[4846]: I0202 13:41:33.460571 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf7v2\" (UniqueName: \"kubernetes.io/projected/9de90124-4479-4ae4-a6c2-e12c41df67b4-kube-api-access-kf7v2\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:33 crc kubenswrapper[4846]: I0202 13:41:33.899165 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sxfkf" Feb 02 13:41:33 crc kubenswrapper[4846]: I0202 13:41:33.899904 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-sxfkf" event={"ID":"9de90124-4479-4ae4-a6c2-e12c41df67b4","Type":"ContainerDied","Data":"91f78cdd3faf79b6f9f7da2edcd793da186c8a079d14a52492caadbf46e3ad4f"} Feb 02 13:41:33 crc kubenswrapper[4846]: I0202 13:41:33.899931 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91f78cdd3faf79b6f9f7da2edcd793da186c8a079d14a52492caadbf46e3ad4f" Feb 02 13:41:34 crc kubenswrapper[4846]: I0202 13:41:34.228363 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-da7f-account-create-update-tbvtt" Feb 02 13:41:34 crc kubenswrapper[4846]: I0202 13:41:34.272352 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wgcg\" (UniqueName: \"kubernetes.io/projected/ff66eedf-5da6-4676-9a52-fd6e5d1df411-kube-api-access-2wgcg\") pod \"ff66eedf-5da6-4676-9a52-fd6e5d1df411\" (UID: \"ff66eedf-5da6-4676-9a52-fd6e5d1df411\") " Feb 02 13:41:34 crc kubenswrapper[4846]: I0202 13:41:34.272606 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff66eedf-5da6-4676-9a52-fd6e5d1df411-operator-scripts\") pod \"ff66eedf-5da6-4676-9a52-fd6e5d1df411\" (UID: \"ff66eedf-5da6-4676-9a52-fd6e5d1df411\") " Feb 02 13:41:34 crc kubenswrapper[4846]: I0202 13:41:34.274293 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff66eedf-5da6-4676-9a52-fd6e5d1df411-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff66eedf-5da6-4676-9a52-fd6e5d1df411" (UID: "ff66eedf-5da6-4676-9a52-fd6e5d1df411"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:41:34 crc kubenswrapper[4846]: I0202 13:41:34.277865 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff66eedf-5da6-4676-9a52-fd6e5d1df411-kube-api-access-2wgcg" (OuterVolumeSpecName: "kube-api-access-2wgcg") pod "ff66eedf-5da6-4676-9a52-fd6e5d1df411" (UID: "ff66eedf-5da6-4676-9a52-fd6e5d1df411"). InnerVolumeSpecName "kube-api-access-2wgcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:41:34 crc kubenswrapper[4846]: I0202 13:41:34.374452 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff66eedf-5da6-4676-9a52-fd6e5d1df411-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:34 crc kubenswrapper[4846]: I0202 13:41:34.374483 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wgcg\" (UniqueName: \"kubernetes.io/projected/ff66eedf-5da6-4676-9a52-fd6e5d1df411-kube-api-access-2wgcg\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:34 crc kubenswrapper[4846]: I0202 13:41:34.908312 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-da7f-account-create-update-tbvtt" event={"ID":"ff66eedf-5da6-4676-9a52-fd6e5d1df411","Type":"ContainerDied","Data":"a3cc759cc3a1a4bfe06b48df7241cbd1461f4cd5e5ad61ae97da39b148f80126"} Feb 02 13:41:34 crc kubenswrapper[4846]: I0202 13:41:34.908365 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3cc759cc3a1a4bfe06b48df7241cbd1461f4cd5e5ad61ae97da39b148f80126" Feb 02 13:41:34 crc kubenswrapper[4846]: I0202 13:41:34.908439 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-da7f-account-create-update-tbvtt" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.096665 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9r479"] Feb 02 13:41:36 crc kubenswrapper[4846]: E0202 13:41:36.104277 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9de90124-4479-4ae4-a6c2-e12c41df67b4" containerName="mariadb-database-create" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.104338 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9de90124-4479-4ae4-a6c2-e12c41df67b4" containerName="mariadb-database-create" Feb 02 13:41:36 crc kubenswrapper[4846]: E0202 13:41:36.104420 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff66eedf-5da6-4676-9a52-fd6e5d1df411" containerName="mariadb-account-create-update" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.104429 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff66eedf-5da6-4676-9a52-fd6e5d1df411" containerName="mariadb-account-create-update" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.104790 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9de90124-4479-4ae4-a6c2-e12c41df67b4" containerName="mariadb-database-create" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.104821 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff66eedf-5da6-4676-9a52-fd6e5d1df411" containerName="mariadb-account-create-update" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.105576 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.108231 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-7k9fd" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.108475 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.115923 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9r479"] Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.201995 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-db-sync-config-data\") pod \"barbican-db-sync-9r479\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.202268 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-combined-ca-bundle\") pod \"barbican-db-sync-9r479\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.202389 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvkw6\" (UniqueName: \"kubernetes.io/projected/957b0ccd-612b-4e78-909e-048644d9fd17-kube-api-access-gvkw6\") pod \"barbican-db-sync-9r479\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.303556 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvkw6\" (UniqueName: \"kubernetes.io/projected/957b0ccd-612b-4e78-909e-048644d9fd17-kube-api-access-gvkw6\") pod \"barbican-db-sync-9r479\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.304068 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-db-sync-config-data\") pod \"barbican-db-sync-9r479\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.304202 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-combined-ca-bundle\") pod \"barbican-db-sync-9r479\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.320566 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-db-sync-config-data\") pod \"barbican-db-sync-9r479\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.320760 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-combined-ca-bundle\") pod \"barbican-db-sync-9r479\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.324568 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvkw6\" (UniqueName: \"kubernetes.io/projected/957b0ccd-612b-4e78-909e-048644d9fd17-kube-api-access-gvkw6\") pod \"barbican-db-sync-9r479\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.435148 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:36 crc kubenswrapper[4846]: I0202 13:41:36.928993 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9r479"] Feb 02 13:41:37 crc kubenswrapper[4846]: I0202 13:41:37.943225 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9r479" event={"ID":"957b0ccd-612b-4e78-909e-048644d9fd17","Type":"ContainerStarted","Data":"f338a53a8f99bcc13f977c2a11df73ae5ad23888cbca856da2180ea35bcd2fe7"} Feb 02 13:41:37 crc kubenswrapper[4846]: I0202 13:41:37.943824 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9r479" event={"ID":"957b0ccd-612b-4e78-909e-048644d9fd17","Type":"ContainerStarted","Data":"69b81100484ca4f46bccdb5abbb4e88d241b549c733b90180cd02442400adc17"} Feb 02 13:41:37 crc kubenswrapper[4846]: I0202 13:41:37.963116 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9r479" podStartSLOduration=1.963095943 podStartE2EDuration="1.963095943s" podCreationTimestamp="2026-02-02 13:41:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:41:37.960677584 +0000 UTC m=+5529.189264447" watchObservedRunningTime="2026-02-02 13:41:37.963095943 +0000 UTC m=+5529.191682806" Feb 02 13:41:40 crc kubenswrapper[4846]: I0202 13:41:40.963989 4846 generic.go:334] "Generic (PLEG): container finished" podID="957b0ccd-612b-4e78-909e-048644d9fd17" containerID="f338a53a8f99bcc13f977c2a11df73ae5ad23888cbca856da2180ea35bcd2fe7" exitCode=0 Feb 02 13:41:40 crc kubenswrapper[4846]: I0202 13:41:40.964081 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9r479" event={"ID":"957b0ccd-612b-4e78-909e-048644d9fd17","Type":"ContainerDied","Data":"f338a53a8f99bcc13f977c2a11df73ae5ad23888cbca856da2180ea35bcd2fe7"} Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.286911 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.412812 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-combined-ca-bundle\") pod \"957b0ccd-612b-4e78-909e-048644d9fd17\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.413154 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-db-sync-config-data\") pod \"957b0ccd-612b-4e78-909e-048644d9fd17\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.413198 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvkw6\" (UniqueName: \"kubernetes.io/projected/957b0ccd-612b-4e78-909e-048644d9fd17-kube-api-access-gvkw6\") pod \"957b0ccd-612b-4e78-909e-048644d9fd17\" (UID: \"957b0ccd-612b-4e78-909e-048644d9fd17\") " Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.419618 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "957b0ccd-612b-4e78-909e-048644d9fd17" (UID: "957b0ccd-612b-4e78-909e-048644d9fd17"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.419701 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/957b0ccd-612b-4e78-909e-048644d9fd17-kube-api-access-gvkw6" (OuterVolumeSpecName: "kube-api-access-gvkw6") pod "957b0ccd-612b-4e78-909e-048644d9fd17" (UID: "957b0ccd-612b-4e78-909e-048644d9fd17"). InnerVolumeSpecName "kube-api-access-gvkw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.437017 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "957b0ccd-612b-4e78-909e-048644d9fd17" (UID: "957b0ccd-612b-4e78-909e-048644d9fd17"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.515359 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.515401 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvkw6\" (UniqueName: \"kubernetes.io/projected/957b0ccd-612b-4e78-909e-048644d9fd17-kube-api-access-gvkw6\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.515411 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/957b0ccd-612b-4e78-909e-048644d9fd17-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.981921 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9r479" event={"ID":"957b0ccd-612b-4e78-909e-048644d9fd17","Type":"ContainerDied","Data":"69b81100484ca4f46bccdb5abbb4e88d241b549c733b90180cd02442400adc17"} Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.981968 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9r479" Feb 02 13:41:42 crc kubenswrapper[4846]: I0202 13:41:42.981971 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69b81100484ca4f46bccdb5abbb4e88d241b549c733b90180cd02442400adc17" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.222926 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7874cbf7b5-2kgbm"] Feb 02 13:41:43 crc kubenswrapper[4846]: E0202 13:41:43.223280 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="957b0ccd-612b-4e78-909e-048644d9fd17" containerName="barbican-db-sync" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.223300 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="957b0ccd-612b-4e78-909e-048644d9fd17" containerName="barbican-db-sync" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.223491 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="957b0ccd-612b-4e78-909e-048644d9fd17" containerName="barbican-db-sync" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.230697 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.239222 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-7k9fd" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.239469 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.239745 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.256410 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7874cbf7b5-2kgbm"] Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.276254 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7687b59478-hmtmb"] Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.277678 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.280002 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.305197 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7687b59478-hmtmb"] Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.321129 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66845c4585-49jp4"] Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.322591 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.330061 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42acd0dc-ac0d-4bc9-a814-12ced76e190a-combined-ca-bundle\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.330412 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42acd0dc-ac0d-4bc9-a814-12ced76e190a-logs\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.330534 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-config-data\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.330709 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42acd0dc-ac0d-4bc9-a814-12ced76e190a-config-data\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.330839 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg67r\" (UniqueName: \"kubernetes.io/projected/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-kube-api-access-dg67r\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.330927 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-combined-ca-bundle\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.331034 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-logs\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.331139 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-config-data-custom\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.331208 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhrlz\" (UniqueName: \"kubernetes.io/projected/42acd0dc-ac0d-4bc9-a814-12ced76e190a-kube-api-access-rhrlz\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.331306 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42acd0dc-ac0d-4bc9-a814-12ced76e190a-config-data-custom\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.343148 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66845c4585-49jp4"] Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.414244 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-54f8585bfb-zsz4k"] Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.420582 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.430714 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42acd0dc-ac0d-4bc9-a814-12ced76e190a-logs\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-config-data\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432540 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42acd0dc-ac0d-4bc9-a814-12ced76e190a-config-data\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432566 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg67r\" (UniqueName: \"kubernetes.io/projected/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-kube-api-access-dg67r\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432590 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-combined-ca-bundle\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432635 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5w45\" (UniqueName: \"kubernetes.io/projected/76bf5119-ebc0-4687-9b32-9c5bf37b18de-kube-api-access-b5w45\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432659 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-config\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432680 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-sb\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432700 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-logs\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432734 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-nb\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-config-data-custom\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432768 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-dns-svc\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432786 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhrlz\" (UniqueName: \"kubernetes.io/projected/42acd0dc-ac0d-4bc9-a814-12ced76e190a-kube-api-access-rhrlz\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432813 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42acd0dc-ac0d-4bc9-a814-12ced76e190a-config-data-custom\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.432864 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42acd0dc-ac0d-4bc9-a814-12ced76e190a-combined-ca-bundle\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.433222 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42acd0dc-ac0d-4bc9-a814-12ced76e190a-logs\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.436164 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-logs\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.449306 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-config-data-custom\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.450943 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-config-data\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.457320 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhrlz\" (UniqueName: \"kubernetes.io/projected/42acd0dc-ac0d-4bc9-a814-12ced76e190a-kube-api-access-rhrlz\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.459221 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54f8585bfb-zsz4k"] Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.461221 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg67r\" (UniqueName: \"kubernetes.io/projected/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-kube-api-access-dg67r\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.461928 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42acd0dc-ac0d-4bc9-a814-12ced76e190a-config-data\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.463589 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42acd0dc-ac0d-4bc9-a814-12ced76e190a-config-data-custom\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.471608 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb-combined-ca-bundle\") pod \"barbican-keystone-listener-7687b59478-hmtmb\" (UID: \"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb\") " pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.481444 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42acd0dc-ac0d-4bc9-a814-12ced76e190a-combined-ca-bundle\") pod \"barbican-worker-7874cbf7b5-2kgbm\" (UID: \"42acd0dc-ac0d-4bc9-a814-12ced76e190a\") " pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.534892 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2bhm\" (UniqueName: \"kubernetes.io/projected/64316e8a-a9d2-42a4-a572-359b33ca510c-kube-api-access-b2bhm\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.534941 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-nb\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.534970 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data-custom\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.535003 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-dns-svc\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.535240 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-combined-ca-bundle\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.535497 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.535600 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5w45\" (UniqueName: \"kubernetes.io/projected/76bf5119-ebc0-4687-9b32-9c5bf37b18de-kube-api-access-b5w45\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.535641 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-config\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.535669 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64316e8a-a9d2-42a4-a572-359b33ca510c-logs\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.535694 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-sb\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.535962 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-dns-svc\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.536087 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-nb\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.536513 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-sb\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.536869 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-config\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.553940 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5w45\" (UniqueName: \"kubernetes.io/projected/76bf5119-ebc0-4687-9b32-9c5bf37b18de-kube-api-access-b5w45\") pod \"dnsmasq-dns-66845c4585-49jp4\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.575296 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7874cbf7b5-2kgbm" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.597177 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.637802 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.637884 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64316e8a-a9d2-42a4-a572-359b33ca510c-logs\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.637932 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2bhm\" (UniqueName: \"kubernetes.io/projected/64316e8a-a9d2-42a4-a572-359b33ca510c-kube-api-access-b2bhm\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.637968 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data-custom\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.638029 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-combined-ca-bundle\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.639001 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64316e8a-a9d2-42a4-a572-359b33ca510c-logs\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.643210 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-combined-ca-bundle\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.643768 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.649202 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data-custom\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.658252 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.659261 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2bhm\" (UniqueName: \"kubernetes.io/projected/64316e8a-a9d2-42a4-a572-359b33ca510c-kube-api-access-b2bhm\") pod \"barbican-api-54f8585bfb-zsz4k\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:43 crc kubenswrapper[4846]: I0202 13:41:43.840841 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:44 crc kubenswrapper[4846]: I0202 13:41:44.072847 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7874cbf7b5-2kgbm"] Feb 02 13:41:44 crc kubenswrapper[4846]: I0202 13:41:44.170501 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66845c4585-49jp4"] Feb 02 13:41:44 crc kubenswrapper[4846]: I0202 13:41:44.181827 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7687b59478-hmtmb"] Feb 02 13:41:44 crc kubenswrapper[4846]: I0202 13:41:44.333999 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54f8585bfb-zsz4k"] Feb 02 13:41:44 crc kubenswrapper[4846]: I0202 13:41:44.998717 4846 generic.go:334] "Generic (PLEG): container finished" podID="76bf5119-ebc0-4687-9b32-9c5bf37b18de" containerID="a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1" exitCode=0 Feb 02 13:41:44 crc kubenswrapper[4846]: I0202 13:41:44.998766 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66845c4585-49jp4" event={"ID":"76bf5119-ebc0-4687-9b32-9c5bf37b18de","Type":"ContainerDied","Data":"a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1"} Feb 02 13:41:44 crc kubenswrapper[4846]: I0202 13:41:44.999068 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66845c4585-49jp4" event={"ID":"76bf5119-ebc0-4687-9b32-9c5bf37b18de","Type":"ContainerStarted","Data":"a0bb4be74e504ada472dcb1ecaa4de1ae0a66674b606ee08a7f41155fd9a2884"} Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.001677 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54f8585bfb-zsz4k" event={"ID":"64316e8a-a9d2-42a4-a572-359b33ca510c","Type":"ContainerStarted","Data":"a47babc40ed3ee554051885b0079ab05b8c4af6cd784a66c8b0f4211a8ee9aa5"} Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.001753 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54f8585bfb-zsz4k" event={"ID":"64316e8a-a9d2-42a4-a572-359b33ca510c","Type":"ContainerStarted","Data":"a5dc7bfb0f923103f317e2f5f0ef410332dc55538a35f56ab9f3e9af692dc182"} Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.001777 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.001793 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.001805 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54f8585bfb-zsz4k" event={"ID":"64316e8a-a9d2-42a4-a572-359b33ca510c","Type":"ContainerStarted","Data":"578da1c18fdaf053ae952e788c6f8704a3c3586c0ae74acce14b5500402d9c8b"} Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.007975 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7874cbf7b5-2kgbm" event={"ID":"42acd0dc-ac0d-4bc9-a814-12ced76e190a","Type":"ContainerStarted","Data":"72280d8b7f3c33825fc59cb68e7c59b347a6d407e37574d9df20c04223a33ce1"} Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.008012 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7874cbf7b5-2kgbm" event={"ID":"42acd0dc-ac0d-4bc9-a814-12ced76e190a","Type":"ContainerStarted","Data":"1c749a94f58e10d4ba2766a25bb746855198867ced410b9e206edad78af8f79a"} Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.008023 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7874cbf7b5-2kgbm" event={"ID":"42acd0dc-ac0d-4bc9-a814-12ced76e190a","Type":"ContainerStarted","Data":"d62c3e827f3e39c2d4b9ef9c0d06fa132114be0b774b0a0412adefbadc725b5f"} Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.011129 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" event={"ID":"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb","Type":"ContainerStarted","Data":"ee1440c73bd768996df657277de27a092d4222a5ba3fb5483aa93ba7c32a364f"} Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.011174 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" event={"ID":"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb","Type":"ContainerStarted","Data":"1826bc588c201c7197051f2af50317812c98e8f24049ba398ba78c5745e52253"} Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.011183 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" event={"ID":"3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb","Type":"ContainerStarted","Data":"2140ecb3392c9dac8f00a606c3b0084c3da0603b140c30fc8972d9af18b85748"} Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.046004 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7687b59478-hmtmb" podStartSLOduration=2.045984984 podStartE2EDuration="2.045984984s" podCreationTimestamp="2026-02-02 13:41:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:41:45.036573625 +0000 UTC m=+5536.265160488" watchObservedRunningTime="2026-02-02 13:41:45.045984984 +0000 UTC m=+5536.274571847" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.065337 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7874cbf7b5-2kgbm" podStartSLOduration=2.065319874 podStartE2EDuration="2.065319874s" podCreationTimestamp="2026-02-02 13:41:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:41:45.060778123 +0000 UTC m=+5536.289364986" watchObservedRunningTime="2026-02-02 13:41:45.065319874 +0000 UTC m=+5536.293906737" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.101453 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-54f8585bfb-zsz4k" podStartSLOduration=2.101429241 podStartE2EDuration="2.101429241s" podCreationTimestamp="2026-02-02 13:41:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:41:45.078956465 +0000 UTC m=+5536.307543318" watchObservedRunningTime="2026-02-02 13:41:45.101429241 +0000 UTC m=+5536.330016104" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.583022 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-857b8ccfcd-55wgg"] Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.585051 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.588246 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.588433 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.605154 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-857b8ccfcd-55wgg"] Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.680876 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-config-data-custom\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.680920 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-config-data\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.680945 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82sdr\" (UniqueName: \"kubernetes.io/projected/a33b00bf-aa6a-4eb8-be29-be3e943e276f-kube-api-access-82sdr\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.680982 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-combined-ca-bundle\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.681008 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-public-tls-certs\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.681032 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a33b00bf-aa6a-4eb8-be29-be3e943e276f-logs\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.681053 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-internal-tls-certs\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.782315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82sdr\" (UniqueName: \"kubernetes.io/projected/a33b00bf-aa6a-4eb8-be29-be3e943e276f-kube-api-access-82sdr\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.782387 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-combined-ca-bundle\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.782426 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-public-tls-certs\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.782449 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a33b00bf-aa6a-4eb8-be29-be3e943e276f-logs\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.782474 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-internal-tls-certs\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.782570 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-config-data-custom\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.782593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-config-data\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.783306 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a33b00bf-aa6a-4eb8-be29-be3e943e276f-logs\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.788578 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-internal-tls-certs\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.788636 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-public-tls-certs\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.789453 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-config-data\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.789956 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-combined-ca-bundle\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.791300 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a33b00bf-aa6a-4eb8-be29-be3e943e276f-config-data-custom\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.805690 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82sdr\" (UniqueName: \"kubernetes.io/projected/a33b00bf-aa6a-4eb8-be29-be3e943e276f-kube-api-access-82sdr\") pod \"barbican-api-857b8ccfcd-55wgg\" (UID: \"a33b00bf-aa6a-4eb8-be29-be3e943e276f\") " pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:45 crc kubenswrapper[4846]: I0202 13:41:45.912089 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:46 crc kubenswrapper[4846]: I0202 13:41:46.036031 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66845c4585-49jp4" event={"ID":"76bf5119-ebc0-4687-9b32-9c5bf37b18de","Type":"ContainerStarted","Data":"1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c"} Feb 02 13:41:46 crc kubenswrapper[4846]: I0202 13:41:46.038264 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:46 crc kubenswrapper[4846]: I0202 13:41:46.061480 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66845c4585-49jp4" podStartSLOduration=3.061461605 podStartE2EDuration="3.061461605s" podCreationTimestamp="2026-02-02 13:41:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:41:46.05503883 +0000 UTC m=+5537.283625693" watchObservedRunningTime="2026-02-02 13:41:46.061461605 +0000 UTC m=+5537.290048468" Feb 02 13:41:46 crc kubenswrapper[4846]: I0202 13:41:46.416890 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-857b8ccfcd-55wgg"] Feb 02 13:41:47 crc kubenswrapper[4846]: I0202 13:41:47.044815 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-857b8ccfcd-55wgg" event={"ID":"a33b00bf-aa6a-4eb8-be29-be3e943e276f","Type":"ContainerStarted","Data":"a7691da9f973ba38b600c21579f4d91686c16795b6dd19c68bb7ddca4d5dcf21"} Feb 02 13:41:47 crc kubenswrapper[4846]: I0202 13:41:47.045388 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:47 crc kubenswrapper[4846]: I0202 13:41:47.045404 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-857b8ccfcd-55wgg" event={"ID":"a33b00bf-aa6a-4eb8-be29-be3e943e276f","Type":"ContainerStarted","Data":"ac5e3e538f8092e38e3b82bb119c2c608150026cfb423d0c26b8baf615180ab4"} Feb 02 13:41:47 crc kubenswrapper[4846]: I0202 13:41:47.045415 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-857b8ccfcd-55wgg" event={"ID":"a33b00bf-aa6a-4eb8-be29-be3e943e276f","Type":"ContainerStarted","Data":"a3d61190a9c45b2f0614388b824106f9f3d6b7102cd029caf8d80f6620594d0e"} Feb 02 13:41:47 crc kubenswrapper[4846]: I0202 13:41:47.070003 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-857b8ccfcd-55wgg" podStartSLOduration=2.069982269 podStartE2EDuration="2.069982269s" podCreationTimestamp="2026-02-02 13:41:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:41:47.063818038 +0000 UTC m=+5538.292404901" watchObservedRunningTime="2026-02-02 13:41:47.069982269 +0000 UTC m=+5538.298569132" Feb 02 13:41:48 crc kubenswrapper[4846]: I0202 13:41:48.058059 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:53 crc kubenswrapper[4846]: I0202 13:41:53.660823 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:41:53 crc kubenswrapper[4846]: I0202 13:41:53.731824 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5757586b9-6tq5q"] Feb 02 13:41:53 crc kubenswrapper[4846]: I0202 13:41:53.732119 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" podUID="b036df9b-ce4c-49f2-8cbc-47bacccabd05" containerName="dnsmasq-dns" containerID="cri-o://5201dae633bb7e7f75cbf443d2c265c3a01e7387321f251ae2d1408f0d13bd4e" gracePeriod=10 Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.118764 4846 generic.go:334] "Generic (PLEG): container finished" podID="b036df9b-ce4c-49f2-8cbc-47bacccabd05" containerID="5201dae633bb7e7f75cbf443d2c265c3a01e7387321f251ae2d1408f0d13bd4e" exitCode=0 Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.118928 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" event={"ID":"b036df9b-ce4c-49f2-8cbc-47bacccabd05","Type":"ContainerDied","Data":"5201dae633bb7e7f75cbf443d2c265c3a01e7387321f251ae2d1408f0d13bd4e"} Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.251556 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.338418 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-config\") pod \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.338652 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62rwj\" (UniqueName: \"kubernetes.io/projected/b036df9b-ce4c-49f2-8cbc-47bacccabd05-kube-api-access-62rwj\") pod \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.338690 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-sb\") pod \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.338745 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-nb\") pod \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.338796 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-dns-svc\") pod \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\" (UID: \"b036df9b-ce4c-49f2-8cbc-47bacccabd05\") " Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.358645 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b036df9b-ce4c-49f2-8cbc-47bacccabd05-kube-api-access-62rwj" (OuterVolumeSpecName: "kube-api-access-62rwj") pod "b036df9b-ce4c-49f2-8cbc-47bacccabd05" (UID: "b036df9b-ce4c-49f2-8cbc-47bacccabd05"). InnerVolumeSpecName "kube-api-access-62rwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.386048 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-config" (OuterVolumeSpecName: "config") pod "b036df9b-ce4c-49f2-8cbc-47bacccabd05" (UID: "b036df9b-ce4c-49f2-8cbc-47bacccabd05"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.386067 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b036df9b-ce4c-49f2-8cbc-47bacccabd05" (UID: "b036df9b-ce4c-49f2-8cbc-47bacccabd05"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.393523 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b036df9b-ce4c-49f2-8cbc-47bacccabd05" (UID: "b036df9b-ce4c-49f2-8cbc-47bacccabd05"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.411389 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b036df9b-ce4c-49f2-8cbc-47bacccabd05" (UID: "b036df9b-ce4c-49f2-8cbc-47bacccabd05"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.441340 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.441737 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62rwj\" (UniqueName: \"kubernetes.io/projected/b036df9b-ce4c-49f2-8cbc-47bacccabd05-kube-api-access-62rwj\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.441750 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.441761 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:54 crc kubenswrapper[4846]: I0202 13:41:54.441770 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b036df9b-ce4c-49f2-8cbc-47bacccabd05-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:41:55 crc kubenswrapper[4846]: I0202 13:41:55.130140 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" event={"ID":"b036df9b-ce4c-49f2-8cbc-47bacccabd05","Type":"ContainerDied","Data":"5985716d5b3b663a6fbe641f671194347395dadd23dd5b13a13311fa5c45dd21"} Feb 02 13:41:55 crc kubenswrapper[4846]: I0202 13:41:55.130203 4846 scope.go:117] "RemoveContainer" containerID="5201dae633bb7e7f75cbf443d2c265c3a01e7387321f251ae2d1408f0d13bd4e" Feb 02 13:41:55 crc kubenswrapper[4846]: I0202 13:41:55.130202 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5757586b9-6tq5q" Feb 02 13:41:55 crc kubenswrapper[4846]: I0202 13:41:55.149139 4846 scope.go:117] "RemoveContainer" containerID="423c31df98a8da5193b64f08771d8d787dcb037074778f4c06f71d3f3a2d8086" Feb 02 13:41:55 crc kubenswrapper[4846]: I0202 13:41:55.178330 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5757586b9-6tq5q"] Feb 02 13:41:55 crc kubenswrapper[4846]: I0202 13:41:55.209027 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5757586b9-6tq5q"] Feb 02 13:41:55 crc kubenswrapper[4846]: I0202 13:41:55.306980 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:55 crc kubenswrapper[4846]: I0202 13:41:55.434105 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b036df9b-ce4c-49f2-8cbc-47bacccabd05" path="/var/lib/kubelet/pods/b036df9b-ce4c-49f2-8cbc-47bacccabd05/volumes" Feb 02 13:41:55 crc kubenswrapper[4846]: I0202 13:41:55.453519 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:41:57 crc kubenswrapper[4846]: I0202 13:41:57.377909 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:57 crc kubenswrapper[4846]: I0202 13:41:57.601983 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-857b8ccfcd-55wgg" Feb 02 13:41:57 crc kubenswrapper[4846]: I0202 13:41:57.667737 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-54f8585bfb-zsz4k"] Feb 02 13:41:57 crc kubenswrapper[4846]: I0202 13:41:57.667977 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-54f8585bfb-zsz4k" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerName="barbican-api-log" containerID="cri-o://a5dc7bfb0f923103f317e2f5f0ef410332dc55538a35f56ab9f3e9af692dc182" gracePeriod=30 Feb 02 13:41:57 crc kubenswrapper[4846]: I0202 13:41:57.668136 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-54f8585bfb-zsz4k" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerName="barbican-api" containerID="cri-o://a47babc40ed3ee554051885b0079ab05b8c4af6cd784a66c8b0f4211a8ee9aa5" gracePeriod=30 Feb 02 13:41:58 crc kubenswrapper[4846]: I0202 13:41:58.173701 4846 generic.go:334] "Generic (PLEG): container finished" podID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerID="a5dc7bfb0f923103f317e2f5f0ef410332dc55538a35f56ab9f3e9af692dc182" exitCode=143 Feb 02 13:41:58 crc kubenswrapper[4846]: I0202 13:41:58.173784 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54f8585bfb-zsz4k" event={"ID":"64316e8a-a9d2-42a4-a572-359b33ca510c","Type":"ContainerDied","Data":"a5dc7bfb0f923103f317e2f5f0ef410332dc55538a35f56ab9f3e9af692dc182"} Feb 02 13:42:00 crc kubenswrapper[4846]: I0202 13:42:00.823139 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-54f8585bfb-zsz4k" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.36:9311/healthcheck\": read tcp 10.217.0.2:60442->10.217.1.36:9311: read: connection reset by peer" Feb 02 13:42:00 crc kubenswrapper[4846]: I0202 13:42:00.823169 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-54f8585bfb-zsz4k" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.36:9311/healthcheck\": read tcp 10.217.0.2:60444->10.217.1.36:9311: read: connection reset by peer" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.206281 4846 generic.go:334] "Generic (PLEG): container finished" podID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerID="a47babc40ed3ee554051885b0079ab05b8c4af6cd784a66c8b0f4211a8ee9aa5" exitCode=0 Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.206326 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54f8585bfb-zsz4k" event={"ID":"64316e8a-a9d2-42a4-a572-359b33ca510c","Type":"ContainerDied","Data":"a47babc40ed3ee554051885b0079ab05b8c4af6cd784a66c8b0f4211a8ee9aa5"} Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.206397 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54f8585bfb-zsz4k" event={"ID":"64316e8a-a9d2-42a4-a572-359b33ca510c","Type":"ContainerDied","Data":"578da1c18fdaf053ae952e788c6f8704a3c3586c0ae74acce14b5500402d9c8b"} Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.206418 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="578da1c18fdaf053ae952e788c6f8704a3c3586c0ae74acce14b5500402d9c8b" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.265235 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.295435 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data-custom\") pod \"64316e8a-a9d2-42a4-a572-359b33ca510c\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.295482 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data\") pod \"64316e8a-a9d2-42a4-a572-359b33ca510c\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.295528 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-combined-ca-bundle\") pod \"64316e8a-a9d2-42a4-a572-359b33ca510c\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.295567 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2bhm\" (UniqueName: \"kubernetes.io/projected/64316e8a-a9d2-42a4-a572-359b33ca510c-kube-api-access-b2bhm\") pod \"64316e8a-a9d2-42a4-a572-359b33ca510c\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.295653 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64316e8a-a9d2-42a4-a572-359b33ca510c-logs\") pod \"64316e8a-a9d2-42a4-a572-359b33ca510c\" (UID: \"64316e8a-a9d2-42a4-a572-359b33ca510c\") " Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.296443 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64316e8a-a9d2-42a4-a572-359b33ca510c-logs" (OuterVolumeSpecName: "logs") pod "64316e8a-a9d2-42a4-a572-359b33ca510c" (UID: "64316e8a-a9d2-42a4-a572-359b33ca510c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.296552 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64316e8a-a9d2-42a4-a572-359b33ca510c-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.308937 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "64316e8a-a9d2-42a4-a572-359b33ca510c" (UID: "64316e8a-a9d2-42a4-a572-359b33ca510c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.316911 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64316e8a-a9d2-42a4-a572-359b33ca510c-kube-api-access-b2bhm" (OuterVolumeSpecName: "kube-api-access-b2bhm") pod "64316e8a-a9d2-42a4-a572-359b33ca510c" (UID: "64316e8a-a9d2-42a4-a572-359b33ca510c"). InnerVolumeSpecName "kube-api-access-b2bhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.346670 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64316e8a-a9d2-42a4-a572-359b33ca510c" (UID: "64316e8a-a9d2-42a4-a572-359b33ca510c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.348993 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data" (OuterVolumeSpecName: "config-data") pod "64316e8a-a9d2-42a4-a572-359b33ca510c" (UID: "64316e8a-a9d2-42a4-a572-359b33ca510c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.397637 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.397680 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.397695 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64316e8a-a9d2-42a4-a572-359b33ca510c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:01 crc kubenswrapper[4846]: I0202 13:42:01.397707 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2bhm\" (UniqueName: \"kubernetes.io/projected/64316e8a-a9d2-42a4-a572-359b33ca510c-kube-api-access-b2bhm\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:02 crc kubenswrapper[4846]: I0202 13:42:02.214234 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54f8585bfb-zsz4k" Feb 02 13:42:02 crc kubenswrapper[4846]: I0202 13:42:02.241359 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-54f8585bfb-zsz4k"] Feb 02 13:42:02 crc kubenswrapper[4846]: I0202 13:42:02.254805 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-54f8585bfb-zsz4k"] Feb 02 13:42:03 crc kubenswrapper[4846]: I0202 13:42:03.454149 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" path="/var/lib/kubelet/pods/64316e8a-a9d2-42a4-a572-359b33ca510c/volumes" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.714976 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-s4ngv"] Feb 02 13:42:04 crc kubenswrapper[4846]: E0202 13:42:04.715384 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerName="barbican-api" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.715402 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerName="barbican-api" Feb 02 13:42:04 crc kubenswrapper[4846]: E0202 13:42:04.715414 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerName="barbican-api-log" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.715422 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerName="barbican-api-log" Feb 02 13:42:04 crc kubenswrapper[4846]: E0202 13:42:04.715432 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b036df9b-ce4c-49f2-8cbc-47bacccabd05" containerName="dnsmasq-dns" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.715439 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b036df9b-ce4c-49f2-8cbc-47bacccabd05" containerName="dnsmasq-dns" Feb 02 13:42:04 crc kubenswrapper[4846]: E0202 13:42:04.715455 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b036df9b-ce4c-49f2-8cbc-47bacccabd05" containerName="init" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.715462 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b036df9b-ce4c-49f2-8cbc-47bacccabd05" containerName="init" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.715659 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerName="barbican-api-log" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.715676 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="64316e8a-a9d2-42a4-a572-359b33ca510c" containerName="barbican-api" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.715694 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b036df9b-ce4c-49f2-8cbc-47bacccabd05" containerName="dnsmasq-dns" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.716375 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s4ngv" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.725037 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-s4ngv"] Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.820212 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-edfe-account-create-update-ks4pd"] Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.821936 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-edfe-account-create-update-ks4pd" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.824229 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.831136 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-edfe-account-create-update-ks4pd"] Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.867209 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw74g\" (UniqueName: \"kubernetes.io/projected/d8175f9f-b434-44ff-b770-252b30d2cf73-kube-api-access-dw74g\") pod \"neutron-db-create-s4ngv\" (UID: \"d8175f9f-b434-44ff-b770-252b30d2cf73\") " pod="openstack/neutron-db-create-s4ngv" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.867325 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8175f9f-b434-44ff-b770-252b30d2cf73-operator-scripts\") pod \"neutron-db-create-s4ngv\" (UID: \"d8175f9f-b434-44ff-b770-252b30d2cf73\") " pod="openstack/neutron-db-create-s4ngv" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.968816 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2726ef7-8ba0-4415-86e0-cf918301446a-operator-scripts\") pod \"neutron-edfe-account-create-update-ks4pd\" (UID: \"d2726ef7-8ba0-4415-86e0-cf918301446a\") " pod="openstack/neutron-edfe-account-create-update-ks4pd" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.969227 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8175f9f-b434-44ff-b770-252b30d2cf73-operator-scripts\") pod \"neutron-db-create-s4ngv\" (UID: \"d8175f9f-b434-44ff-b770-252b30d2cf73\") " pod="openstack/neutron-db-create-s4ngv" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.969381 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccjm2\" (UniqueName: \"kubernetes.io/projected/d2726ef7-8ba0-4415-86e0-cf918301446a-kube-api-access-ccjm2\") pod \"neutron-edfe-account-create-update-ks4pd\" (UID: \"d2726ef7-8ba0-4415-86e0-cf918301446a\") " pod="openstack/neutron-edfe-account-create-update-ks4pd" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.969490 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw74g\" (UniqueName: \"kubernetes.io/projected/d8175f9f-b434-44ff-b770-252b30d2cf73-kube-api-access-dw74g\") pod \"neutron-db-create-s4ngv\" (UID: \"d8175f9f-b434-44ff-b770-252b30d2cf73\") " pod="openstack/neutron-db-create-s4ngv" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.970173 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8175f9f-b434-44ff-b770-252b30d2cf73-operator-scripts\") pod \"neutron-db-create-s4ngv\" (UID: \"d8175f9f-b434-44ff-b770-252b30d2cf73\") " pod="openstack/neutron-db-create-s4ngv" Feb 02 13:42:04 crc kubenswrapper[4846]: I0202 13:42:04.994494 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw74g\" (UniqueName: \"kubernetes.io/projected/d8175f9f-b434-44ff-b770-252b30d2cf73-kube-api-access-dw74g\") pod \"neutron-db-create-s4ngv\" (UID: \"d8175f9f-b434-44ff-b770-252b30d2cf73\") " pod="openstack/neutron-db-create-s4ngv" Feb 02 13:42:05 crc kubenswrapper[4846]: I0202 13:42:05.032689 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s4ngv" Feb 02 13:42:05 crc kubenswrapper[4846]: I0202 13:42:05.071549 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccjm2\" (UniqueName: \"kubernetes.io/projected/d2726ef7-8ba0-4415-86e0-cf918301446a-kube-api-access-ccjm2\") pod \"neutron-edfe-account-create-update-ks4pd\" (UID: \"d2726ef7-8ba0-4415-86e0-cf918301446a\") " pod="openstack/neutron-edfe-account-create-update-ks4pd" Feb 02 13:42:05 crc kubenswrapper[4846]: I0202 13:42:05.071735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2726ef7-8ba0-4415-86e0-cf918301446a-operator-scripts\") pod \"neutron-edfe-account-create-update-ks4pd\" (UID: \"d2726ef7-8ba0-4415-86e0-cf918301446a\") " pod="openstack/neutron-edfe-account-create-update-ks4pd" Feb 02 13:42:05 crc kubenswrapper[4846]: I0202 13:42:05.073002 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2726ef7-8ba0-4415-86e0-cf918301446a-operator-scripts\") pod \"neutron-edfe-account-create-update-ks4pd\" (UID: \"d2726ef7-8ba0-4415-86e0-cf918301446a\") " pod="openstack/neutron-edfe-account-create-update-ks4pd" Feb 02 13:42:05 crc kubenswrapper[4846]: I0202 13:42:05.090228 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccjm2\" (UniqueName: \"kubernetes.io/projected/d2726ef7-8ba0-4415-86e0-cf918301446a-kube-api-access-ccjm2\") pod \"neutron-edfe-account-create-update-ks4pd\" (UID: \"d2726ef7-8ba0-4415-86e0-cf918301446a\") " pod="openstack/neutron-edfe-account-create-update-ks4pd" Feb 02 13:42:05 crc kubenswrapper[4846]: I0202 13:42:05.138643 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-edfe-account-create-update-ks4pd" Feb 02 13:42:05 crc kubenswrapper[4846]: I0202 13:42:05.480881 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-s4ngv"] Feb 02 13:42:05 crc kubenswrapper[4846]: W0202 13:42:05.482146 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8175f9f_b434_44ff_b770_252b30d2cf73.slice/crio-66f33a7087bd2715b27bbbc7381b825a91eed249218aa725c07799f02a975b2f WatchSource:0}: Error finding container 66f33a7087bd2715b27bbbc7381b825a91eed249218aa725c07799f02a975b2f: Status 404 returned error can't find the container with id 66f33a7087bd2715b27bbbc7381b825a91eed249218aa725c07799f02a975b2f Feb 02 13:42:05 crc kubenswrapper[4846]: I0202 13:42:05.607779 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-edfe-account-create-update-ks4pd"] Feb 02 13:42:05 crc kubenswrapper[4846]: W0202 13:42:05.617514 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2726ef7_8ba0_4415_86e0_cf918301446a.slice/crio-53d98161ad18ef7fd934f72cee1fe546e22683ffe43c6300f54acd9ce5ef34c7 WatchSource:0}: Error finding container 53d98161ad18ef7fd934f72cee1fe546e22683ffe43c6300f54acd9ce5ef34c7: Status 404 returned error can't find the container with id 53d98161ad18ef7fd934f72cee1fe546e22683ffe43c6300f54acd9ce5ef34c7 Feb 02 13:42:06 crc kubenswrapper[4846]: I0202 13:42:06.248790 4846 generic.go:334] "Generic (PLEG): container finished" podID="d8175f9f-b434-44ff-b770-252b30d2cf73" containerID="6156327aecadf498e51f3ef6d8243ebb8c487db008b029f9fb2954bb8cacaaa2" exitCode=0 Feb 02 13:42:06 crc kubenswrapper[4846]: I0202 13:42:06.248883 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s4ngv" event={"ID":"d8175f9f-b434-44ff-b770-252b30d2cf73","Type":"ContainerDied","Data":"6156327aecadf498e51f3ef6d8243ebb8c487db008b029f9fb2954bb8cacaaa2"} Feb 02 13:42:06 crc kubenswrapper[4846]: I0202 13:42:06.249063 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s4ngv" event={"ID":"d8175f9f-b434-44ff-b770-252b30d2cf73","Type":"ContainerStarted","Data":"66f33a7087bd2715b27bbbc7381b825a91eed249218aa725c07799f02a975b2f"} Feb 02 13:42:06 crc kubenswrapper[4846]: I0202 13:42:06.252686 4846 generic.go:334] "Generic (PLEG): container finished" podID="d2726ef7-8ba0-4415-86e0-cf918301446a" containerID="66d119cc3db1736125c3cf810a2c411382e995cc2ec2648c44eb775d460acd9e" exitCode=0 Feb 02 13:42:06 crc kubenswrapper[4846]: I0202 13:42:06.252732 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-edfe-account-create-update-ks4pd" event={"ID":"d2726ef7-8ba0-4415-86e0-cf918301446a","Type":"ContainerDied","Data":"66d119cc3db1736125c3cf810a2c411382e995cc2ec2648c44eb775d460acd9e"} Feb 02 13:42:06 crc kubenswrapper[4846]: I0202 13:42:06.252766 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-edfe-account-create-update-ks4pd" event={"ID":"d2726ef7-8ba0-4415-86e0-cf918301446a","Type":"ContainerStarted","Data":"53d98161ad18ef7fd934f72cee1fe546e22683ffe43c6300f54acd9ce5ef34c7"} Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.704219 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-edfe-account-create-update-ks4pd" Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.712092 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s4ngv" Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.822327 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8175f9f-b434-44ff-b770-252b30d2cf73-operator-scripts\") pod \"d8175f9f-b434-44ff-b770-252b30d2cf73\" (UID: \"d8175f9f-b434-44ff-b770-252b30d2cf73\") " Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.822936 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2726ef7-8ba0-4415-86e0-cf918301446a-operator-scripts\") pod \"d2726ef7-8ba0-4415-86e0-cf918301446a\" (UID: \"d2726ef7-8ba0-4415-86e0-cf918301446a\") " Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.823092 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw74g\" (UniqueName: \"kubernetes.io/projected/d8175f9f-b434-44ff-b770-252b30d2cf73-kube-api-access-dw74g\") pod \"d8175f9f-b434-44ff-b770-252b30d2cf73\" (UID: \"d8175f9f-b434-44ff-b770-252b30d2cf73\") " Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.823333 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccjm2\" (UniqueName: \"kubernetes.io/projected/d2726ef7-8ba0-4415-86e0-cf918301446a-kube-api-access-ccjm2\") pod \"d2726ef7-8ba0-4415-86e0-cf918301446a\" (UID: \"d2726ef7-8ba0-4415-86e0-cf918301446a\") " Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.824590 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2726ef7-8ba0-4415-86e0-cf918301446a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2726ef7-8ba0-4415-86e0-cf918301446a" (UID: "d2726ef7-8ba0-4415-86e0-cf918301446a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.824892 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8175f9f-b434-44ff-b770-252b30d2cf73-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d8175f9f-b434-44ff-b770-252b30d2cf73" (UID: "d8175f9f-b434-44ff-b770-252b30d2cf73"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.825131 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8175f9f-b434-44ff-b770-252b30d2cf73-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.825166 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2726ef7-8ba0-4415-86e0-cf918301446a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.835036 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2726ef7-8ba0-4415-86e0-cf918301446a-kube-api-access-ccjm2" (OuterVolumeSpecName: "kube-api-access-ccjm2") pod "d2726ef7-8ba0-4415-86e0-cf918301446a" (UID: "d2726ef7-8ba0-4415-86e0-cf918301446a"). InnerVolumeSpecName "kube-api-access-ccjm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.835588 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8175f9f-b434-44ff-b770-252b30d2cf73-kube-api-access-dw74g" (OuterVolumeSpecName: "kube-api-access-dw74g") pod "d8175f9f-b434-44ff-b770-252b30d2cf73" (UID: "d8175f9f-b434-44ff-b770-252b30d2cf73"). InnerVolumeSpecName "kube-api-access-dw74g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.926762 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw74g\" (UniqueName: \"kubernetes.io/projected/d8175f9f-b434-44ff-b770-252b30d2cf73-kube-api-access-dw74g\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:07 crc kubenswrapper[4846]: I0202 13:42:07.926805 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccjm2\" (UniqueName: \"kubernetes.io/projected/d2726ef7-8ba0-4415-86e0-cf918301446a-kube-api-access-ccjm2\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:08 crc kubenswrapper[4846]: I0202 13:42:08.272408 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-edfe-account-create-update-ks4pd" event={"ID":"d2726ef7-8ba0-4415-86e0-cf918301446a","Type":"ContainerDied","Data":"53d98161ad18ef7fd934f72cee1fe546e22683ffe43c6300f54acd9ce5ef34c7"} Feb 02 13:42:08 crc kubenswrapper[4846]: I0202 13:42:08.272463 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53d98161ad18ef7fd934f72cee1fe546e22683ffe43c6300f54acd9ce5ef34c7" Feb 02 13:42:08 crc kubenswrapper[4846]: I0202 13:42:08.272458 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-edfe-account-create-update-ks4pd" Feb 02 13:42:08 crc kubenswrapper[4846]: I0202 13:42:08.274817 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s4ngv" event={"ID":"d8175f9f-b434-44ff-b770-252b30d2cf73","Type":"ContainerDied","Data":"66f33a7087bd2715b27bbbc7381b825a91eed249218aa725c07799f02a975b2f"} Feb 02 13:42:08 crc kubenswrapper[4846]: I0202 13:42:08.274864 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66f33a7087bd2715b27bbbc7381b825a91eed249218aa725c07799f02a975b2f" Feb 02 13:42:08 crc kubenswrapper[4846]: I0202 13:42:08.274872 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s4ngv" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.158998 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4glkm"] Feb 02 13:42:10 crc kubenswrapper[4846]: E0202 13:42:10.159802 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2726ef7-8ba0-4415-86e0-cf918301446a" containerName="mariadb-account-create-update" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.159822 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2726ef7-8ba0-4415-86e0-cf918301446a" containerName="mariadb-account-create-update" Feb 02 13:42:10 crc kubenswrapper[4846]: E0202 13:42:10.159851 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8175f9f-b434-44ff-b770-252b30d2cf73" containerName="mariadb-database-create" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.159860 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8175f9f-b434-44ff-b770-252b30d2cf73" containerName="mariadb-database-create" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.160048 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8175f9f-b434-44ff-b770-252b30d2cf73" containerName="mariadb-database-create" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.160075 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2726ef7-8ba0-4415-86e0-cf918301446a" containerName="mariadb-account-create-update" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.160768 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.163529 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dqhcz" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.163813 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.165686 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.177456 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4glkm"] Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.265725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-config\") pod \"neutron-db-sync-4glkm\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.265779 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkbnp\" (UniqueName: \"kubernetes.io/projected/77c19163-c5ff-45f2-8410-87c0eee441a0-kube-api-access-rkbnp\") pod \"neutron-db-sync-4glkm\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.265901 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-combined-ca-bundle\") pod \"neutron-db-sync-4glkm\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.367669 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-config\") pod \"neutron-db-sync-4glkm\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.367729 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkbnp\" (UniqueName: \"kubernetes.io/projected/77c19163-c5ff-45f2-8410-87c0eee441a0-kube-api-access-rkbnp\") pod \"neutron-db-sync-4glkm\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.367771 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-combined-ca-bundle\") pod \"neutron-db-sync-4glkm\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.372747 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-config\") pod \"neutron-db-sync-4glkm\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.373915 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-combined-ca-bundle\") pod \"neutron-db-sync-4glkm\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.389557 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkbnp\" (UniqueName: \"kubernetes.io/projected/77c19163-c5ff-45f2-8410-87c0eee441a0-kube-api-access-rkbnp\") pod \"neutron-db-sync-4glkm\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.479454 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:10 crc kubenswrapper[4846]: I0202 13:42:10.975013 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4glkm"] Feb 02 13:42:11 crc kubenswrapper[4846]: I0202 13:42:11.315344 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4glkm" event={"ID":"77c19163-c5ff-45f2-8410-87c0eee441a0","Type":"ContainerStarted","Data":"9ae519a9b309c537001888466260ad26927cfdf3c7b19f69c4f9ea422bff5265"} Feb 02 13:42:11 crc kubenswrapper[4846]: I0202 13:42:11.315416 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4glkm" event={"ID":"77c19163-c5ff-45f2-8410-87c0eee441a0","Type":"ContainerStarted","Data":"57e4aa2341dfa73ee0264380389581f6f4c2b3281c4292f26aac9e04adf99a04"} Feb 02 13:42:11 crc kubenswrapper[4846]: I0202 13:42:11.338460 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4glkm" podStartSLOduration=1.3384353629999999 podStartE2EDuration="1.338435363s" podCreationTimestamp="2026-02-02 13:42:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:42:11.330838529 +0000 UTC m=+5562.559425392" watchObservedRunningTime="2026-02-02 13:42:11.338435363 +0000 UTC m=+5562.567022226" Feb 02 13:42:12 crc kubenswrapper[4846]: I0202 13:42:12.068692 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-mqzll"] Feb 02 13:42:12 crc kubenswrapper[4846]: I0202 13:42:12.077610 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-mqzll"] Feb 02 13:42:13 crc kubenswrapper[4846]: I0202 13:42:13.433477 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ab60cda-aa99-4608-9011-6e9e569ca203" path="/var/lib/kubelet/pods/8ab60cda-aa99-4608-9011-6e9e569ca203/volumes" Feb 02 13:42:16 crc kubenswrapper[4846]: I0202 13:42:16.378916 4846 generic.go:334] "Generic (PLEG): container finished" podID="77c19163-c5ff-45f2-8410-87c0eee441a0" containerID="9ae519a9b309c537001888466260ad26927cfdf3c7b19f69c4f9ea422bff5265" exitCode=0 Feb 02 13:42:16 crc kubenswrapper[4846]: I0202 13:42:16.379013 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4glkm" event={"ID":"77c19163-c5ff-45f2-8410-87c0eee441a0","Type":"ContainerDied","Data":"9ae519a9b309c537001888466260ad26927cfdf3c7b19f69c4f9ea422bff5265"} Feb 02 13:42:17 crc kubenswrapper[4846]: I0202 13:42:17.728369 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:17 crc kubenswrapper[4846]: I0202 13:42:17.819284 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-config\") pod \"77c19163-c5ff-45f2-8410-87c0eee441a0\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " Feb 02 13:42:17 crc kubenswrapper[4846]: I0202 13:42:17.819393 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-combined-ca-bundle\") pod \"77c19163-c5ff-45f2-8410-87c0eee441a0\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " Feb 02 13:42:17 crc kubenswrapper[4846]: I0202 13:42:17.819549 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkbnp\" (UniqueName: \"kubernetes.io/projected/77c19163-c5ff-45f2-8410-87c0eee441a0-kube-api-access-rkbnp\") pod \"77c19163-c5ff-45f2-8410-87c0eee441a0\" (UID: \"77c19163-c5ff-45f2-8410-87c0eee441a0\") " Feb 02 13:42:17 crc kubenswrapper[4846]: I0202 13:42:17.826993 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77c19163-c5ff-45f2-8410-87c0eee441a0-kube-api-access-rkbnp" (OuterVolumeSpecName: "kube-api-access-rkbnp") pod "77c19163-c5ff-45f2-8410-87c0eee441a0" (UID: "77c19163-c5ff-45f2-8410-87c0eee441a0"). InnerVolumeSpecName "kube-api-access-rkbnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:42:17 crc kubenswrapper[4846]: I0202 13:42:17.854011 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77c19163-c5ff-45f2-8410-87c0eee441a0" (UID: "77c19163-c5ff-45f2-8410-87c0eee441a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:42:17 crc kubenswrapper[4846]: I0202 13:42:17.855299 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-config" (OuterVolumeSpecName: "config") pod "77c19163-c5ff-45f2-8410-87c0eee441a0" (UID: "77c19163-c5ff-45f2-8410-87c0eee441a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:42:17 crc kubenswrapper[4846]: I0202 13:42:17.922328 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkbnp\" (UniqueName: \"kubernetes.io/projected/77c19163-c5ff-45f2-8410-87c0eee441a0-kube-api-access-rkbnp\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:17 crc kubenswrapper[4846]: I0202 13:42:17.922375 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:17 crc kubenswrapper[4846]: I0202 13:42:17.922387 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77c19163-c5ff-45f2-8410-87c0eee441a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.399807 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4glkm" event={"ID":"77c19163-c5ff-45f2-8410-87c0eee441a0","Type":"ContainerDied","Data":"57e4aa2341dfa73ee0264380389581f6f4c2b3281c4292f26aac9e04adf99a04"} Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.399866 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57e4aa2341dfa73ee0264380389581f6f4c2b3281c4292f26aac9e04adf99a04" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.399901 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4glkm" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.562722 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77d8d6f48f-vhhwk"] Feb 02 13:42:18 crc kubenswrapper[4846]: E0202 13:42:18.563853 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77c19163-c5ff-45f2-8410-87c0eee441a0" containerName="neutron-db-sync" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.563879 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="77c19163-c5ff-45f2-8410-87c0eee441a0" containerName="neutron-db-sync" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.564074 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="77c19163-c5ff-45f2-8410-87c0eee441a0" containerName="neutron-db-sync" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.565313 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.580173 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77d8d6f48f-vhhwk"] Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.635726 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.635954 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-config\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.636126 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.636349 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqgfb\" (UniqueName: \"kubernetes.io/projected/60adbe37-7588-48e4-8d77-3129ac673426-kube-api-access-mqgfb\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.636485 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-dns-svc\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.693592 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7c8ff4b488-6qprd"] Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.698138 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.703717 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.703982 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dqhcz" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.704191 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.704346 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.716300 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c8ff4b488-6qprd"] Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.739265 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqgfb\" (UniqueName: \"kubernetes.io/projected/60adbe37-7588-48e4-8d77-3129ac673426-kube-api-access-mqgfb\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.739359 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-httpd-config\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.739389 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-combined-ca-bundle\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.739416 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-dns-svc\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.739447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.739502 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-config\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.739530 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-ovndb-tls-certs\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.739555 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-config\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.739578 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z84x\" (UniqueName: \"kubernetes.io/projected/5525c4be-eef3-486b-80d7-57e17a645f24-kube-api-access-9z84x\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.739648 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.741699 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-sb\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.745504 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-nb\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.746145 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-config\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.746504 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-dns-svc\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.768916 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqgfb\" (UniqueName: \"kubernetes.io/projected/60adbe37-7588-48e4-8d77-3129ac673426-kube-api-access-mqgfb\") pod \"dnsmasq-dns-77d8d6f48f-vhhwk\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.845384 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-httpd-config\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.845504 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-combined-ca-bundle\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.845595 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-config\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.845651 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-ovndb-tls-certs\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.845685 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z84x\" (UniqueName: \"kubernetes.io/projected/5525c4be-eef3-486b-80d7-57e17a645f24-kube-api-access-9z84x\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.852491 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-httpd-config\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.854188 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-ovndb-tls-certs\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.854371 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-config\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.855382 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-combined-ca-bundle\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.874638 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z84x\" (UniqueName: \"kubernetes.io/projected/5525c4be-eef3-486b-80d7-57e17a645f24-kube-api-access-9z84x\") pod \"neutron-7c8ff4b488-6qprd\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:18 crc kubenswrapper[4846]: I0202 13:42:18.890691 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:19 crc kubenswrapper[4846]: I0202 13:42:19.031281 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:19 crc kubenswrapper[4846]: I0202 13:42:19.443691 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77d8d6f48f-vhhwk"] Feb 02 13:42:19 crc kubenswrapper[4846]: I0202 13:42:19.726877 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c8ff4b488-6qprd"] Feb 02 13:42:19 crc kubenswrapper[4846]: W0202 13:42:19.734808 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5525c4be_eef3_486b_80d7_57e17a645f24.slice/crio-13ade089a1e65c2fedc0c4520b6efda990ddf291a34196fe9ced7d65481e7662 WatchSource:0}: Error finding container 13ade089a1e65c2fedc0c4520b6efda990ddf291a34196fe9ced7d65481e7662: Status 404 returned error can't find the container with id 13ade089a1e65c2fedc0c4520b6efda990ddf291a34196fe9ced7d65481e7662 Feb 02 13:42:20 crc kubenswrapper[4846]: I0202 13:42:20.426696 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8ff4b488-6qprd" event={"ID":"5525c4be-eef3-486b-80d7-57e17a645f24","Type":"ContainerStarted","Data":"90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60"} Feb 02 13:42:20 crc kubenswrapper[4846]: I0202 13:42:20.427176 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:20 crc kubenswrapper[4846]: I0202 13:42:20.427192 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8ff4b488-6qprd" event={"ID":"5525c4be-eef3-486b-80d7-57e17a645f24","Type":"ContainerStarted","Data":"19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377"} Feb 02 13:42:20 crc kubenswrapper[4846]: I0202 13:42:20.427207 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8ff4b488-6qprd" event={"ID":"5525c4be-eef3-486b-80d7-57e17a645f24","Type":"ContainerStarted","Data":"13ade089a1e65c2fedc0c4520b6efda990ddf291a34196fe9ced7d65481e7662"} Feb 02 13:42:20 crc kubenswrapper[4846]: I0202 13:42:20.429200 4846 generic.go:334] "Generic (PLEG): container finished" podID="60adbe37-7588-48e4-8d77-3129ac673426" containerID="afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809" exitCode=0 Feb 02 13:42:20 crc kubenswrapper[4846]: I0202 13:42:20.429237 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" event={"ID":"60adbe37-7588-48e4-8d77-3129ac673426","Type":"ContainerDied","Data":"afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809"} Feb 02 13:42:20 crc kubenswrapper[4846]: I0202 13:42:20.429260 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" event={"ID":"60adbe37-7588-48e4-8d77-3129ac673426","Type":"ContainerStarted","Data":"946c65cd446c4db7692e20603573e38e1fd6f1419be22771887e6a6ebe04a7e2"} Feb 02 13:42:20 crc kubenswrapper[4846]: I0202 13:42:20.490713 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7c8ff4b488-6qprd" podStartSLOduration=2.490686192 podStartE2EDuration="2.490686192s" podCreationTimestamp="2026-02-02 13:42:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:42:20.457704562 +0000 UTC m=+5571.686291435" watchObservedRunningTime="2026-02-02 13:42:20.490686192 +0000 UTC m=+5571.719273075" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.446425 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" event={"ID":"60adbe37-7588-48e4-8d77-3129ac673426","Type":"ContainerStarted","Data":"bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b"} Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.446847 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.472185 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" podStartSLOduration=3.472164598 podStartE2EDuration="3.472164598s" podCreationTimestamp="2026-02-02 13:42:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:42:21.463052667 +0000 UTC m=+5572.691639530" watchObservedRunningTime="2026-02-02 13:42:21.472164598 +0000 UTC m=+5572.700751461" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.749727 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d7f794577-lckxg"] Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.751338 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.753707 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.755331 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.762386 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d7f794577-lckxg"] Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.855538 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmsvg\" (UniqueName: \"kubernetes.io/projected/04c833b7-c2fd-4697-b77b-5a777c7678e8-kube-api-access-pmsvg\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.855588 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-internal-tls-certs\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.855691 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-public-tls-certs\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.855715 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-config\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.855745 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-combined-ca-bundle\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.855786 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-httpd-config\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.855822 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-ovndb-tls-certs\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.957654 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-ovndb-tls-certs\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.957755 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmsvg\" (UniqueName: \"kubernetes.io/projected/04c833b7-c2fd-4697-b77b-5a777c7678e8-kube-api-access-pmsvg\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.957781 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-internal-tls-certs\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.957829 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-public-tls-certs\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.957851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-config\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.957879 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-combined-ca-bundle\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.957918 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-httpd-config\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.966373 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-public-tls-certs\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.975331 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-internal-tls-certs\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.975398 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-ovndb-tls-certs\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.975951 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-httpd-config\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.976893 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-combined-ca-bundle\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.978108 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/04c833b7-c2fd-4697-b77b-5a777c7678e8-config\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:21 crc kubenswrapper[4846]: I0202 13:42:21.981890 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmsvg\" (UniqueName: \"kubernetes.io/projected/04c833b7-c2fd-4697-b77b-5a777c7678e8-kube-api-access-pmsvg\") pod \"neutron-5d7f794577-lckxg\" (UID: \"04c833b7-c2fd-4697-b77b-5a777c7678e8\") " pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:22 crc kubenswrapper[4846]: I0202 13:42:22.071372 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:22 crc kubenswrapper[4846]: I0202 13:42:22.641524 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d7f794577-lckxg"] Feb 02 13:42:23 crc kubenswrapper[4846]: I0202 13:42:23.464669 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d7f794577-lckxg" event={"ID":"04c833b7-c2fd-4697-b77b-5a777c7678e8","Type":"ContainerStarted","Data":"b79a2a02f35af5ad82254633346b12596ce3d2e46043201b9ebaf6fb3f6347bc"} Feb 02 13:42:23 crc kubenswrapper[4846]: I0202 13:42:23.465071 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d7f794577-lckxg" event={"ID":"04c833b7-c2fd-4697-b77b-5a777c7678e8","Type":"ContainerStarted","Data":"40492ded599542aa4b0c8e15f4b1f2106d509f61f1c0efc8390b5c3d0d28e5e1"} Feb 02 13:42:23 crc kubenswrapper[4846]: I0202 13:42:23.465089 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d7f794577-lckxg" event={"ID":"04c833b7-c2fd-4697-b77b-5a777c7678e8","Type":"ContainerStarted","Data":"711adb9ef13f72d4a3356aed7241ec5e4520d3b04e96bb91ad4cb283a6d7133e"} Feb 02 13:42:23 crc kubenswrapper[4846]: I0202 13:42:23.466478 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:23 crc kubenswrapper[4846]: I0202 13:42:23.496867 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d7f794577-lckxg" podStartSLOduration=2.496849279 podStartE2EDuration="2.496849279s" podCreationTimestamp="2026-02-02 13:42:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:42:23.484081129 +0000 UTC m=+5574.712668002" watchObservedRunningTime="2026-02-02 13:42:23.496849279 +0000 UTC m=+5574.725436142" Feb 02 13:42:28 crc kubenswrapper[4846]: I0202 13:42:28.892753 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:42:28 crc kubenswrapper[4846]: I0202 13:42:28.952129 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66845c4585-49jp4"] Feb 02 13:42:28 crc kubenswrapper[4846]: I0202 13:42:28.952390 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66845c4585-49jp4" podUID="76bf5119-ebc0-4687-9b32-9c5bf37b18de" containerName="dnsmasq-dns" containerID="cri-o://1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c" gracePeriod=10 Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.431948 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.496116 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-sb\") pod \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.496266 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5w45\" (UniqueName: \"kubernetes.io/projected/76bf5119-ebc0-4687-9b32-9c5bf37b18de-kube-api-access-b5w45\") pod \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.496322 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-config\") pod \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.496373 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-dns-svc\") pod \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.496396 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-nb\") pod \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\" (UID: \"76bf5119-ebc0-4687-9b32-9c5bf37b18de\") " Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.506537 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76bf5119-ebc0-4687-9b32-9c5bf37b18de-kube-api-access-b5w45" (OuterVolumeSpecName: "kube-api-access-b5w45") pod "76bf5119-ebc0-4687-9b32-9c5bf37b18de" (UID: "76bf5119-ebc0-4687-9b32-9c5bf37b18de"). InnerVolumeSpecName "kube-api-access-b5w45". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.522363 4846 generic.go:334] "Generic (PLEG): container finished" podID="76bf5119-ebc0-4687-9b32-9c5bf37b18de" containerID="1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c" exitCode=0 Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.522414 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66845c4585-49jp4" event={"ID":"76bf5119-ebc0-4687-9b32-9c5bf37b18de","Type":"ContainerDied","Data":"1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c"} Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.522416 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66845c4585-49jp4" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.522462 4846 scope.go:117] "RemoveContainer" containerID="1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.522446 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66845c4585-49jp4" event={"ID":"76bf5119-ebc0-4687-9b32-9c5bf37b18de","Type":"ContainerDied","Data":"a0bb4be74e504ada472dcb1ecaa4de1ae0a66674b606ee08a7f41155fd9a2884"} Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.550383 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "76bf5119-ebc0-4687-9b32-9c5bf37b18de" (UID: "76bf5119-ebc0-4687-9b32-9c5bf37b18de"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.552321 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-config" (OuterVolumeSpecName: "config") pod "76bf5119-ebc0-4687-9b32-9c5bf37b18de" (UID: "76bf5119-ebc0-4687-9b32-9c5bf37b18de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.554241 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "76bf5119-ebc0-4687-9b32-9c5bf37b18de" (UID: "76bf5119-ebc0-4687-9b32-9c5bf37b18de"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.558394 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "76bf5119-ebc0-4687-9b32-9c5bf37b18de" (UID: "76bf5119-ebc0-4687-9b32-9c5bf37b18de"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.596403 4846 scope.go:117] "RemoveContainer" containerID="a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.598230 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5w45\" (UniqueName: \"kubernetes.io/projected/76bf5119-ebc0-4687-9b32-9c5bf37b18de-kube-api-access-b5w45\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.598261 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.598278 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.598290 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.598304 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76bf5119-ebc0-4687-9b32-9c5bf37b18de-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.611810 4846 scope.go:117] "RemoveContainer" containerID="1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c" Feb 02 13:42:29 crc kubenswrapper[4846]: E0202 13:42:29.612481 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c\": container with ID starting with 1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c not found: ID does not exist" containerID="1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.612519 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c"} err="failed to get container status \"1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c\": rpc error: code = NotFound desc = could not find container \"1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c\": container with ID starting with 1de28be6cd9f24ff21311f12f702fede8d837da7671839eb5f9e4bd6994df30c not found: ID does not exist" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.612561 4846 scope.go:117] "RemoveContainer" containerID="a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1" Feb 02 13:42:29 crc kubenswrapper[4846]: E0202 13:42:29.613657 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1\": container with ID starting with a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1 not found: ID does not exist" containerID="a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.613688 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1"} err="failed to get container status \"a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1\": rpc error: code = NotFound desc = could not find container \"a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1\": container with ID starting with a314db1b95d8d7a6a3faae4200bb85417d8c71a04472fbc42e5c22c51bf1e7b1 not found: ID does not exist" Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.852919 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66845c4585-49jp4"] Feb 02 13:42:29 crc kubenswrapper[4846]: I0202 13:42:29.859987 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66845c4585-49jp4"] Feb 02 13:42:30 crc kubenswrapper[4846]: I0202 13:42:30.479194 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:42:30 crc kubenswrapper[4846]: I0202 13:42:30.479262 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:42:31 crc kubenswrapper[4846]: I0202 13:42:31.434537 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76bf5119-ebc0-4687-9b32-9c5bf37b18de" path="/var/lib/kubelet/pods/76bf5119-ebc0-4687-9b32-9c5bf37b18de/volumes" Feb 02 13:42:38 crc kubenswrapper[4846]: I0202 13:42:38.391557 4846 scope.go:117] "RemoveContainer" containerID="2c0fd3b791357a80593703b92ed8aa00a4c584b1e20dcf0f65e60ee14a6d6f58" Feb 02 13:42:49 crc kubenswrapper[4846]: I0202 13:42:49.041424 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:52 crc kubenswrapper[4846]: I0202 13:42:52.083789 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d7f794577-lckxg" Feb 02 13:42:52 crc kubenswrapper[4846]: I0202 13:42:52.173993 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7c8ff4b488-6qprd"] Feb 02 13:42:52 crc kubenswrapper[4846]: I0202 13:42:52.174205 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7c8ff4b488-6qprd" podUID="5525c4be-eef3-486b-80d7-57e17a645f24" containerName="neutron-api" containerID="cri-o://19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377" gracePeriod=30 Feb 02 13:42:52 crc kubenswrapper[4846]: I0202 13:42:52.174764 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7c8ff4b488-6qprd" podUID="5525c4be-eef3-486b-80d7-57e17a645f24" containerName="neutron-httpd" containerID="cri-o://90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60" gracePeriod=30 Feb 02 13:42:52 crc kubenswrapper[4846]: I0202 13:42:52.725312 4846 generic.go:334] "Generic (PLEG): container finished" podID="5525c4be-eef3-486b-80d7-57e17a645f24" containerID="90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60" exitCode=0 Feb 02 13:42:52 crc kubenswrapper[4846]: I0202 13:42:52.725371 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8ff4b488-6qprd" event={"ID":"5525c4be-eef3-486b-80d7-57e17a645f24","Type":"ContainerDied","Data":"90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60"} Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.427376 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.595874 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-ovndb-tls-certs\") pod \"5525c4be-eef3-486b-80d7-57e17a645f24\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.595997 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-config\") pod \"5525c4be-eef3-486b-80d7-57e17a645f24\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.596070 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-httpd-config\") pod \"5525c4be-eef3-486b-80d7-57e17a645f24\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.596118 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-combined-ca-bundle\") pod \"5525c4be-eef3-486b-80d7-57e17a645f24\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.596738 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z84x\" (UniqueName: \"kubernetes.io/projected/5525c4be-eef3-486b-80d7-57e17a645f24-kube-api-access-9z84x\") pod \"5525c4be-eef3-486b-80d7-57e17a645f24\" (UID: \"5525c4be-eef3-486b-80d7-57e17a645f24\") " Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.604247 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5525c4be-eef3-486b-80d7-57e17a645f24-kube-api-access-9z84x" (OuterVolumeSpecName: "kube-api-access-9z84x") pod "5525c4be-eef3-486b-80d7-57e17a645f24" (UID: "5525c4be-eef3-486b-80d7-57e17a645f24"). InnerVolumeSpecName "kube-api-access-9z84x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.604884 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5525c4be-eef3-486b-80d7-57e17a645f24" (UID: "5525c4be-eef3-486b-80d7-57e17a645f24"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.648751 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5525c4be-eef3-486b-80d7-57e17a645f24" (UID: "5525c4be-eef3-486b-80d7-57e17a645f24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.656324 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-config" (OuterVolumeSpecName: "config") pod "5525c4be-eef3-486b-80d7-57e17a645f24" (UID: "5525c4be-eef3-486b-80d7-57e17a645f24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.672023 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5525c4be-eef3-486b-80d7-57e17a645f24" (UID: "5525c4be-eef3-486b-80d7-57e17a645f24"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.699411 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z84x\" (UniqueName: \"kubernetes.io/projected/5525c4be-eef3-486b-80d7-57e17a645f24-kube-api-access-9z84x\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.699459 4846 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.699470 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.699479 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.699490 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5525c4be-eef3-486b-80d7-57e17a645f24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.763752 4846 generic.go:334] "Generic (PLEG): container finished" podID="5525c4be-eef3-486b-80d7-57e17a645f24" containerID="19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377" exitCode=0 Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.763807 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8ff4b488-6qprd" event={"ID":"5525c4be-eef3-486b-80d7-57e17a645f24","Type":"ContainerDied","Data":"19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377"} Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.763845 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c8ff4b488-6qprd" event={"ID":"5525c4be-eef3-486b-80d7-57e17a645f24","Type":"ContainerDied","Data":"13ade089a1e65c2fedc0c4520b6efda990ddf291a34196fe9ced7d65481e7662"} Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.763868 4846 scope.go:117] "RemoveContainer" containerID="90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.764160 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c8ff4b488-6qprd" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.804047 4846 scope.go:117] "RemoveContainer" containerID="19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.812679 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7c8ff4b488-6qprd"] Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.819331 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7c8ff4b488-6qprd"] Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.824228 4846 scope.go:117] "RemoveContainer" containerID="90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60" Feb 02 13:42:56 crc kubenswrapper[4846]: E0202 13:42:56.824879 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60\": container with ID starting with 90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60 not found: ID does not exist" containerID="90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.824930 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60"} err="failed to get container status \"90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60\": rpc error: code = NotFound desc = could not find container \"90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60\": container with ID starting with 90ff94872d250d9b8b736acfd7f8627ba2a1ee3aa9e79f9e25552bee42a6de60 not found: ID does not exist" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.824958 4846 scope.go:117] "RemoveContainer" containerID="19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377" Feb 02 13:42:56 crc kubenswrapper[4846]: E0202 13:42:56.825462 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377\": container with ID starting with 19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377 not found: ID does not exist" containerID="19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377" Feb 02 13:42:56 crc kubenswrapper[4846]: I0202 13:42:56.825500 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377"} err="failed to get container status \"19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377\": rpc error: code = NotFound desc = could not find container \"19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377\": container with ID starting with 19bbafaab2ae53a93df1847a30f4a1c0a93d16240da514f0e19636c338739377 not found: ID does not exist" Feb 02 13:42:57 crc kubenswrapper[4846]: I0202 13:42:57.434480 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5525c4be-eef3-486b-80d7-57e17a645f24" path="/var/lib/kubelet/pods/5525c4be-eef3-486b-80d7-57e17a645f24/volumes" Feb 02 13:43:00 crc kubenswrapper[4846]: I0202 13:43:00.479092 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:43:00 crc kubenswrapper[4846]: I0202 13:43:00.479424 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.658755 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-grx5c"] Feb 02 13:43:15 crc kubenswrapper[4846]: E0202 13:43:15.659772 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76bf5119-ebc0-4687-9b32-9c5bf37b18de" containerName="dnsmasq-dns" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.659792 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="76bf5119-ebc0-4687-9b32-9c5bf37b18de" containerName="dnsmasq-dns" Feb 02 13:43:15 crc kubenswrapper[4846]: E0202 13:43:15.659821 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5525c4be-eef3-486b-80d7-57e17a645f24" containerName="neutron-api" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.659829 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5525c4be-eef3-486b-80d7-57e17a645f24" containerName="neutron-api" Feb 02 13:43:15 crc kubenswrapper[4846]: E0202 13:43:15.659844 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5525c4be-eef3-486b-80d7-57e17a645f24" containerName="neutron-httpd" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.659851 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5525c4be-eef3-486b-80d7-57e17a645f24" containerName="neutron-httpd" Feb 02 13:43:15 crc kubenswrapper[4846]: E0202 13:43:15.659872 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76bf5119-ebc0-4687-9b32-9c5bf37b18de" containerName="init" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.659879 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="76bf5119-ebc0-4687-9b32-9c5bf37b18de" containerName="init" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.660055 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="76bf5119-ebc0-4687-9b32-9c5bf37b18de" containerName="dnsmasq-dns" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.660076 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5525c4be-eef3-486b-80d7-57e17a645f24" containerName="neutron-api" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.660095 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5525c4be-eef3-486b-80d7-57e17a645f24" containerName="neutron-httpd" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.660792 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.664048 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.664299 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-6dfrs" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.672155 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.672986 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.673184 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.687492 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-grx5c"] Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.749156 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-ring-data-devices\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.749238 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-swiftconf\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.749352 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0c65201-a430-4889-a853-00c23cf8658e-etc-swift\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.749477 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-scripts\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.749858 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-combined-ca-bundle\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.749939 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-dispersionconf\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.750063 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7wzc\" (UniqueName: \"kubernetes.io/projected/f0c65201-a430-4889-a853-00c23cf8658e-kube-api-access-v7wzc\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.795011 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f87f9d645-qnvj7"] Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.796973 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.833691 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f87f9d645-qnvj7"] Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.852683 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0c65201-a430-4889-a853-00c23cf8658e-etc-swift\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.852743 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-scripts\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.852820 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-combined-ca-bundle\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.852868 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-dispersionconf\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.852913 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7wzc\" (UniqueName: \"kubernetes.io/projected/f0c65201-a430-4889-a853-00c23cf8658e-kube-api-access-v7wzc\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.852962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-ring-data-devices\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.852995 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-swiftconf\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.854784 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-scripts\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.854807 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0c65201-a430-4889-a853-00c23cf8658e-etc-swift\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.855272 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-ring-data-devices\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.858505 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-combined-ca-bundle\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.861154 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-dispersionconf\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.867988 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-swiftconf\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.881868 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7wzc\" (UniqueName: \"kubernetes.io/projected/f0c65201-a430-4889-a853-00c23cf8658e-kube-api-access-v7wzc\") pod \"swift-ring-rebalance-grx5c\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.954510 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-nb\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.954573 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-dns-svc\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.954614 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-config\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.954820 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pthqx\" (UniqueName: \"kubernetes.io/projected/6dac67a9-bcff-44fe-b4f5-a52de8033984-kube-api-access-pthqx\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:15 crc kubenswrapper[4846]: I0202 13:43:15.954930 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-sb\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.018099 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.056987 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-nb\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.057036 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-dns-svc\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.057059 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-config\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.057124 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pthqx\" (UniqueName: \"kubernetes.io/projected/6dac67a9-bcff-44fe-b4f5-a52de8033984-kube-api-access-pthqx\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.057163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-sb\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.058277 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-nb\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.058301 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-sb\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.059519 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-config\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.061211 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-dns-svc\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.075780 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pthqx\" (UniqueName: \"kubernetes.io/projected/6dac67a9-bcff-44fe-b4f5-a52de8033984-kube-api-access-pthqx\") pod \"dnsmasq-dns-f87f9d645-qnvj7\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.132290 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.586090 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-grx5c"] Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.705182 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f87f9d645-qnvj7"] Feb 02 13:43:16 crc kubenswrapper[4846]: W0202 13:43:16.709425 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dac67a9_bcff_44fe_b4f5_a52de8033984.slice/crio-6c4dec8ae6cf287dbdb5b11299053b2dc14b9cd124497a4b481d796cc7d4a4a4 WatchSource:0}: Error finding container 6c4dec8ae6cf287dbdb5b11299053b2dc14b9cd124497a4b481d796cc7d4a4a4: Status 404 returned error can't find the container with id 6c4dec8ae6cf287dbdb5b11299053b2dc14b9cd124497a4b481d796cc7d4a4a4 Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.936262 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" event={"ID":"6dac67a9-bcff-44fe-b4f5-a52de8033984","Type":"ContainerStarted","Data":"6c4dec8ae6cf287dbdb5b11299053b2dc14b9cd124497a4b481d796cc7d4a4a4"} Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.937416 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-grx5c" event={"ID":"f0c65201-a430-4889-a853-00c23cf8658e","Type":"ContainerStarted","Data":"a221103e1555cac3bb08f7cf5797a0654a4461f8628b0afdd6c76645c32bc7c5"} Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.937444 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-grx5c" event={"ID":"f0c65201-a430-4889-a853-00c23cf8658e","Type":"ContainerStarted","Data":"88c78c5f4101c3f85b6d33f94387a6aca2a40c8f3c963e972eabf0bf9cc0ac73"} Feb 02 13:43:16 crc kubenswrapper[4846]: I0202 13:43:16.960198 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-grx5c" podStartSLOduration=1.9601809270000001 podStartE2EDuration="1.960180927s" podCreationTimestamp="2026-02-02 13:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:43:16.95824393 +0000 UTC m=+5628.186830793" watchObservedRunningTime="2026-02-02 13:43:16.960180927 +0000 UTC m=+5628.188767790" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.789820 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7b4d5c8894-74dxm"] Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.791812 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.797568 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.798791 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b4d5c8894-74dxm"] Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.890488 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-log-httpd\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.890575 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-combined-ca-bundle\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.890610 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-etc-swift\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.890885 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-config-data\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.891245 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-run-httpd\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.891300 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmxrr\" (UniqueName: \"kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-kube-api-access-bmxrr\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.946046 4846 generic.go:334] "Generic (PLEG): container finished" podID="6dac67a9-bcff-44fe-b4f5-a52de8033984" containerID="1be1f1eb7f13040ffddb93fdb888ff9b387c46be744cd5ba0613e4acefae0187" exitCode=0 Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.946975 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" event={"ID":"6dac67a9-bcff-44fe-b4f5-a52de8033984","Type":"ContainerDied","Data":"1be1f1eb7f13040ffddb93fdb888ff9b387c46be744cd5ba0613e4acefae0187"} Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.999015 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-log-httpd\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.999119 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-combined-ca-bundle\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.999142 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-etc-swift\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.999182 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-config-data\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.999298 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-run-httpd\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.999321 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmxrr\" (UniqueName: \"kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-kube-api-access-bmxrr\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:17 crc kubenswrapper[4846]: I0202 13:43:17.999532 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-log-httpd\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:18 crc kubenswrapper[4846]: I0202 13:43:18.000052 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-run-httpd\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:18 crc kubenswrapper[4846]: I0202 13:43:18.004005 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-combined-ca-bundle\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:18 crc kubenswrapper[4846]: I0202 13:43:18.004954 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-config-data\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:18 crc kubenswrapper[4846]: I0202 13:43:18.011704 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-etc-swift\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:18 crc kubenswrapper[4846]: I0202 13:43:18.020181 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmxrr\" (UniqueName: \"kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-kube-api-access-bmxrr\") pod \"swift-proxy-7b4d5c8894-74dxm\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:18 crc kubenswrapper[4846]: I0202 13:43:18.135342 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:18 crc kubenswrapper[4846]: I0202 13:43:18.877649 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b4d5c8894-74dxm"] Feb 02 13:43:18 crc kubenswrapper[4846]: W0202 13:43:18.888782 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91dad3d1_0de8_44bb_ab9b_6671b119f709.slice/crio-cbf4ce53d9994b32c5235e5206bd9d8036a843c81ab6c3845dabd4c5f8be2542 WatchSource:0}: Error finding container cbf4ce53d9994b32c5235e5206bd9d8036a843c81ab6c3845dabd4c5f8be2542: Status 404 returned error can't find the container with id cbf4ce53d9994b32c5235e5206bd9d8036a843c81ab6c3845dabd4c5f8be2542 Feb 02 13:43:18 crc kubenswrapper[4846]: I0202 13:43:18.958446 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b4d5c8894-74dxm" event={"ID":"91dad3d1-0de8-44bb-ab9b-6671b119f709","Type":"ContainerStarted","Data":"cbf4ce53d9994b32c5235e5206bd9d8036a843c81ab6c3845dabd4c5f8be2542"} Feb 02 13:43:18 crc kubenswrapper[4846]: I0202 13:43:18.970887 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" event={"ID":"6dac67a9-bcff-44fe-b4f5-a52de8033984","Type":"ContainerStarted","Data":"674e23d0e99a48bd2b58cef4a2d38ab23dbf7ff2689d81a5938d2e82f97afa98"} Feb 02 13:43:18 crc kubenswrapper[4846]: I0202 13:43:18.971155 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:19 crc kubenswrapper[4846]: I0202 13:43:19.002583 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" podStartSLOduration=4.002566648 podStartE2EDuration="4.002566648s" podCreationTimestamp="2026-02-02 13:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:43:18.990889355 +0000 UTC m=+5630.219476218" watchObservedRunningTime="2026-02-02 13:43:19.002566648 +0000 UTC m=+5630.231153511" Feb 02 13:43:19 crc kubenswrapper[4846]: I0202 13:43:19.982890 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b4d5c8894-74dxm" event={"ID":"91dad3d1-0de8-44bb-ab9b-6671b119f709","Type":"ContainerStarted","Data":"bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f"} Feb 02 13:43:19 crc kubenswrapper[4846]: I0202 13:43:19.983254 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b4d5c8894-74dxm" event={"ID":"91dad3d1-0de8-44bb-ab9b-6671b119f709","Type":"ContainerStarted","Data":"d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897"} Feb 02 13:43:19 crc kubenswrapper[4846]: I0202 13:43:19.983310 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:19 crc kubenswrapper[4846]: I0202 13:43:19.983350 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.011833 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7b4d5c8894-74dxm" podStartSLOduration=3.011813408 podStartE2EDuration="3.011813408s" podCreationTimestamp="2026-02-02 13:43:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:43:20.005872674 +0000 UTC m=+5631.234459537" watchObservedRunningTime="2026-02-02 13:43:20.011813408 +0000 UTC m=+5631.240400271" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.260735 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6977cb8458-tl6r6"] Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.262601 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.265033 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.265360 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.284356 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6977cb8458-tl6r6"] Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.346681 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37cf7ba0-1c89-417e-958b-883454800ce9-log-httpd\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.347217 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-config-data\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.347353 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-combined-ca-bundle\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.347446 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-internal-tls-certs\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.347560 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-public-tls-certs\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.347667 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/37cf7ba0-1c89-417e-958b-883454800ce9-etc-swift\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.347770 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37cf7ba0-1c89-417e-958b-883454800ce9-run-httpd\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.347867 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6jvb\" (UniqueName: \"kubernetes.io/projected/37cf7ba0-1c89-417e-958b-883454800ce9-kube-api-access-s6jvb\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.449874 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-public-tls-certs\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.449967 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/37cf7ba0-1c89-417e-958b-883454800ce9-etc-swift\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.450038 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37cf7ba0-1c89-417e-958b-883454800ce9-run-httpd\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.450095 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6jvb\" (UniqueName: \"kubernetes.io/projected/37cf7ba0-1c89-417e-958b-883454800ce9-kube-api-access-s6jvb\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.450156 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37cf7ba0-1c89-417e-958b-883454800ce9-log-httpd\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.450187 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-config-data\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.450255 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-combined-ca-bundle\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.450287 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-internal-tls-certs\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.451080 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37cf7ba0-1c89-417e-958b-883454800ce9-run-httpd\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.451097 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37cf7ba0-1c89-417e-958b-883454800ce9-log-httpd\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.455769 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-combined-ca-bundle\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.457318 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/37cf7ba0-1c89-417e-958b-883454800ce9-etc-swift\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.461312 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-public-tls-certs\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.461399 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-internal-tls-certs\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.473613 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37cf7ba0-1c89-417e-958b-883454800ce9-config-data\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.482535 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6jvb\" (UniqueName: \"kubernetes.io/projected/37cf7ba0-1c89-417e-958b-883454800ce9-kube-api-access-s6jvb\") pod \"swift-proxy-6977cb8458-tl6r6\" (UID: \"37cf7ba0-1c89-417e-958b-883454800ce9\") " pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:20 crc kubenswrapper[4846]: I0202 13:43:20.579662 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:21 crc kubenswrapper[4846]: I0202 13:43:21.369275 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6977cb8458-tl6r6"] Feb 02 13:43:21 crc kubenswrapper[4846]: W0202 13:43:21.378139 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37cf7ba0_1c89_417e_958b_883454800ce9.slice/crio-2c32a4b89639581a3fc7a3a82cc0ce72c9f972a1c06082ddbce1dc204cb73541 WatchSource:0}: Error finding container 2c32a4b89639581a3fc7a3a82cc0ce72c9f972a1c06082ddbce1dc204cb73541: Status 404 returned error can't find the container with id 2c32a4b89639581a3fc7a3a82cc0ce72c9f972a1c06082ddbce1dc204cb73541 Feb 02 13:43:21 crc kubenswrapper[4846]: I0202 13:43:21.998282 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6977cb8458-tl6r6" event={"ID":"37cf7ba0-1c89-417e-958b-883454800ce9","Type":"ContainerStarted","Data":"e0d0b5006e660d9b91f8347d3fde80d5f806c0067b24f7e6636b57e8029e58bf"} Feb 02 13:43:21 crc kubenswrapper[4846]: I0202 13:43:21.998634 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6977cb8458-tl6r6" event={"ID":"37cf7ba0-1c89-417e-958b-883454800ce9","Type":"ContainerStarted","Data":"dbe49c37f6a3e8da579e00090ab7a0413f1960e2a0d5862602a8aa95611eaaf7"} Feb 02 13:43:21 crc kubenswrapper[4846]: I0202 13:43:21.998648 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6977cb8458-tl6r6" event={"ID":"37cf7ba0-1c89-417e-958b-883454800ce9","Type":"ContainerStarted","Data":"2c32a4b89639581a3fc7a3a82cc0ce72c9f972a1c06082ddbce1dc204cb73541"} Feb 02 13:43:21 crc kubenswrapper[4846]: I0202 13:43:21.998676 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:21 crc kubenswrapper[4846]: I0202 13:43:21.998694 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:22 crc kubenswrapper[4846]: I0202 13:43:22.002032 4846 generic.go:334] "Generic (PLEG): container finished" podID="f0c65201-a430-4889-a853-00c23cf8658e" containerID="a221103e1555cac3bb08f7cf5797a0654a4461f8628b0afdd6c76645c32bc7c5" exitCode=0 Feb 02 13:43:22 crc kubenswrapper[4846]: I0202 13:43:22.002320 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-grx5c" event={"ID":"f0c65201-a430-4889-a853-00c23cf8658e","Type":"ContainerDied","Data":"a221103e1555cac3bb08f7cf5797a0654a4461f8628b0afdd6c76645c32bc7c5"} Feb 02 13:43:22 crc kubenswrapper[4846]: I0202 13:43:22.021248 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6977cb8458-tl6r6" podStartSLOduration=2.021229758 podStartE2EDuration="2.021229758s" podCreationTimestamp="2026-02-02 13:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:43:22.019026655 +0000 UTC m=+5633.247613518" watchObservedRunningTime="2026-02-02 13:43:22.021229758 +0000 UTC m=+5633.249816621" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.345275 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.406647 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-ring-data-devices\") pod \"f0c65201-a430-4889-a853-00c23cf8658e\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.406765 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-scripts\") pod \"f0c65201-a430-4889-a853-00c23cf8658e\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.406797 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-swiftconf\") pod \"f0c65201-a430-4889-a853-00c23cf8658e\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.406821 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-combined-ca-bundle\") pod \"f0c65201-a430-4889-a853-00c23cf8658e\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.406850 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7wzc\" (UniqueName: \"kubernetes.io/projected/f0c65201-a430-4889-a853-00c23cf8658e-kube-api-access-v7wzc\") pod \"f0c65201-a430-4889-a853-00c23cf8658e\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.406928 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0c65201-a430-4889-a853-00c23cf8658e-etc-swift\") pod \"f0c65201-a430-4889-a853-00c23cf8658e\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.406952 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-dispersionconf\") pod \"f0c65201-a430-4889-a853-00c23cf8658e\" (UID: \"f0c65201-a430-4889-a853-00c23cf8658e\") " Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.411799 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f0c65201-a430-4889-a853-00c23cf8658e" (UID: "f0c65201-a430-4889-a853-00c23cf8658e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.413168 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0c65201-a430-4889-a853-00c23cf8658e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f0c65201-a430-4889-a853-00c23cf8658e" (UID: "f0c65201-a430-4889-a853-00c23cf8658e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.413492 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0c65201-a430-4889-a853-00c23cf8658e-kube-api-access-v7wzc" (OuterVolumeSpecName: "kube-api-access-v7wzc") pod "f0c65201-a430-4889-a853-00c23cf8658e" (UID: "f0c65201-a430-4889-a853-00c23cf8658e"). InnerVolumeSpecName "kube-api-access-v7wzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.417596 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f0c65201-a430-4889-a853-00c23cf8658e" (UID: "f0c65201-a430-4889-a853-00c23cf8658e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.431676 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-scripts" (OuterVolumeSpecName: "scripts") pod "f0c65201-a430-4889-a853-00c23cf8658e" (UID: "f0c65201-a430-4889-a853-00c23cf8658e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.435533 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0c65201-a430-4889-a853-00c23cf8658e" (UID: "f0c65201-a430-4889-a853-00c23cf8658e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.444769 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f0c65201-a430-4889-a853-00c23cf8658e" (UID: "f0c65201-a430-4889-a853-00c23cf8658e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.509803 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0c65201-a430-4889-a853-00c23cf8658e-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.509845 4846 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.509856 4846 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.509867 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0c65201-a430-4889-a853-00c23cf8658e-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.509877 4846 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.509888 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0c65201-a430-4889-a853-00c23cf8658e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:23 crc kubenswrapper[4846]: I0202 13:43:23.509898 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7wzc\" (UniqueName: \"kubernetes.io/projected/f0c65201-a430-4889-a853-00c23cf8658e-kube-api-access-v7wzc\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:24 crc kubenswrapper[4846]: I0202 13:43:24.022772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-grx5c" event={"ID":"f0c65201-a430-4889-a853-00c23cf8658e","Type":"ContainerDied","Data":"88c78c5f4101c3f85b6d33f94387a6aca2a40c8f3c963e972eabf0bf9cc0ac73"} Feb 02 13:43:24 crc kubenswrapper[4846]: I0202 13:43:24.022835 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88c78c5f4101c3f85b6d33f94387a6aca2a40c8f3c963e972eabf0bf9cc0ac73" Feb 02 13:43:24 crc kubenswrapper[4846]: I0202 13:43:24.022854 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-grx5c" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.135961 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.208637 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77d8d6f48f-vhhwk"] Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.208983 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" podUID="60adbe37-7588-48e4-8d77-3129ac673426" containerName="dnsmasq-dns" containerID="cri-o://bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b" gracePeriod=10 Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.705301 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.771833 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-config\") pod \"60adbe37-7588-48e4-8d77-3129ac673426\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.771950 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-dns-svc\") pod \"60adbe37-7588-48e4-8d77-3129ac673426\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.772073 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-nb\") pod \"60adbe37-7588-48e4-8d77-3129ac673426\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.772126 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqgfb\" (UniqueName: \"kubernetes.io/projected/60adbe37-7588-48e4-8d77-3129ac673426-kube-api-access-mqgfb\") pod \"60adbe37-7588-48e4-8d77-3129ac673426\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.772237 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-sb\") pod \"60adbe37-7588-48e4-8d77-3129ac673426\" (UID: \"60adbe37-7588-48e4-8d77-3129ac673426\") " Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.787899 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60adbe37-7588-48e4-8d77-3129ac673426-kube-api-access-mqgfb" (OuterVolumeSpecName: "kube-api-access-mqgfb") pod "60adbe37-7588-48e4-8d77-3129ac673426" (UID: "60adbe37-7588-48e4-8d77-3129ac673426"). InnerVolumeSpecName "kube-api-access-mqgfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.829197 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "60adbe37-7588-48e4-8d77-3129ac673426" (UID: "60adbe37-7588-48e4-8d77-3129ac673426"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.829747 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "60adbe37-7588-48e4-8d77-3129ac673426" (UID: "60adbe37-7588-48e4-8d77-3129ac673426"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.836345 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "60adbe37-7588-48e4-8d77-3129ac673426" (UID: "60adbe37-7588-48e4-8d77-3129ac673426"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.854283 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-config" (OuterVolumeSpecName: "config") pod "60adbe37-7588-48e4-8d77-3129ac673426" (UID: "60adbe37-7588-48e4-8d77-3129ac673426"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.875001 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.875045 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqgfb\" (UniqueName: \"kubernetes.io/projected/60adbe37-7588-48e4-8d77-3129ac673426-kube-api-access-mqgfb\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.875058 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.875073 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:26 crc kubenswrapper[4846]: I0202 13:43:26.875086 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60adbe37-7588-48e4-8d77-3129ac673426-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.045732 4846 generic.go:334] "Generic (PLEG): container finished" podID="60adbe37-7588-48e4-8d77-3129ac673426" containerID="bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b" exitCode=0 Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.045784 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.045787 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" event={"ID":"60adbe37-7588-48e4-8d77-3129ac673426","Type":"ContainerDied","Data":"bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b"} Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.045812 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d8d6f48f-vhhwk" event={"ID":"60adbe37-7588-48e4-8d77-3129ac673426","Type":"ContainerDied","Data":"946c65cd446c4db7692e20603573e38e1fd6f1419be22771887e6a6ebe04a7e2"} Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.045829 4846 scope.go:117] "RemoveContainer" containerID="bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b" Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.083702 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77d8d6f48f-vhhwk"] Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.096133 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77d8d6f48f-vhhwk"] Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.104838 4846 scope.go:117] "RemoveContainer" containerID="afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809" Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.132727 4846 scope.go:117] "RemoveContainer" containerID="bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b" Feb 02 13:43:27 crc kubenswrapper[4846]: E0202 13:43:27.133636 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b\": container with ID starting with bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b not found: ID does not exist" containerID="bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b" Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.133688 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b"} err="failed to get container status \"bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b\": rpc error: code = NotFound desc = could not find container \"bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b\": container with ID starting with bdb840bd7ff468d56304b16591af77395ae22582c5c9cce554df65ca01d8905b not found: ID does not exist" Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.133709 4846 scope.go:117] "RemoveContainer" containerID="afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809" Feb 02 13:43:27 crc kubenswrapper[4846]: E0202 13:43:27.134137 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809\": container with ID starting with afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809 not found: ID does not exist" containerID="afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809" Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.134177 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809"} err="failed to get container status \"afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809\": rpc error: code = NotFound desc = could not find container \"afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809\": container with ID starting with afe7c8375e6bc54f957cf874d19ba2902559c8224d558815bc5d4b745882e809 not found: ID does not exist" Feb 02 13:43:27 crc kubenswrapper[4846]: I0202 13:43:27.433468 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60adbe37-7588-48e4-8d77-3129ac673426" path="/var/lib/kubelet/pods/60adbe37-7588-48e4-8d77-3129ac673426/volumes" Feb 02 13:43:28 crc kubenswrapper[4846]: I0202 13:43:28.138493 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:28 crc kubenswrapper[4846]: I0202 13:43:28.139152 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:30 crc kubenswrapper[4846]: I0202 13:43:30.478907 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:43:30 crc kubenswrapper[4846]: I0202 13:43:30.479470 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:43:30 crc kubenswrapper[4846]: I0202 13:43:30.479514 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 13:43:30 crc kubenswrapper[4846]: I0202 13:43:30.488333 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 13:43:30 crc kubenswrapper[4846]: I0202 13:43:30.488573 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" gracePeriod=600 Feb 02 13:43:30 crc kubenswrapper[4846]: I0202 13:43:30.586895 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:30 crc kubenswrapper[4846]: I0202 13:43:30.588598 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6977cb8458-tl6r6" Feb 02 13:43:30 crc kubenswrapper[4846]: E0202 13:43:30.607877 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:43:30 crc kubenswrapper[4846]: I0202 13:43:30.672103 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7b4d5c8894-74dxm"] Feb 02 13:43:30 crc kubenswrapper[4846]: I0202 13:43:30.672327 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7b4d5c8894-74dxm" podUID="91dad3d1-0de8-44bb-ab9b-6671b119f709" containerName="proxy-httpd" containerID="cri-o://d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897" gracePeriod=30 Feb 02 13:43:30 crc kubenswrapper[4846]: I0202 13:43:30.675231 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7b4d5c8894-74dxm" podUID="91dad3d1-0de8-44bb-ab9b-6671b119f709" containerName="proxy-server" containerID="cri-o://bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f" gracePeriod=30 Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.088029 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" exitCode=0 Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.088101 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4"} Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.088512 4846 scope.go:117] "RemoveContainer" containerID="cc41b07cc0804d15d8ee13c8c57c6dfb00243755829e268a57694312efd7b440" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.089511 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:43:31 crc kubenswrapper[4846]: E0202 13:43:31.089845 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.094183 4846 generic.go:334] "Generic (PLEG): container finished" podID="91dad3d1-0de8-44bb-ab9b-6671b119f709" containerID="d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897" exitCode=0 Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.094228 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b4d5c8894-74dxm" event={"ID":"91dad3d1-0de8-44bb-ab9b-6671b119f709","Type":"ContainerDied","Data":"d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897"} Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.760012 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.883931 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-log-httpd\") pod \"91dad3d1-0de8-44bb-ab9b-6671b119f709\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.884019 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-run-httpd\") pod \"91dad3d1-0de8-44bb-ab9b-6671b119f709\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.884148 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-config-data\") pod \"91dad3d1-0de8-44bb-ab9b-6671b119f709\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.884202 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-etc-swift\") pod \"91dad3d1-0de8-44bb-ab9b-6671b119f709\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.884259 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmxrr\" (UniqueName: \"kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-kube-api-access-bmxrr\") pod \"91dad3d1-0de8-44bb-ab9b-6671b119f709\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.884322 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-combined-ca-bundle\") pod \"91dad3d1-0de8-44bb-ab9b-6671b119f709\" (UID: \"91dad3d1-0de8-44bb-ab9b-6671b119f709\") " Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.884676 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "91dad3d1-0de8-44bb-ab9b-6671b119f709" (UID: "91dad3d1-0de8-44bb-ab9b-6671b119f709"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.884892 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "91dad3d1-0de8-44bb-ab9b-6671b119f709" (UID: "91dad3d1-0de8-44bb-ab9b-6671b119f709"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.886534 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.886562 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/91dad3d1-0de8-44bb-ab9b-6671b119f709-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.893843 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-kube-api-access-bmxrr" (OuterVolumeSpecName: "kube-api-access-bmxrr") pod "91dad3d1-0de8-44bb-ab9b-6671b119f709" (UID: "91dad3d1-0de8-44bb-ab9b-6671b119f709"). InnerVolumeSpecName "kube-api-access-bmxrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.912696 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "91dad3d1-0de8-44bb-ab9b-6671b119f709" (UID: "91dad3d1-0de8-44bb-ab9b-6671b119f709"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.945021 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91dad3d1-0de8-44bb-ab9b-6671b119f709" (UID: "91dad3d1-0de8-44bb-ab9b-6671b119f709"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.946274 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-config-data" (OuterVolumeSpecName: "config-data") pod "91dad3d1-0de8-44bb-ab9b-6671b119f709" (UID: "91dad3d1-0de8-44bb-ab9b-6671b119f709"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.987622 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmxrr\" (UniqueName: \"kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-kube-api-access-bmxrr\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.987672 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.987681 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91dad3d1-0de8-44bb-ab9b-6671b119f709-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:31 crc kubenswrapper[4846]: I0202 13:43:31.987690 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/91dad3d1-0de8-44bb-ab9b-6671b119f709-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.108970 4846 generic.go:334] "Generic (PLEG): container finished" podID="91dad3d1-0de8-44bb-ab9b-6671b119f709" containerID="bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f" exitCode=0 Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.109019 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b4d5c8894-74dxm" event={"ID":"91dad3d1-0de8-44bb-ab9b-6671b119f709","Type":"ContainerDied","Data":"bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f"} Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.109044 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b4d5c8894-74dxm" event={"ID":"91dad3d1-0de8-44bb-ab9b-6671b119f709","Type":"ContainerDied","Data":"cbf4ce53d9994b32c5235e5206bd9d8036a843c81ab6c3845dabd4c5f8be2542"} Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.109063 4846 scope.go:117] "RemoveContainer" containerID="bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f" Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.109171 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b4d5c8894-74dxm" Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.146651 4846 scope.go:117] "RemoveContainer" containerID="d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897" Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.158286 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7b4d5c8894-74dxm"] Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.165316 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-7b4d5c8894-74dxm"] Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.169403 4846 scope.go:117] "RemoveContainer" containerID="bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f" Feb 02 13:43:32 crc kubenswrapper[4846]: E0202 13:43:32.169907 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f\": container with ID starting with bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f not found: ID does not exist" containerID="bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f" Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.169953 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f"} err="failed to get container status \"bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f\": rpc error: code = NotFound desc = could not find container \"bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f\": container with ID starting with bfc50d88462d9e2c34b7ef0689174bbf24de69ccd8c5e203ed674424c1fd584f not found: ID does not exist" Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.170003 4846 scope.go:117] "RemoveContainer" containerID="d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897" Feb 02 13:43:32 crc kubenswrapper[4846]: E0202 13:43:32.170515 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897\": container with ID starting with d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897 not found: ID does not exist" containerID="d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897" Feb 02 13:43:32 crc kubenswrapper[4846]: I0202 13:43:32.170545 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897"} err="failed to get container status \"d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897\": rpc error: code = NotFound desc = could not find container \"d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897\": container with ID starting with d35baf0f2cc7eb8118f3e7eb35303aafc0003f47d0967d03fca935b085c36897 not found: ID does not exist" Feb 02 13:43:33 crc kubenswrapper[4846]: I0202 13:43:33.434755 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91dad3d1-0de8-44bb-ab9b-6671b119f709" path="/var/lib/kubelet/pods/91dad3d1-0de8-44bb-ab9b-6671b119f709/volumes" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.464997 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-cgm6d"] Feb 02 13:43:36 crc kubenswrapper[4846]: E0202 13:43:36.469415 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60adbe37-7588-48e4-8d77-3129ac673426" containerName="dnsmasq-dns" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.469484 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="60adbe37-7588-48e4-8d77-3129ac673426" containerName="dnsmasq-dns" Feb 02 13:43:36 crc kubenswrapper[4846]: E0202 13:43:36.469502 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0c65201-a430-4889-a853-00c23cf8658e" containerName="swift-ring-rebalance" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.469511 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0c65201-a430-4889-a853-00c23cf8658e" containerName="swift-ring-rebalance" Feb 02 13:43:36 crc kubenswrapper[4846]: E0202 13:43:36.469537 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91dad3d1-0de8-44bb-ab9b-6671b119f709" containerName="proxy-httpd" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.469546 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="91dad3d1-0de8-44bb-ab9b-6671b119f709" containerName="proxy-httpd" Feb 02 13:43:36 crc kubenswrapper[4846]: E0202 13:43:36.469561 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91dad3d1-0de8-44bb-ab9b-6671b119f709" containerName="proxy-server" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.469578 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="91dad3d1-0de8-44bb-ab9b-6671b119f709" containerName="proxy-server" Feb 02 13:43:36 crc kubenswrapper[4846]: E0202 13:43:36.469595 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60adbe37-7588-48e4-8d77-3129ac673426" containerName="init" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.469602 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="60adbe37-7588-48e4-8d77-3129ac673426" containerName="init" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.469901 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="91dad3d1-0de8-44bb-ab9b-6671b119f709" containerName="proxy-httpd" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.469936 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="91dad3d1-0de8-44bb-ab9b-6671b119f709" containerName="proxy-server" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.469965 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0c65201-a430-4889-a853-00c23cf8658e" containerName="swift-ring-rebalance" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.469979 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="60adbe37-7588-48e4-8d77-3129ac673426" containerName="dnsmasq-dns" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.471061 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cgm6d" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.482509 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-cgm6d"] Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.554247 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5fd7-account-create-update-2k6hx"] Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.555695 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5fd7-account-create-update-2k6hx" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.561072 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.565462 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5fd7-account-create-update-2k6hx"] Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.671675 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb3025e-a013-471a-bf21-d28e12c189f1-operator-scripts\") pod \"cinder-db-create-cgm6d\" (UID: \"9eb3025e-a013-471a-bf21-d28e12c189f1\") " pod="openstack/cinder-db-create-cgm6d" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.672019 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85622\" (UniqueName: \"kubernetes.io/projected/9eb3025e-a013-471a-bf21-d28e12c189f1-kube-api-access-85622\") pod \"cinder-db-create-cgm6d\" (UID: \"9eb3025e-a013-471a-bf21-d28e12c189f1\") " pod="openstack/cinder-db-create-cgm6d" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.672125 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsq2d\" (UniqueName: \"kubernetes.io/projected/81aec00d-a484-44aa-a06d-57bc3a25c481-kube-api-access-jsq2d\") pod \"cinder-5fd7-account-create-update-2k6hx\" (UID: \"81aec00d-a484-44aa-a06d-57bc3a25c481\") " pod="openstack/cinder-5fd7-account-create-update-2k6hx" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.672166 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81aec00d-a484-44aa-a06d-57bc3a25c481-operator-scripts\") pod \"cinder-5fd7-account-create-update-2k6hx\" (UID: \"81aec00d-a484-44aa-a06d-57bc3a25c481\") " pod="openstack/cinder-5fd7-account-create-update-2k6hx" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.773265 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85622\" (UniqueName: \"kubernetes.io/projected/9eb3025e-a013-471a-bf21-d28e12c189f1-kube-api-access-85622\") pod \"cinder-db-create-cgm6d\" (UID: \"9eb3025e-a013-471a-bf21-d28e12c189f1\") " pod="openstack/cinder-db-create-cgm6d" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.773384 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsq2d\" (UniqueName: \"kubernetes.io/projected/81aec00d-a484-44aa-a06d-57bc3a25c481-kube-api-access-jsq2d\") pod \"cinder-5fd7-account-create-update-2k6hx\" (UID: \"81aec00d-a484-44aa-a06d-57bc3a25c481\") " pod="openstack/cinder-5fd7-account-create-update-2k6hx" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.773433 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81aec00d-a484-44aa-a06d-57bc3a25c481-operator-scripts\") pod \"cinder-5fd7-account-create-update-2k6hx\" (UID: \"81aec00d-a484-44aa-a06d-57bc3a25c481\") " pod="openstack/cinder-5fd7-account-create-update-2k6hx" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.773489 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb3025e-a013-471a-bf21-d28e12c189f1-operator-scripts\") pod \"cinder-db-create-cgm6d\" (UID: \"9eb3025e-a013-471a-bf21-d28e12c189f1\") " pod="openstack/cinder-db-create-cgm6d" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.774378 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81aec00d-a484-44aa-a06d-57bc3a25c481-operator-scripts\") pod \"cinder-5fd7-account-create-update-2k6hx\" (UID: \"81aec00d-a484-44aa-a06d-57bc3a25c481\") " pod="openstack/cinder-5fd7-account-create-update-2k6hx" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.774698 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb3025e-a013-471a-bf21-d28e12c189f1-operator-scripts\") pod \"cinder-db-create-cgm6d\" (UID: \"9eb3025e-a013-471a-bf21-d28e12c189f1\") " pod="openstack/cinder-db-create-cgm6d" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.792735 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85622\" (UniqueName: \"kubernetes.io/projected/9eb3025e-a013-471a-bf21-d28e12c189f1-kube-api-access-85622\") pod \"cinder-db-create-cgm6d\" (UID: \"9eb3025e-a013-471a-bf21-d28e12c189f1\") " pod="openstack/cinder-db-create-cgm6d" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.802907 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cgm6d" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.804845 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsq2d\" (UniqueName: \"kubernetes.io/projected/81aec00d-a484-44aa-a06d-57bc3a25c481-kube-api-access-jsq2d\") pod \"cinder-5fd7-account-create-update-2k6hx\" (UID: \"81aec00d-a484-44aa-a06d-57bc3a25c481\") " pod="openstack/cinder-5fd7-account-create-update-2k6hx" Feb 02 13:43:36 crc kubenswrapper[4846]: I0202 13:43:36.874191 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5fd7-account-create-update-2k6hx" Feb 02 13:43:37 crc kubenswrapper[4846]: W0202 13:43:37.272433 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9eb3025e_a013_471a_bf21_d28e12c189f1.slice/crio-8a3017def92c6550cef4de81ebf0650ac0e78625af8c809b35c22e1b890536eb WatchSource:0}: Error finding container 8a3017def92c6550cef4de81ebf0650ac0e78625af8c809b35c22e1b890536eb: Status 404 returned error can't find the container with id 8a3017def92c6550cef4de81ebf0650ac0e78625af8c809b35c22e1b890536eb Feb 02 13:43:37 crc kubenswrapper[4846]: I0202 13:43:37.273843 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-cgm6d"] Feb 02 13:43:37 crc kubenswrapper[4846]: W0202 13:43:37.395402 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81aec00d_a484_44aa_a06d_57bc3a25c481.slice/crio-e89be6b11b7cb339869fe15b2047c9bbc444d3855186657150ab66828c2a5b86 WatchSource:0}: Error finding container e89be6b11b7cb339869fe15b2047c9bbc444d3855186657150ab66828c2a5b86: Status 404 returned error can't find the container with id e89be6b11b7cb339869fe15b2047c9bbc444d3855186657150ab66828c2a5b86 Feb 02 13:43:37 crc kubenswrapper[4846]: I0202 13:43:37.396884 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5fd7-account-create-update-2k6hx"] Feb 02 13:43:38 crc kubenswrapper[4846]: I0202 13:43:38.161785 4846 generic.go:334] "Generic (PLEG): container finished" podID="81aec00d-a484-44aa-a06d-57bc3a25c481" containerID="3482dc67a33cb3f7e8b5da214b372820fb55bcf723a7053e78d52e2e5d2be8f8" exitCode=0 Feb 02 13:43:38 crc kubenswrapper[4846]: I0202 13:43:38.162155 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5fd7-account-create-update-2k6hx" event={"ID":"81aec00d-a484-44aa-a06d-57bc3a25c481","Type":"ContainerDied","Data":"3482dc67a33cb3f7e8b5da214b372820fb55bcf723a7053e78d52e2e5d2be8f8"} Feb 02 13:43:38 crc kubenswrapper[4846]: I0202 13:43:38.162229 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5fd7-account-create-update-2k6hx" event={"ID":"81aec00d-a484-44aa-a06d-57bc3a25c481","Type":"ContainerStarted","Data":"e89be6b11b7cb339869fe15b2047c9bbc444d3855186657150ab66828c2a5b86"} Feb 02 13:43:38 crc kubenswrapper[4846]: I0202 13:43:38.164240 4846 generic.go:334] "Generic (PLEG): container finished" podID="9eb3025e-a013-471a-bf21-d28e12c189f1" containerID="7b29da07bbab88e9e8a5d6b1cb8e79cc7b27f452040d70b52362efa8845f14a9" exitCode=0 Feb 02 13:43:38 crc kubenswrapper[4846]: I0202 13:43:38.164278 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cgm6d" event={"ID":"9eb3025e-a013-471a-bf21-d28e12c189f1","Type":"ContainerDied","Data":"7b29da07bbab88e9e8a5d6b1cb8e79cc7b27f452040d70b52362efa8845f14a9"} Feb 02 13:43:38 crc kubenswrapper[4846]: I0202 13:43:38.164304 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cgm6d" event={"ID":"9eb3025e-a013-471a-bf21-d28e12c189f1","Type":"ContainerStarted","Data":"8a3017def92c6550cef4de81ebf0650ac0e78625af8c809b35c22e1b890536eb"} Feb 02 13:43:38 crc kubenswrapper[4846]: I0202 13:43:38.464835 4846 scope.go:117] "RemoveContainer" containerID="90d90ae85de5986e2fa7d084c94a411a460ca65f71721f007924a4127ed2274c" Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.624595 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cgm6d" Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.634939 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5fd7-account-create-update-2k6hx" Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.737015 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb3025e-a013-471a-bf21-d28e12c189f1-operator-scripts\") pod \"9eb3025e-a013-471a-bf21-d28e12c189f1\" (UID: \"9eb3025e-a013-471a-bf21-d28e12c189f1\") " Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.737130 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85622\" (UniqueName: \"kubernetes.io/projected/9eb3025e-a013-471a-bf21-d28e12c189f1-kube-api-access-85622\") pod \"9eb3025e-a013-471a-bf21-d28e12c189f1\" (UID: \"9eb3025e-a013-471a-bf21-d28e12c189f1\") " Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.737206 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81aec00d-a484-44aa-a06d-57bc3a25c481-operator-scripts\") pod \"81aec00d-a484-44aa-a06d-57bc3a25c481\" (UID: \"81aec00d-a484-44aa-a06d-57bc3a25c481\") " Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.737262 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsq2d\" (UniqueName: \"kubernetes.io/projected/81aec00d-a484-44aa-a06d-57bc3a25c481-kube-api-access-jsq2d\") pod \"81aec00d-a484-44aa-a06d-57bc3a25c481\" (UID: \"81aec00d-a484-44aa-a06d-57bc3a25c481\") " Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.738202 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eb3025e-a013-471a-bf21-d28e12c189f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9eb3025e-a013-471a-bf21-d28e12c189f1" (UID: "9eb3025e-a013-471a-bf21-d28e12c189f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.738287 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81aec00d-a484-44aa-a06d-57bc3a25c481-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "81aec00d-a484-44aa-a06d-57bc3a25c481" (UID: "81aec00d-a484-44aa-a06d-57bc3a25c481"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.744130 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81aec00d-a484-44aa-a06d-57bc3a25c481-kube-api-access-jsq2d" (OuterVolumeSpecName: "kube-api-access-jsq2d") pod "81aec00d-a484-44aa-a06d-57bc3a25c481" (UID: "81aec00d-a484-44aa-a06d-57bc3a25c481"). InnerVolumeSpecName "kube-api-access-jsq2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.745464 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eb3025e-a013-471a-bf21-d28e12c189f1-kube-api-access-85622" (OuterVolumeSpecName: "kube-api-access-85622") pod "9eb3025e-a013-471a-bf21-d28e12c189f1" (UID: "9eb3025e-a013-471a-bf21-d28e12c189f1"). InnerVolumeSpecName "kube-api-access-85622". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.839606 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9eb3025e-a013-471a-bf21-d28e12c189f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.839955 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85622\" (UniqueName: \"kubernetes.io/projected/9eb3025e-a013-471a-bf21-d28e12c189f1-kube-api-access-85622\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.839968 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81aec00d-a484-44aa-a06d-57bc3a25c481-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:39 crc kubenswrapper[4846]: I0202 13:43:39.839977 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsq2d\" (UniqueName: \"kubernetes.io/projected/81aec00d-a484-44aa-a06d-57bc3a25c481-kube-api-access-jsq2d\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:40 crc kubenswrapper[4846]: I0202 13:43:40.184224 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cgm6d" Feb 02 13:43:40 crc kubenswrapper[4846]: I0202 13:43:40.184231 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cgm6d" event={"ID":"9eb3025e-a013-471a-bf21-d28e12c189f1","Type":"ContainerDied","Data":"8a3017def92c6550cef4de81ebf0650ac0e78625af8c809b35c22e1b890536eb"} Feb 02 13:43:40 crc kubenswrapper[4846]: I0202 13:43:40.184304 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a3017def92c6550cef4de81ebf0650ac0e78625af8c809b35c22e1b890536eb" Feb 02 13:43:40 crc kubenswrapper[4846]: I0202 13:43:40.186180 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5fd7-account-create-update-2k6hx" event={"ID":"81aec00d-a484-44aa-a06d-57bc3a25c481","Type":"ContainerDied","Data":"e89be6b11b7cb339869fe15b2047c9bbc444d3855186657150ab66828c2a5b86"} Feb 02 13:43:40 crc kubenswrapper[4846]: I0202 13:43:40.186215 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e89be6b11b7cb339869fe15b2047c9bbc444d3855186657150ab66828c2a5b86" Feb 02 13:43:40 crc kubenswrapper[4846]: I0202 13:43:40.186326 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5fd7-account-create-update-2k6hx" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.424488 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:43:41 crc kubenswrapper[4846]: E0202 13:43:41.424862 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.746193 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-wxpwk"] Feb 02 13:43:41 crc kubenswrapper[4846]: E0202 13:43:41.746686 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb3025e-a013-471a-bf21-d28e12c189f1" containerName="mariadb-database-create" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.746720 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb3025e-a013-471a-bf21-d28e12c189f1" containerName="mariadb-database-create" Feb 02 13:43:41 crc kubenswrapper[4846]: E0202 13:43:41.746746 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81aec00d-a484-44aa-a06d-57bc3a25c481" containerName="mariadb-account-create-update" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.746752 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="81aec00d-a484-44aa-a06d-57bc3a25c481" containerName="mariadb-account-create-update" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.746931 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eb3025e-a013-471a-bf21-d28e12c189f1" containerName="mariadb-database-create" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.746968 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="81aec00d-a484-44aa-a06d-57bc3a25c481" containerName="mariadb-account-create-update" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.747553 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.750042 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2mrs5" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.750686 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.750964 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.766427 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-wxpwk"] Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.883516 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-combined-ca-bundle\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.883590 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-db-sync-config-data\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.883640 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5a1aebe-43c7-4791-a2ff-b81a10230d01-etc-machine-id\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.883668 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsf8v\" (UniqueName: \"kubernetes.io/projected/a5a1aebe-43c7-4791-a2ff-b81a10230d01-kube-api-access-vsf8v\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.883732 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-scripts\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.883748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-config-data\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.985699 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-db-sync-config-data\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.985783 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5a1aebe-43c7-4791-a2ff-b81a10230d01-etc-machine-id\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.985805 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsf8v\" (UniqueName: \"kubernetes.io/projected/a5a1aebe-43c7-4791-a2ff-b81a10230d01-kube-api-access-vsf8v\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.985883 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-scripts\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.985905 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-config-data\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.985997 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-combined-ca-bundle\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.986565 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5a1aebe-43c7-4791-a2ff-b81a10230d01-etc-machine-id\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.990986 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-combined-ca-bundle\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.991366 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-config-data\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.996646 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-scripts\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:41 crc kubenswrapper[4846]: I0202 13:43:41.996744 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-db-sync-config-data\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:42 crc kubenswrapper[4846]: I0202 13:43:42.010190 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsf8v\" (UniqueName: \"kubernetes.io/projected/a5a1aebe-43c7-4791-a2ff-b81a10230d01-kube-api-access-vsf8v\") pod \"cinder-db-sync-wxpwk\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:42 crc kubenswrapper[4846]: I0202 13:43:42.104189 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:42 crc kubenswrapper[4846]: I0202 13:43:42.534496 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-wxpwk"] Feb 02 13:43:43 crc kubenswrapper[4846]: I0202 13:43:43.226159 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wxpwk" event={"ID":"a5a1aebe-43c7-4791-a2ff-b81a10230d01","Type":"ContainerStarted","Data":"4197896df10ef7e7b0b43c79de5dc6cb20f0090a9950f89e32264e20708bb932"} Feb 02 13:43:43 crc kubenswrapper[4846]: I0202 13:43:43.226506 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wxpwk" event={"ID":"a5a1aebe-43c7-4791-a2ff-b81a10230d01","Type":"ContainerStarted","Data":"e8ff3e66b5999a5f7a34030430b27a5da2137e7108da0c771f8286e20cf00725"} Feb 02 13:43:43 crc kubenswrapper[4846]: I0202 13:43:43.245542 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-wxpwk" podStartSLOduration=2.245520196 podStartE2EDuration="2.245520196s" podCreationTimestamp="2026-02-02 13:43:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:43:43.241440217 +0000 UTC m=+5654.470027080" watchObservedRunningTime="2026-02-02 13:43:43.245520196 +0000 UTC m=+5654.474107069" Feb 02 13:43:46 crc kubenswrapper[4846]: I0202 13:43:46.249691 4846 generic.go:334] "Generic (PLEG): container finished" podID="a5a1aebe-43c7-4791-a2ff-b81a10230d01" containerID="4197896df10ef7e7b0b43c79de5dc6cb20f0090a9950f89e32264e20708bb932" exitCode=0 Feb 02 13:43:46 crc kubenswrapper[4846]: I0202 13:43:46.249766 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wxpwk" event={"ID":"a5a1aebe-43c7-4791-a2ff-b81a10230d01","Type":"ContainerDied","Data":"4197896df10ef7e7b0b43c79de5dc6cb20f0090a9950f89e32264e20708bb932"} Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.588413 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.690493 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5a1aebe-43c7-4791-a2ff-b81a10230d01-etc-machine-id\") pod \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.690586 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-db-sync-config-data\") pod \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.690614 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5a1aebe-43c7-4791-a2ff-b81a10230d01-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a5a1aebe-43c7-4791-a2ff-b81a10230d01" (UID: "a5a1aebe-43c7-4791-a2ff-b81a10230d01"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.690647 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-config-data\") pod \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.690918 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-scripts\") pod \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.690957 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsf8v\" (UniqueName: \"kubernetes.io/projected/a5a1aebe-43c7-4791-a2ff-b81a10230d01-kube-api-access-vsf8v\") pod \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.690996 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-combined-ca-bundle\") pod \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\" (UID: \"a5a1aebe-43c7-4791-a2ff-b81a10230d01\") " Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.691420 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5a1aebe-43c7-4791-a2ff-b81a10230d01-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.696553 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a5a1aebe-43c7-4791-a2ff-b81a10230d01" (UID: "a5a1aebe-43c7-4791-a2ff-b81a10230d01"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.696757 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a1aebe-43c7-4791-a2ff-b81a10230d01-kube-api-access-vsf8v" (OuterVolumeSpecName: "kube-api-access-vsf8v") pod "a5a1aebe-43c7-4791-a2ff-b81a10230d01" (UID: "a5a1aebe-43c7-4791-a2ff-b81a10230d01"). InnerVolumeSpecName "kube-api-access-vsf8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.697087 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-scripts" (OuterVolumeSpecName: "scripts") pod "a5a1aebe-43c7-4791-a2ff-b81a10230d01" (UID: "a5a1aebe-43c7-4791-a2ff-b81a10230d01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.715264 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5a1aebe-43c7-4791-a2ff-b81a10230d01" (UID: "a5a1aebe-43c7-4791-a2ff-b81a10230d01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.734149 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-config-data" (OuterVolumeSpecName: "config-data") pod "a5a1aebe-43c7-4791-a2ff-b81a10230d01" (UID: "a5a1aebe-43c7-4791-a2ff-b81a10230d01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.793533 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.793570 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.793581 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.793592 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsf8v\" (UniqueName: \"kubernetes.io/projected/a5a1aebe-43c7-4791-a2ff-b81a10230d01-kube-api-access-vsf8v\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:47 crc kubenswrapper[4846]: I0202 13:43:47.793643 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a1aebe-43c7-4791-a2ff-b81a10230d01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.267483 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-wxpwk" event={"ID":"a5a1aebe-43c7-4791-a2ff-b81a10230d01","Type":"ContainerDied","Data":"e8ff3e66b5999a5f7a34030430b27a5da2137e7108da0c771f8286e20cf00725"} Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.267901 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8ff3e66b5999a5f7a34030430b27a5da2137e7108da0c771f8286e20cf00725" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.267813 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-wxpwk" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.537060 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58c54f7d8f-kscsh"] Feb 02 13:43:48 crc kubenswrapper[4846]: E0202 13:43:48.537567 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a1aebe-43c7-4791-a2ff-b81a10230d01" containerName="cinder-db-sync" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.537592 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a1aebe-43c7-4791-a2ff-b81a10230d01" containerName="cinder-db-sync" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.537802 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a1aebe-43c7-4791-a2ff-b81a10230d01" containerName="cinder-db-sync" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.539344 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.549750 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58c54f7d8f-kscsh"] Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.607405 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br5t6\" (UniqueName: \"kubernetes.io/projected/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-kube-api-access-br5t6\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.607564 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-sb\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.607596 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-config\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.607744 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-nb\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.607830 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-dns-svc\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.720401 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-sb\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.721445 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-config\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.721527 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-nb\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.721589 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-dns-svc\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.721736 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br5t6\" (UniqueName: \"kubernetes.io/projected/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-kube-api-access-br5t6\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.721399 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-sb\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.722683 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-config\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.723263 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-nb\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.723855 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-dns-svc\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.782747 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br5t6\" (UniqueName: \"kubernetes.io/projected/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-kube-api-access-br5t6\") pod \"dnsmasq-dns-58c54f7d8f-kscsh\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.807712 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.809665 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.822010 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.822266 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2mrs5" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.825746 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.825922 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.872695 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.881228 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.928509 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data-custom\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.928773 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpd5m\" (UniqueName: \"kubernetes.io/projected/00124999-6b8a-42b0-a831-b14a82307e21-kube-api-access-zpd5m\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.928791 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-scripts\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.928822 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.928847 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/00124999-6b8a-42b0-a831-b14a82307e21-etc-machine-id\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.928885 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00124999-6b8a-42b0-a831-b14a82307e21-logs\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:48 crc kubenswrapper[4846]: I0202 13:43:48.928958 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.030081 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.030145 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/00124999-6b8a-42b0-a831-b14a82307e21-etc-machine-id\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.030193 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00124999-6b8a-42b0-a831-b14a82307e21-logs\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.030311 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.030368 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data-custom\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.030396 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpd5m\" (UniqueName: \"kubernetes.io/projected/00124999-6b8a-42b0-a831-b14a82307e21-kube-api-access-zpd5m\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.030419 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-scripts\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.031096 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00124999-6b8a-42b0-a831-b14a82307e21-logs\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.031586 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/00124999-6b8a-42b0-a831-b14a82307e21-etc-machine-id\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.034260 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-scripts\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.036857 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.036981 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.037442 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data-custom\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.052786 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpd5m\" (UniqueName: \"kubernetes.io/projected/00124999-6b8a-42b0-a831-b14a82307e21-kube-api-access-zpd5m\") pod \"cinder-api-0\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.139530 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.625656 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58c54f7d8f-kscsh"] Feb 02 13:43:49 crc kubenswrapper[4846]: I0202 13:43:49.810268 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:43:49 crc kubenswrapper[4846]: W0202 13:43:49.823698 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00124999_6b8a_42b0_a831_b14a82307e21.slice/crio-04ede0a097807d80c98cef12066fcf39005818fbeb47b5144f974f6c2487fe97 WatchSource:0}: Error finding container 04ede0a097807d80c98cef12066fcf39005818fbeb47b5144f974f6c2487fe97: Status 404 returned error can't find the container with id 04ede0a097807d80c98cef12066fcf39005818fbeb47b5144f974f6c2487fe97 Feb 02 13:43:50 crc kubenswrapper[4846]: I0202 13:43:50.294937 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"00124999-6b8a-42b0-a831-b14a82307e21","Type":"ContainerStarted","Data":"04ede0a097807d80c98cef12066fcf39005818fbeb47b5144f974f6c2487fe97"} Feb 02 13:43:50 crc kubenswrapper[4846]: I0202 13:43:50.309114 4846 generic.go:334] "Generic (PLEG): container finished" podID="efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" containerID="55cd3b9adbf7d2dc7c46d93d3ea7a77904d06edac6918a9fe761ef3b6d533465" exitCode=0 Feb 02 13:43:50 crc kubenswrapper[4846]: I0202 13:43:50.309168 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" event={"ID":"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d","Type":"ContainerDied","Data":"55cd3b9adbf7d2dc7c46d93d3ea7a77904d06edac6918a9fe761ef3b6d533465"} Feb 02 13:43:50 crc kubenswrapper[4846]: I0202 13:43:50.309200 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" event={"ID":"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d","Type":"ContainerStarted","Data":"7206bc7ec0510f8f6f925298363f0d21eeb8986a586499865a719a5204003b09"} Feb 02 13:43:51 crc kubenswrapper[4846]: I0202 13:43:51.318395 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"00124999-6b8a-42b0-a831-b14a82307e21","Type":"ContainerStarted","Data":"d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c"} Feb 02 13:43:51 crc kubenswrapper[4846]: I0202 13:43:51.318889 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"00124999-6b8a-42b0-a831-b14a82307e21","Type":"ContainerStarted","Data":"839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87"} Feb 02 13:43:51 crc kubenswrapper[4846]: I0202 13:43:51.318912 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 02 13:43:51 crc kubenswrapper[4846]: I0202 13:43:51.321135 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" event={"ID":"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d","Type":"ContainerStarted","Data":"23fa90dc256fc9e685fbd9ec837bc4860fec55e3efe78e2ea4d3c97dc6bf3559"} Feb 02 13:43:51 crc kubenswrapper[4846]: I0202 13:43:51.321282 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:51 crc kubenswrapper[4846]: I0202 13:43:51.338936 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.3389144379999998 podStartE2EDuration="3.338914438s" podCreationTimestamp="2026-02-02 13:43:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:43:51.336808988 +0000 UTC m=+5662.565395871" watchObservedRunningTime="2026-02-02 13:43:51.338914438 +0000 UTC m=+5662.567501311" Feb 02 13:43:51 crc kubenswrapper[4846]: I0202 13:43:51.366782 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" podStartSLOduration=3.366759515 podStartE2EDuration="3.366759515s" podCreationTimestamp="2026-02-02 13:43:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:43:51.36034837 +0000 UTC m=+5662.588935243" watchObservedRunningTime="2026-02-02 13:43:51.366759515 +0000 UTC m=+5662.595346378" Feb 02 13:43:51 crc kubenswrapper[4846]: I0202 13:43:51.488140 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:43:53 crc kubenswrapper[4846]: I0202 13:43:53.340086 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="00124999-6b8a-42b0-a831-b14a82307e21" containerName="cinder-api-log" containerID="cri-o://839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87" gracePeriod=30 Feb 02 13:43:53 crc kubenswrapper[4846]: I0202 13:43:53.340169 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="00124999-6b8a-42b0-a831-b14a82307e21" containerName="cinder-api" containerID="cri-o://d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c" gracePeriod=30 Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.005268 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.140855 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-scripts\") pod \"00124999-6b8a-42b0-a831-b14a82307e21\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.141007 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data-custom\") pod \"00124999-6b8a-42b0-a831-b14a82307e21\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.141170 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data\") pod \"00124999-6b8a-42b0-a831-b14a82307e21\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.141199 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/00124999-6b8a-42b0-a831-b14a82307e21-etc-machine-id\") pod \"00124999-6b8a-42b0-a831-b14a82307e21\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.142080 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00124999-6b8a-42b0-a831-b14a82307e21-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "00124999-6b8a-42b0-a831-b14a82307e21" (UID: "00124999-6b8a-42b0-a831-b14a82307e21"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.142330 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-combined-ca-bundle\") pod \"00124999-6b8a-42b0-a831-b14a82307e21\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.142746 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00124999-6b8a-42b0-a831-b14a82307e21-logs\") pod \"00124999-6b8a-42b0-a831-b14a82307e21\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.143138 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00124999-6b8a-42b0-a831-b14a82307e21-logs" (OuterVolumeSpecName: "logs") pod "00124999-6b8a-42b0-a831-b14a82307e21" (UID: "00124999-6b8a-42b0-a831-b14a82307e21"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.143464 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpd5m\" (UniqueName: \"kubernetes.io/projected/00124999-6b8a-42b0-a831-b14a82307e21-kube-api-access-zpd5m\") pod \"00124999-6b8a-42b0-a831-b14a82307e21\" (UID: \"00124999-6b8a-42b0-a831-b14a82307e21\") " Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.144809 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/00124999-6b8a-42b0-a831-b14a82307e21-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.144846 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00124999-6b8a-42b0-a831-b14a82307e21-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.148775 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "00124999-6b8a-42b0-a831-b14a82307e21" (UID: "00124999-6b8a-42b0-a831-b14a82307e21"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.156476 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-scripts" (OuterVolumeSpecName: "scripts") pod "00124999-6b8a-42b0-a831-b14a82307e21" (UID: "00124999-6b8a-42b0-a831-b14a82307e21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.156750 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00124999-6b8a-42b0-a831-b14a82307e21-kube-api-access-zpd5m" (OuterVolumeSpecName: "kube-api-access-zpd5m") pod "00124999-6b8a-42b0-a831-b14a82307e21" (UID: "00124999-6b8a-42b0-a831-b14a82307e21"). InnerVolumeSpecName "kube-api-access-zpd5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.172957 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00124999-6b8a-42b0-a831-b14a82307e21" (UID: "00124999-6b8a-42b0-a831-b14a82307e21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.196032 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data" (OuterVolumeSpecName: "config-data") pod "00124999-6b8a-42b0-a831-b14a82307e21" (UID: "00124999-6b8a-42b0-a831-b14a82307e21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.247116 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.247165 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.247184 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpd5m\" (UniqueName: \"kubernetes.io/projected/00124999-6b8a-42b0-a831-b14a82307e21-kube-api-access-zpd5m\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.247197 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.247207 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00124999-6b8a-42b0-a831-b14a82307e21-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.349398 4846 generic.go:334] "Generic (PLEG): container finished" podID="00124999-6b8a-42b0-a831-b14a82307e21" containerID="d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c" exitCode=0 Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.349438 4846 generic.go:334] "Generic (PLEG): container finished" podID="00124999-6b8a-42b0-a831-b14a82307e21" containerID="839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87" exitCode=143 Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.349457 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"00124999-6b8a-42b0-a831-b14a82307e21","Type":"ContainerDied","Data":"d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c"} Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.349528 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"00124999-6b8a-42b0-a831-b14a82307e21","Type":"ContainerDied","Data":"839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87"} Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.349545 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"00124999-6b8a-42b0-a831-b14a82307e21","Type":"ContainerDied","Data":"04ede0a097807d80c98cef12066fcf39005818fbeb47b5144f974f6c2487fe97"} Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.349566 4846 scope.go:117] "RemoveContainer" containerID="d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.349460 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.371987 4846 scope.go:117] "RemoveContainer" containerID="839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.385713 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.394653 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.405020 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:43:54 crc kubenswrapper[4846]: E0202 13:43:54.405378 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00124999-6b8a-42b0-a831-b14a82307e21" containerName="cinder-api-log" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.405397 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="00124999-6b8a-42b0-a831-b14a82307e21" containerName="cinder-api-log" Feb 02 13:43:54 crc kubenswrapper[4846]: E0202 13:43:54.405421 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00124999-6b8a-42b0-a831-b14a82307e21" containerName="cinder-api" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.405427 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="00124999-6b8a-42b0-a831-b14a82307e21" containerName="cinder-api" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.405614 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="00124999-6b8a-42b0-a831-b14a82307e21" containerName="cinder-api-log" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.405642 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="00124999-6b8a-42b0-a831-b14a82307e21" containerName="cinder-api" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.406591 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.410415 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.411314 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.411493 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2mrs5" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.411793 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.412089 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.412251 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.415388 4846 scope.go:117] "RemoveContainer" containerID="d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c" Feb 02 13:43:54 crc kubenswrapper[4846]: E0202 13:43:54.417604 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c\": container with ID starting with d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c not found: ID does not exist" containerID="d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.417721 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c"} err="failed to get container status \"d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c\": rpc error: code = NotFound desc = could not find container \"d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c\": container with ID starting with d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c not found: ID does not exist" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.417749 4846 scope.go:117] "RemoveContainer" containerID="839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87" Feb 02 13:43:54 crc kubenswrapper[4846]: E0202 13:43:54.418448 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87\": container with ID starting with 839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87 not found: ID does not exist" containerID="839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.418477 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87"} err="failed to get container status \"839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87\": rpc error: code = NotFound desc = could not find container \"839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87\": container with ID starting with 839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87 not found: ID does not exist" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.418499 4846 scope.go:117] "RemoveContainer" containerID="d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.419097 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c"} err="failed to get container status \"d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c\": rpc error: code = NotFound desc = could not find container \"d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c\": container with ID starting with d01a7b927497988f66916a7d312f5b92898c7ee417cd48557ed3064ddbe0127c not found: ID does not exist" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.419153 4846 scope.go:117] "RemoveContainer" containerID="839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.422308 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.425089 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87"} err="failed to get container status \"839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87\": rpc error: code = NotFound desc = could not find container \"839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87\": container with ID starting with 839ed82da5540b2a17a4b43ce0993cb6a0993bb9530febbda03bddaaa72f0c87 not found: ID does not exist" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.551458 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5964a748-1a08-4359-b050-45153f45b3b7-logs\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.551507 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvkks\" (UniqueName: \"kubernetes.io/projected/5964a748-1a08-4359-b050-45153f45b3b7-kube-api-access-pvkks\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.551529 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.551565 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-scripts\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.551593 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.551693 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5964a748-1a08-4359-b050-45153f45b3b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.551748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.551921 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.552015 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.653778 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.654361 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5964a748-1a08-4359-b050-45153f45b3b7-logs\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.654617 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvkks\" (UniqueName: \"kubernetes.io/projected/5964a748-1a08-4359-b050-45153f45b3b7-kube-api-access-pvkks\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.654889 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.654781 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5964a748-1a08-4359-b050-45153f45b3b7-logs\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.655390 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-scripts\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.655705 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.655964 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5964a748-1a08-4359-b050-45153f45b3b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.656196 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.656082 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5964a748-1a08-4359-b050-45153f45b3b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.656429 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.658033 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.658077 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.658827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-scripts\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.659671 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.659997 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.660358 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.679512 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvkks\" (UniqueName: \"kubernetes.io/projected/5964a748-1a08-4359-b050-45153f45b3b7-kube-api-access-pvkks\") pod \"cinder-api-0\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " pod="openstack/cinder-api-0" Feb 02 13:43:54 crc kubenswrapper[4846]: I0202 13:43:54.731141 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 13:43:55 crc kubenswrapper[4846]: I0202 13:43:55.407956 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:43:55 crc kubenswrapper[4846]: W0202 13:43:55.411642 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5964a748_1a08_4359_b050_45153f45b3b7.slice/crio-a825103d3d837cb44d6841b086e1d529433037d6a820887c2e42aa15a6490d19 WatchSource:0}: Error finding container a825103d3d837cb44d6841b086e1d529433037d6a820887c2e42aa15a6490d19: Status 404 returned error can't find the container with id a825103d3d837cb44d6841b086e1d529433037d6a820887c2e42aa15a6490d19 Feb 02 13:43:55 crc kubenswrapper[4846]: I0202 13:43:55.424215 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:43:55 crc kubenswrapper[4846]: E0202 13:43:55.424930 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:43:55 crc kubenswrapper[4846]: I0202 13:43:55.434256 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00124999-6b8a-42b0-a831-b14a82307e21" path="/var/lib/kubelet/pods/00124999-6b8a-42b0-a831-b14a82307e21/volumes" Feb 02 13:43:56 crc kubenswrapper[4846]: I0202 13:43:56.368714 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5964a748-1a08-4359-b050-45153f45b3b7","Type":"ContainerStarted","Data":"5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335"} Feb 02 13:43:56 crc kubenswrapper[4846]: I0202 13:43:56.369256 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5964a748-1a08-4359-b050-45153f45b3b7","Type":"ContainerStarted","Data":"a825103d3d837cb44d6841b086e1d529433037d6a820887c2e42aa15a6490d19"} Feb 02 13:43:57 crc kubenswrapper[4846]: I0202 13:43:57.378503 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5964a748-1a08-4359-b050-45153f45b3b7","Type":"ContainerStarted","Data":"99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f"} Feb 02 13:43:57 crc kubenswrapper[4846]: I0202 13:43:57.378676 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 02 13:43:57 crc kubenswrapper[4846]: I0202 13:43:57.404149 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.404126256 podStartE2EDuration="3.404126256s" podCreationTimestamp="2026-02-02 13:43:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:43:57.401516453 +0000 UTC m=+5668.630103326" watchObservedRunningTime="2026-02-02 13:43:57.404126256 +0000 UTC m=+5668.632713119" Feb 02 13:43:58 crc kubenswrapper[4846]: I0202 13:43:58.883407 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:43:58 crc kubenswrapper[4846]: I0202 13:43:58.950121 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f87f9d645-qnvj7"] Feb 02 13:43:58 crc kubenswrapper[4846]: I0202 13:43:58.950409 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" podUID="6dac67a9-bcff-44fe-b4f5-a52de8033984" containerName="dnsmasq-dns" containerID="cri-o://674e23d0e99a48bd2b58cef4a2d38ab23dbf7ff2689d81a5938d2e82f97afa98" gracePeriod=10 Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.398440 4846 generic.go:334] "Generic (PLEG): container finished" podID="6dac67a9-bcff-44fe-b4f5-a52de8033984" containerID="674e23d0e99a48bd2b58cef4a2d38ab23dbf7ff2689d81a5938d2e82f97afa98" exitCode=0 Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.398486 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" event={"ID":"6dac67a9-bcff-44fe-b4f5-a52de8033984","Type":"ContainerDied","Data":"674e23d0e99a48bd2b58cef4a2d38ab23dbf7ff2689d81a5938d2e82f97afa98"} Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.398835 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" event={"ID":"6dac67a9-bcff-44fe-b4f5-a52de8033984","Type":"ContainerDied","Data":"6c4dec8ae6cf287dbdb5b11299053b2dc14b9cd124497a4b481d796cc7d4a4a4"} Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.398851 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c4dec8ae6cf287dbdb5b11299053b2dc14b9cd124497a4b481d796cc7d4a4a4" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.463083 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.654496 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-dns-svc\") pod \"6dac67a9-bcff-44fe-b4f5-a52de8033984\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.654632 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-config\") pod \"6dac67a9-bcff-44fe-b4f5-a52de8033984\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.654657 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-nb\") pod \"6dac67a9-bcff-44fe-b4f5-a52de8033984\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.654823 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-sb\") pod \"6dac67a9-bcff-44fe-b4f5-a52de8033984\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.654922 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pthqx\" (UniqueName: \"kubernetes.io/projected/6dac67a9-bcff-44fe-b4f5-a52de8033984-kube-api-access-pthqx\") pod \"6dac67a9-bcff-44fe-b4f5-a52de8033984\" (UID: \"6dac67a9-bcff-44fe-b4f5-a52de8033984\") " Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.660355 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dac67a9-bcff-44fe-b4f5-a52de8033984-kube-api-access-pthqx" (OuterVolumeSpecName: "kube-api-access-pthqx") pod "6dac67a9-bcff-44fe-b4f5-a52de8033984" (UID: "6dac67a9-bcff-44fe-b4f5-a52de8033984"). InnerVolumeSpecName "kube-api-access-pthqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.700314 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-config" (OuterVolumeSpecName: "config") pod "6dac67a9-bcff-44fe-b4f5-a52de8033984" (UID: "6dac67a9-bcff-44fe-b4f5-a52de8033984"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.701120 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6dac67a9-bcff-44fe-b4f5-a52de8033984" (UID: "6dac67a9-bcff-44fe-b4f5-a52de8033984"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.703301 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6dac67a9-bcff-44fe-b4f5-a52de8033984" (UID: "6dac67a9-bcff-44fe-b4f5-a52de8033984"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.721907 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6dac67a9-bcff-44fe-b4f5-a52de8033984" (UID: "6dac67a9-bcff-44fe-b4f5-a52de8033984"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.756557 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.756592 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pthqx\" (UniqueName: \"kubernetes.io/projected/6dac67a9-bcff-44fe-b4f5-a52de8033984-kube-api-access-pthqx\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.756632 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.756643 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:43:59 crc kubenswrapper[4846]: I0202 13:43:59.756652 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6dac67a9-bcff-44fe-b4f5-a52de8033984-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:00 crc kubenswrapper[4846]: I0202 13:44:00.407314 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f87f9d645-qnvj7" Feb 02 13:44:00 crc kubenswrapper[4846]: I0202 13:44:00.442108 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f87f9d645-qnvj7"] Feb 02 13:44:00 crc kubenswrapper[4846]: I0202 13:44:00.451297 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f87f9d645-qnvj7"] Feb 02 13:44:01 crc kubenswrapper[4846]: I0202 13:44:01.440950 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dac67a9-bcff-44fe-b4f5-a52de8033984" path="/var/lib/kubelet/pods/6dac67a9-bcff-44fe-b4f5-a52de8033984/volumes" Feb 02 13:44:06 crc kubenswrapper[4846]: I0202 13:44:06.423654 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:44:06 crc kubenswrapper[4846]: E0202 13:44:06.424555 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:44:06 crc kubenswrapper[4846]: I0202 13:44:06.619091 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 02 13:44:19 crc kubenswrapper[4846]: I0202 13:44:19.428726 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:44:19 crc kubenswrapper[4846]: E0202 13:44:19.429546 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.291295 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 13:44:22 crc kubenswrapper[4846]: E0202 13:44:22.292042 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dac67a9-bcff-44fe-b4f5-a52de8033984" containerName="init" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.292059 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dac67a9-bcff-44fe-b4f5-a52de8033984" containerName="init" Feb 02 13:44:22 crc kubenswrapper[4846]: E0202 13:44:22.292072 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dac67a9-bcff-44fe-b4f5-a52de8033984" containerName="dnsmasq-dns" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.292079 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dac67a9-bcff-44fe-b4f5-a52de8033984" containerName="dnsmasq-dns" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.292305 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dac67a9-bcff-44fe-b4f5-a52de8033984" containerName="dnsmasq-dns" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.293425 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.295138 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.307239 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.385227 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.385278 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjmjk\" (UniqueName: \"kubernetes.io/projected/b908dc5f-4777-40d9-ab29-35c36ba0b72a-kube-api-access-zjmjk\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.385345 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.385424 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-scripts\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.385475 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.385737 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b908dc5f-4777-40d9-ab29-35c36ba0b72a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.487112 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b908dc5f-4777-40d9-ab29-35c36ba0b72a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.487176 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.487178 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b908dc5f-4777-40d9-ab29-35c36ba0b72a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.487217 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjmjk\" (UniqueName: \"kubernetes.io/projected/b908dc5f-4777-40d9-ab29-35c36ba0b72a-kube-api-access-zjmjk\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.487246 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.487286 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-scripts\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.487312 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.492523 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.493089 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.493297 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-scripts\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.493482 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.502612 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjmjk\" (UniqueName: \"kubernetes.io/projected/b908dc5f-4777-40d9-ab29-35c36ba0b72a-kube-api-access-zjmjk\") pod \"cinder-scheduler-0\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:22 crc kubenswrapper[4846]: I0202 13:44:22.620056 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 13:44:23 crc kubenswrapper[4846]: I0202 13:44:23.060266 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 13:44:23 crc kubenswrapper[4846]: I0202 13:44:23.603036 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b908dc5f-4777-40d9-ab29-35c36ba0b72a","Type":"ContainerStarted","Data":"56c352b0cdae8ced4c6c10f627a0894b00bf6a516269d5288794cc2e3185b6d1"} Feb 02 13:44:23 crc kubenswrapper[4846]: I0202 13:44:23.672433 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:44:23 crc kubenswrapper[4846]: I0202 13:44:23.672736 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5964a748-1a08-4359-b050-45153f45b3b7" containerName="cinder-api-log" containerID="cri-o://5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335" gracePeriod=30 Feb 02 13:44:23 crc kubenswrapper[4846]: I0202 13:44:23.672804 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5964a748-1a08-4359-b050-45153f45b3b7" containerName="cinder-api" containerID="cri-o://99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f" gracePeriod=30 Feb 02 13:44:24 crc kubenswrapper[4846]: I0202 13:44:24.612468 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b908dc5f-4777-40d9-ab29-35c36ba0b72a","Type":"ContainerStarted","Data":"6ed4533e5375c214d5de44ef693e420691e4c178bf31279c841114f333645f1f"} Feb 02 13:44:24 crc kubenswrapper[4846]: I0202 13:44:24.612788 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b908dc5f-4777-40d9-ab29-35c36ba0b72a","Type":"ContainerStarted","Data":"16af17f8b3b7cd907cbd033da73c0b0205c8e1bc8ea31aeaf297e5567b71ff91"} Feb 02 13:44:24 crc kubenswrapper[4846]: I0202 13:44:24.614134 4846 generic.go:334] "Generic (PLEG): container finished" podID="5964a748-1a08-4359-b050-45153f45b3b7" containerID="5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335" exitCode=143 Feb 02 13:44:24 crc kubenswrapper[4846]: I0202 13:44:24.614185 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5964a748-1a08-4359-b050-45153f45b3b7","Type":"ContainerDied","Data":"5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335"} Feb 02 13:44:24 crc kubenswrapper[4846]: I0202 13:44:24.634775 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.634757671 podStartE2EDuration="2.634757671s" podCreationTimestamp="2026-02-02 13:44:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:44:24.629021941 +0000 UTC m=+5695.857608854" watchObservedRunningTime="2026-02-02 13:44:24.634757671 +0000 UTC m=+5695.863344534" Feb 02 13:44:26 crc kubenswrapper[4846]: I0202 13:44:26.816453 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="5964a748-1a08-4359-b050-45153f45b3b7" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.1.53:8776/healthcheck\": read tcp 10.217.0.2:33576->10.217.1.53:8776: read: connection reset by peer" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.247943 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.382336 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5964a748-1a08-4359-b050-45153f45b3b7-logs\") pod \"5964a748-1a08-4359-b050-45153f45b3b7\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.382398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data\") pod \"5964a748-1a08-4359-b050-45153f45b3b7\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.382490 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-public-tls-certs\") pod \"5964a748-1a08-4359-b050-45153f45b3b7\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.382547 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5964a748-1a08-4359-b050-45153f45b3b7-etc-machine-id\") pod \"5964a748-1a08-4359-b050-45153f45b3b7\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.382615 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvkks\" (UniqueName: \"kubernetes.io/projected/5964a748-1a08-4359-b050-45153f45b3b7-kube-api-access-pvkks\") pod \"5964a748-1a08-4359-b050-45153f45b3b7\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.382662 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-scripts\") pod \"5964a748-1a08-4359-b050-45153f45b3b7\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.382690 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-internal-tls-certs\") pod \"5964a748-1a08-4359-b050-45153f45b3b7\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.382746 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data-custom\") pod \"5964a748-1a08-4359-b050-45153f45b3b7\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.382844 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-combined-ca-bundle\") pod \"5964a748-1a08-4359-b050-45153f45b3b7\" (UID: \"5964a748-1a08-4359-b050-45153f45b3b7\") " Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.383850 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5964a748-1a08-4359-b050-45153f45b3b7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5964a748-1a08-4359-b050-45153f45b3b7" (UID: "5964a748-1a08-4359-b050-45153f45b3b7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.385772 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5964a748-1a08-4359-b050-45153f45b3b7-logs" (OuterVolumeSpecName: "logs") pod "5964a748-1a08-4359-b050-45153f45b3b7" (UID: "5964a748-1a08-4359-b050-45153f45b3b7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.391085 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-scripts" (OuterVolumeSpecName: "scripts") pod "5964a748-1a08-4359-b050-45153f45b3b7" (UID: "5964a748-1a08-4359-b050-45153f45b3b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.399022 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5964a748-1a08-4359-b050-45153f45b3b7" (UID: "5964a748-1a08-4359-b050-45153f45b3b7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.406920 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5964a748-1a08-4359-b050-45153f45b3b7-kube-api-access-pvkks" (OuterVolumeSpecName: "kube-api-access-pvkks") pod "5964a748-1a08-4359-b050-45153f45b3b7" (UID: "5964a748-1a08-4359-b050-45153f45b3b7"). InnerVolumeSpecName "kube-api-access-pvkks". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.448147 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5964a748-1a08-4359-b050-45153f45b3b7" (UID: "5964a748-1a08-4359-b050-45153f45b3b7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.448281 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5964a748-1a08-4359-b050-45153f45b3b7" (UID: "5964a748-1a08-4359-b050-45153f45b3b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.471559 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data" (OuterVolumeSpecName: "config-data") pod "5964a748-1a08-4359-b050-45153f45b3b7" (UID: "5964a748-1a08-4359-b050-45153f45b3b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.478555 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5964a748-1a08-4359-b050-45153f45b3b7" (UID: "5964a748-1a08-4359-b050-45153f45b3b7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.485171 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.485207 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.485216 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5964a748-1a08-4359-b050-45153f45b3b7-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.485225 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.485232 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.485245 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5964a748-1a08-4359-b050-45153f45b3b7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.485262 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvkks\" (UniqueName: \"kubernetes.io/projected/5964a748-1a08-4359-b050-45153f45b3b7-kube-api-access-pvkks\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.485274 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.485289 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5964a748-1a08-4359-b050-45153f45b3b7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.620535 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.656761 4846 generic.go:334] "Generic (PLEG): container finished" podID="5964a748-1a08-4359-b050-45153f45b3b7" containerID="99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f" exitCode=0 Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.656812 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5964a748-1a08-4359-b050-45153f45b3b7","Type":"ContainerDied","Data":"99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f"} Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.656884 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5964a748-1a08-4359-b050-45153f45b3b7","Type":"ContainerDied","Data":"a825103d3d837cb44d6841b086e1d529433037d6a820887c2e42aa15a6490d19"} Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.656907 4846 scope.go:117] "RemoveContainer" containerID="99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.656993 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.700911 4846 scope.go:117] "RemoveContainer" containerID="5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.706965 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.722837 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.728418 4846 scope.go:117] "RemoveContainer" containerID="99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f" Feb 02 13:44:27 crc kubenswrapper[4846]: E0202 13:44:27.729004 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f\": container with ID starting with 99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f not found: ID does not exist" containerID="99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.729107 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f"} err="failed to get container status \"99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f\": rpc error: code = NotFound desc = could not find container \"99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f\": container with ID starting with 99a42ea5ea4cfc8754a1188e4f513cd99cea69cbe3ed2564f4a869fcfe58c21f not found: ID does not exist" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.729199 4846 scope.go:117] "RemoveContainer" containerID="5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335" Feb 02 13:44:27 crc kubenswrapper[4846]: E0202 13:44:27.729557 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335\": container with ID starting with 5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335 not found: ID does not exist" containerID="5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.729668 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335"} err="failed to get container status \"5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335\": rpc error: code = NotFound desc = could not find container \"5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335\": container with ID starting with 5357fb982e790bb25aecf77f315215654566ced0e0d9287d516c01036c8c4335 not found: ID does not exist" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.733487 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:44:27 crc kubenswrapper[4846]: E0202 13:44:27.734118 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5964a748-1a08-4359-b050-45153f45b3b7" containerName="cinder-api-log" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.734247 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5964a748-1a08-4359-b050-45153f45b3b7" containerName="cinder-api-log" Feb 02 13:44:27 crc kubenswrapper[4846]: E0202 13:44:27.734360 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5964a748-1a08-4359-b050-45153f45b3b7" containerName="cinder-api" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.734448 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5964a748-1a08-4359-b050-45153f45b3b7" containerName="cinder-api" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.734761 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5964a748-1a08-4359-b050-45153f45b3b7" containerName="cinder-api" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.734873 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5964a748-1a08-4359-b050-45153f45b3b7" containerName="cinder-api-log" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.745068 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.745190 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.747367 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.747574 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.747798 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.893370 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd759\" (UniqueName: \"kubernetes.io/projected/f4d63fe7-c925-40de-b144-8ec6f667f22f-kube-api-access-jd759\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.893507 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.893559 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-config-data\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.893669 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-config-data-custom\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.893721 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.893845 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4d63fe7-c925-40de-b144-8ec6f667f22f-logs\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.893877 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.893923 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4d63fe7-c925-40de-b144-8ec6f667f22f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.893987 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-scripts\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.995611 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-scripts\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.996058 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd759\" (UniqueName: \"kubernetes.io/projected/f4d63fe7-c925-40de-b144-8ec6f667f22f-kube-api-access-jd759\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.996139 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.996163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-config-data\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.996195 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-config-data-custom\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.996231 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.996274 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4d63fe7-c925-40de-b144-8ec6f667f22f-logs\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.996297 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.996332 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4d63fe7-c925-40de-b144-8ec6f667f22f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.996453 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f4d63fe7-c925-40de-b144-8ec6f667f22f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:27 crc kubenswrapper[4846]: I0202 13:44:27.996992 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4d63fe7-c925-40de-b144-8ec6f667f22f-logs\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:28 crc kubenswrapper[4846]: I0202 13:44:28.000360 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:28 crc kubenswrapper[4846]: I0202 13:44:28.000891 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:28 crc kubenswrapper[4846]: I0202 13:44:28.001113 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-scripts\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:28 crc kubenswrapper[4846]: I0202 13:44:28.001721 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-config-data-custom\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:28 crc kubenswrapper[4846]: I0202 13:44:28.003824 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-config-data\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:28 crc kubenswrapper[4846]: I0202 13:44:28.005702 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4d63fe7-c925-40de-b144-8ec6f667f22f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:28 crc kubenswrapper[4846]: I0202 13:44:28.015733 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd759\" (UniqueName: \"kubernetes.io/projected/f4d63fe7-c925-40de-b144-8ec6f667f22f-kube-api-access-jd759\") pod \"cinder-api-0\" (UID: \"f4d63fe7-c925-40de-b144-8ec6f667f22f\") " pod="openstack/cinder-api-0" Feb 02 13:44:28 crc kubenswrapper[4846]: I0202 13:44:28.069048 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 02 13:44:28 crc kubenswrapper[4846]: I0202 13:44:28.484600 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 02 13:44:28 crc kubenswrapper[4846]: W0202 13:44:28.493862 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4d63fe7_c925_40de_b144_8ec6f667f22f.slice/crio-fdb61ebb83dcf847dff3f5de34db4bbe48d0c2f70eba1eb19338e3a163eaa26f WatchSource:0}: Error finding container fdb61ebb83dcf847dff3f5de34db4bbe48d0c2f70eba1eb19338e3a163eaa26f: Status 404 returned error can't find the container with id fdb61ebb83dcf847dff3f5de34db4bbe48d0c2f70eba1eb19338e3a163eaa26f Feb 02 13:44:28 crc kubenswrapper[4846]: I0202 13:44:28.685404 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f4d63fe7-c925-40de-b144-8ec6f667f22f","Type":"ContainerStarted","Data":"fdb61ebb83dcf847dff3f5de34db4bbe48d0c2f70eba1eb19338e3a163eaa26f"} Feb 02 13:44:29 crc kubenswrapper[4846]: I0202 13:44:29.439242 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5964a748-1a08-4359-b050-45153f45b3b7" path="/var/lib/kubelet/pods/5964a748-1a08-4359-b050-45153f45b3b7/volumes" Feb 02 13:44:29 crc kubenswrapper[4846]: I0202 13:44:29.695260 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f4d63fe7-c925-40de-b144-8ec6f667f22f","Type":"ContainerStarted","Data":"efa9dcdeb38e42404cdd33273d71f0a5b9c809679ee02c11b7dfb3364b2b53dd"} Feb 02 13:44:29 crc kubenswrapper[4846]: I0202 13:44:29.695298 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f4d63fe7-c925-40de-b144-8ec6f667f22f","Type":"ContainerStarted","Data":"d275e8a7fac5819aa3540f29875bd6957538f3cd9aa3a491d883531a5ea07cfa"} Feb 02 13:44:29 crc kubenswrapper[4846]: I0202 13:44:29.695402 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 02 13:44:29 crc kubenswrapper[4846]: I0202 13:44:29.714434 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.714395284 podStartE2EDuration="2.714395284s" podCreationTimestamp="2026-02-02 13:44:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:44:29.710340435 +0000 UTC m=+5700.938927318" watchObservedRunningTime="2026-02-02 13:44:29.714395284 +0000 UTC m=+5700.942982157" Feb 02 13:44:31 crc kubenswrapper[4846]: I0202 13:44:31.424113 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:44:31 crc kubenswrapper[4846]: E0202 13:44:31.424851 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:44:32 crc kubenswrapper[4846]: I0202 13:44:32.834169 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 02 13:44:32 crc kubenswrapper[4846]: I0202 13:44:32.882435 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 13:44:33 crc kubenswrapper[4846]: I0202 13:44:33.736770 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" containerName="cinder-scheduler" containerID="cri-o://16af17f8b3b7cd907cbd033da73c0b0205c8e1bc8ea31aeaf297e5567b71ff91" gracePeriod=30 Feb 02 13:44:33 crc kubenswrapper[4846]: I0202 13:44:33.736871 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" containerName="probe" containerID="cri-o://6ed4533e5375c214d5de44ef693e420691e4c178bf31279c841114f333645f1f" gracePeriod=30 Feb 02 13:44:34 crc kubenswrapper[4846]: I0202 13:44:34.746579 4846 generic.go:334] "Generic (PLEG): container finished" podID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" containerID="6ed4533e5375c214d5de44ef693e420691e4c178bf31279c841114f333645f1f" exitCode=0 Feb 02 13:44:34 crc kubenswrapper[4846]: I0202 13:44:34.746656 4846 generic.go:334] "Generic (PLEG): container finished" podID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" containerID="16af17f8b3b7cd907cbd033da73c0b0205c8e1bc8ea31aeaf297e5567b71ff91" exitCode=0 Feb 02 13:44:34 crc kubenswrapper[4846]: I0202 13:44:34.746661 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b908dc5f-4777-40d9-ab29-35c36ba0b72a","Type":"ContainerDied","Data":"6ed4533e5375c214d5de44ef693e420691e4c178bf31279c841114f333645f1f"} Feb 02 13:44:34 crc kubenswrapper[4846]: I0202 13:44:34.746708 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b908dc5f-4777-40d9-ab29-35c36ba0b72a","Type":"ContainerDied","Data":"16af17f8b3b7cd907cbd033da73c0b0205c8e1bc8ea31aeaf297e5567b71ff91"} Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.284434 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.431519 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-scripts\") pod \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.431608 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data\") pod \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.431773 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data-custom\") pod \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.431973 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjmjk\" (UniqueName: \"kubernetes.io/projected/b908dc5f-4777-40d9-ab29-35c36ba0b72a-kube-api-access-zjmjk\") pod \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.432010 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b908dc5f-4777-40d9-ab29-35c36ba0b72a-etc-machine-id\") pod \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.432062 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-combined-ca-bundle\") pod \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\" (UID: \"b908dc5f-4777-40d9-ab29-35c36ba0b72a\") " Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.432217 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b908dc5f-4777-40d9-ab29-35c36ba0b72a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b908dc5f-4777-40d9-ab29-35c36ba0b72a" (UID: "b908dc5f-4777-40d9-ab29-35c36ba0b72a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.432895 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b908dc5f-4777-40d9-ab29-35c36ba0b72a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.447026 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b908dc5f-4777-40d9-ab29-35c36ba0b72a" (UID: "b908dc5f-4777-40d9-ab29-35c36ba0b72a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.447141 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-scripts" (OuterVolumeSpecName: "scripts") pod "b908dc5f-4777-40d9-ab29-35c36ba0b72a" (UID: "b908dc5f-4777-40d9-ab29-35c36ba0b72a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.459810 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b908dc5f-4777-40d9-ab29-35c36ba0b72a-kube-api-access-zjmjk" (OuterVolumeSpecName: "kube-api-access-zjmjk") pod "b908dc5f-4777-40d9-ab29-35c36ba0b72a" (UID: "b908dc5f-4777-40d9-ab29-35c36ba0b72a"). InnerVolumeSpecName "kube-api-access-zjmjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.491136 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b908dc5f-4777-40d9-ab29-35c36ba0b72a" (UID: "b908dc5f-4777-40d9-ab29-35c36ba0b72a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.534148 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjmjk\" (UniqueName: \"kubernetes.io/projected/b908dc5f-4777-40d9-ab29-35c36ba0b72a-kube-api-access-zjmjk\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.534417 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.534431 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.534443 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.544924 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data" (OuterVolumeSpecName: "config-data") pod "b908dc5f-4777-40d9-ab29-35c36ba0b72a" (UID: "b908dc5f-4777-40d9-ab29-35c36ba0b72a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.637273 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b908dc5f-4777-40d9-ab29-35c36ba0b72a-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.757297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b908dc5f-4777-40d9-ab29-35c36ba0b72a","Type":"ContainerDied","Data":"56c352b0cdae8ced4c6c10f627a0894b00bf6a516269d5288794cc2e3185b6d1"} Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.757349 4846 scope.go:117] "RemoveContainer" containerID="6ed4533e5375c214d5de44ef693e420691e4c178bf31279c841114f333645f1f" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.757481 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.779266 4846 scope.go:117] "RemoveContainer" containerID="16af17f8b3b7cd907cbd033da73c0b0205c8e1bc8ea31aeaf297e5567b71ff91" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.792820 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.810504 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.819847 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 13:44:35 crc kubenswrapper[4846]: E0202 13:44:35.821233 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" containerName="probe" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.821387 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" containerName="probe" Feb 02 13:44:35 crc kubenswrapper[4846]: E0202 13:44:35.821486 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" containerName="cinder-scheduler" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.821551 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" containerName="cinder-scheduler" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.821910 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" containerName="probe" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.822014 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" containerName="cinder-scheduler" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.823379 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.829123 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.852839 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.942092 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpt2m\" (UniqueName: \"kubernetes.io/projected/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-kube-api-access-wpt2m\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.942192 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.942313 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.942340 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.942364 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-scripts\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:35 crc kubenswrapper[4846]: I0202 13:44:35.942394 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-config-data\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.044200 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-config-data\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.044281 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpt2m\" (UniqueName: \"kubernetes.io/projected/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-kube-api-access-wpt2m\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.044345 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.044465 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.044508 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.044535 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-scripts\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.044782 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.049096 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.049113 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-scripts\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.049444 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.050261 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-config-data\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.061917 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpt2m\" (UniqueName: \"kubernetes.io/projected/f67f8ace-1e94-4f90-b8cf-d5a732eed4dd-kube-api-access-wpt2m\") pod \"cinder-scheduler-0\" (UID: \"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd\") " pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.154834 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.570905 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 02 13:44:36 crc kubenswrapper[4846]: I0202 13:44:36.765974 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd","Type":"ContainerStarted","Data":"c17b8c4c4ad7e623ca67903fbf4256c1f864fb3611a8d85a5b7179142b6eac76"} Feb 02 13:44:37 crc kubenswrapper[4846]: I0202 13:44:37.433360 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b908dc5f-4777-40d9-ab29-35c36ba0b72a" path="/var/lib/kubelet/pods/b908dc5f-4777-40d9-ab29-35c36ba0b72a/volumes" Feb 02 13:44:37 crc kubenswrapper[4846]: I0202 13:44:37.777709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd","Type":"ContainerStarted","Data":"63bbde23e5d57ea0167207d67926d8e20b68ac300980c6ed37e772848e15c33f"} Feb 02 13:44:37 crc kubenswrapper[4846]: I0202 13:44:37.777750 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f67f8ace-1e94-4f90-b8cf-d5a732eed4dd","Type":"ContainerStarted","Data":"aab9c3ca468d9ba639177d345532a8e06af05a9cdfa1c5fbcd5a03695e85d6c8"} Feb 02 13:44:37 crc kubenswrapper[4846]: I0202 13:44:37.799793 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.799774481 podStartE2EDuration="2.799774481s" podCreationTimestamp="2026-02-02 13:44:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:44:37.793498459 +0000 UTC m=+5709.022085352" watchObservedRunningTime="2026-02-02 13:44:37.799774481 +0000 UTC m=+5709.028361344" Feb 02 13:44:38 crc kubenswrapper[4846]: I0202 13:44:38.560060 4846 scope.go:117] "RemoveContainer" containerID="772335beca7d1b5e1d0b58c203fa5b7d7d0ef58f1904c8998b8a7ce8a4b930e2" Feb 02 13:44:38 crc kubenswrapper[4846]: I0202 13:44:38.612186 4846 scope.go:117] "RemoveContainer" containerID="3769afb30f73489bf34fbf983573ff8deab91b1d95e5418d2ef191f747eb9b14" Feb 02 13:44:40 crc kubenswrapper[4846]: I0202 13:44:40.026101 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 02 13:44:41 crc kubenswrapper[4846]: I0202 13:44:41.155360 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 02 13:44:45 crc kubenswrapper[4846]: I0202 13:44:45.424171 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:44:45 crc kubenswrapper[4846]: E0202 13:44:45.425515 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:44:46 crc kubenswrapper[4846]: I0202 13:44:46.371827 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.273789 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-zb62r"] Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.276534 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zb62r" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.282313 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zb62r"] Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.360208 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/083851d9-5552-40fb-8b8f-0185a23e1f36-operator-scripts\") pod \"glance-db-create-zb62r\" (UID: \"083851d9-5552-40fb-8b8f-0185a23e1f36\") " pod="openstack/glance-db-create-zb62r" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.360277 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlgwk\" (UniqueName: \"kubernetes.io/projected/083851d9-5552-40fb-8b8f-0185a23e1f36-kube-api-access-mlgwk\") pod \"glance-db-create-zb62r\" (UID: \"083851d9-5552-40fb-8b8f-0185a23e1f36\") " pod="openstack/glance-db-create-zb62r" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.376977 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b982-account-create-update-6p2b9"] Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.378680 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b982-account-create-update-6p2b9" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.382657 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.384052 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b982-account-create-update-6p2b9"] Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.461995 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752948fa-aec6-4754-baa0-c04dcd8dfb94-operator-scripts\") pod \"glance-b982-account-create-update-6p2b9\" (UID: \"752948fa-aec6-4754-baa0-c04dcd8dfb94\") " pod="openstack/glance-b982-account-create-update-6p2b9" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.462048 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w59rt\" (UniqueName: \"kubernetes.io/projected/752948fa-aec6-4754-baa0-c04dcd8dfb94-kube-api-access-w59rt\") pod \"glance-b982-account-create-update-6p2b9\" (UID: \"752948fa-aec6-4754-baa0-c04dcd8dfb94\") " pod="openstack/glance-b982-account-create-update-6p2b9" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.462103 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/083851d9-5552-40fb-8b8f-0185a23e1f36-operator-scripts\") pod \"glance-db-create-zb62r\" (UID: \"083851d9-5552-40fb-8b8f-0185a23e1f36\") " pod="openstack/glance-db-create-zb62r" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.462953 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/083851d9-5552-40fb-8b8f-0185a23e1f36-operator-scripts\") pod \"glance-db-create-zb62r\" (UID: \"083851d9-5552-40fb-8b8f-0185a23e1f36\") " pod="openstack/glance-db-create-zb62r" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.463000 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlgwk\" (UniqueName: \"kubernetes.io/projected/083851d9-5552-40fb-8b8f-0185a23e1f36-kube-api-access-mlgwk\") pod \"glance-db-create-zb62r\" (UID: \"083851d9-5552-40fb-8b8f-0185a23e1f36\") " pod="openstack/glance-db-create-zb62r" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.480701 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlgwk\" (UniqueName: \"kubernetes.io/projected/083851d9-5552-40fb-8b8f-0185a23e1f36-kube-api-access-mlgwk\") pod \"glance-db-create-zb62r\" (UID: \"083851d9-5552-40fb-8b8f-0185a23e1f36\") " pod="openstack/glance-db-create-zb62r" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.565332 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752948fa-aec6-4754-baa0-c04dcd8dfb94-operator-scripts\") pod \"glance-b982-account-create-update-6p2b9\" (UID: \"752948fa-aec6-4754-baa0-c04dcd8dfb94\") " pod="openstack/glance-b982-account-create-update-6p2b9" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.565400 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w59rt\" (UniqueName: \"kubernetes.io/projected/752948fa-aec6-4754-baa0-c04dcd8dfb94-kube-api-access-w59rt\") pod \"glance-b982-account-create-update-6p2b9\" (UID: \"752948fa-aec6-4754-baa0-c04dcd8dfb94\") " pod="openstack/glance-b982-account-create-update-6p2b9" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.566121 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752948fa-aec6-4754-baa0-c04dcd8dfb94-operator-scripts\") pod \"glance-b982-account-create-update-6p2b9\" (UID: \"752948fa-aec6-4754-baa0-c04dcd8dfb94\") " pod="openstack/glance-b982-account-create-update-6p2b9" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.580092 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w59rt\" (UniqueName: \"kubernetes.io/projected/752948fa-aec6-4754-baa0-c04dcd8dfb94-kube-api-access-w59rt\") pod \"glance-b982-account-create-update-6p2b9\" (UID: \"752948fa-aec6-4754-baa0-c04dcd8dfb94\") " pod="openstack/glance-b982-account-create-update-6p2b9" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.597819 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zb62r" Feb 02 13:44:47 crc kubenswrapper[4846]: I0202 13:44:47.698236 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b982-account-create-update-6p2b9" Feb 02 13:44:48 crc kubenswrapper[4846]: I0202 13:44:48.042150 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zb62r"] Feb 02 13:44:48 crc kubenswrapper[4846]: W0202 13:44:48.045873 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod083851d9_5552_40fb_8b8f_0185a23e1f36.slice/crio-450e9ab5c57c57c0863b9d6a95b4ce5ba7be1ed8c1e378f16254888037f43ec8 WatchSource:0}: Error finding container 450e9ab5c57c57c0863b9d6a95b4ce5ba7be1ed8c1e378f16254888037f43ec8: Status 404 returned error can't find the container with id 450e9ab5c57c57c0863b9d6a95b4ce5ba7be1ed8c1e378f16254888037f43ec8 Feb 02 13:44:48 crc kubenswrapper[4846]: I0202 13:44:48.162735 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b982-account-create-update-6p2b9"] Feb 02 13:44:48 crc kubenswrapper[4846]: W0202 13:44:48.162847 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod752948fa_aec6_4754_baa0_c04dcd8dfb94.slice/crio-3bd02b89727b748484fe4bff769c81a2af67413a41ab89439bcef7f1affd8e31 WatchSource:0}: Error finding container 3bd02b89727b748484fe4bff769c81a2af67413a41ab89439bcef7f1affd8e31: Status 404 returned error can't find the container with id 3bd02b89727b748484fe4bff769c81a2af67413a41ab89439bcef7f1affd8e31 Feb 02 13:44:48 crc kubenswrapper[4846]: I0202 13:44:48.874566 4846 generic.go:334] "Generic (PLEG): container finished" podID="083851d9-5552-40fb-8b8f-0185a23e1f36" containerID="f2eee019dc41b815eed61d6e84b86bf012d759d7822a6245d408781e923c3808" exitCode=0 Feb 02 13:44:48 crc kubenswrapper[4846]: I0202 13:44:48.874688 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zb62r" event={"ID":"083851d9-5552-40fb-8b8f-0185a23e1f36","Type":"ContainerDied","Data":"f2eee019dc41b815eed61d6e84b86bf012d759d7822a6245d408781e923c3808"} Feb 02 13:44:48 crc kubenswrapper[4846]: I0202 13:44:48.875035 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zb62r" event={"ID":"083851d9-5552-40fb-8b8f-0185a23e1f36","Type":"ContainerStarted","Data":"450e9ab5c57c57c0863b9d6a95b4ce5ba7be1ed8c1e378f16254888037f43ec8"} Feb 02 13:44:48 crc kubenswrapper[4846]: I0202 13:44:48.877002 4846 generic.go:334] "Generic (PLEG): container finished" podID="752948fa-aec6-4754-baa0-c04dcd8dfb94" containerID="78715439dbb4defd7ed7a34b9d10c49e8b1b8c0d3000953fbbad0f003ef1f97c" exitCode=0 Feb 02 13:44:48 crc kubenswrapper[4846]: I0202 13:44:48.877030 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b982-account-create-update-6p2b9" event={"ID":"752948fa-aec6-4754-baa0-c04dcd8dfb94","Type":"ContainerDied","Data":"78715439dbb4defd7ed7a34b9d10c49e8b1b8c0d3000953fbbad0f003ef1f97c"} Feb 02 13:44:48 crc kubenswrapper[4846]: I0202 13:44:48.877050 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b982-account-create-update-6p2b9" event={"ID":"752948fa-aec6-4754-baa0-c04dcd8dfb94","Type":"ContainerStarted","Data":"3bd02b89727b748484fe4bff769c81a2af67413a41ab89439bcef7f1affd8e31"} Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.296318 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b982-account-create-update-6p2b9" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.304543 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zb62r" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.419287 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752948fa-aec6-4754-baa0-c04dcd8dfb94-operator-scripts\") pod \"752948fa-aec6-4754-baa0-c04dcd8dfb94\" (UID: \"752948fa-aec6-4754-baa0-c04dcd8dfb94\") " Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.419466 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w59rt\" (UniqueName: \"kubernetes.io/projected/752948fa-aec6-4754-baa0-c04dcd8dfb94-kube-api-access-w59rt\") pod \"752948fa-aec6-4754-baa0-c04dcd8dfb94\" (UID: \"752948fa-aec6-4754-baa0-c04dcd8dfb94\") " Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.419573 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlgwk\" (UniqueName: \"kubernetes.io/projected/083851d9-5552-40fb-8b8f-0185a23e1f36-kube-api-access-mlgwk\") pod \"083851d9-5552-40fb-8b8f-0185a23e1f36\" (UID: \"083851d9-5552-40fb-8b8f-0185a23e1f36\") " Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.420255 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/752948fa-aec6-4754-baa0-c04dcd8dfb94-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "752948fa-aec6-4754-baa0-c04dcd8dfb94" (UID: "752948fa-aec6-4754-baa0-c04dcd8dfb94"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.420397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/083851d9-5552-40fb-8b8f-0185a23e1f36-operator-scripts\") pod \"083851d9-5552-40fb-8b8f-0185a23e1f36\" (UID: \"083851d9-5552-40fb-8b8f-0185a23e1f36\") " Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.420819 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/083851d9-5552-40fb-8b8f-0185a23e1f36-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "083851d9-5552-40fb-8b8f-0185a23e1f36" (UID: "083851d9-5552-40fb-8b8f-0185a23e1f36"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.421265 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/083851d9-5552-40fb-8b8f-0185a23e1f36-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.421286 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752948fa-aec6-4754-baa0-c04dcd8dfb94-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.424405 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/752948fa-aec6-4754-baa0-c04dcd8dfb94-kube-api-access-w59rt" (OuterVolumeSpecName: "kube-api-access-w59rt") pod "752948fa-aec6-4754-baa0-c04dcd8dfb94" (UID: "752948fa-aec6-4754-baa0-c04dcd8dfb94"). InnerVolumeSpecName "kube-api-access-w59rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.425080 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/083851d9-5552-40fb-8b8f-0185a23e1f36-kube-api-access-mlgwk" (OuterVolumeSpecName: "kube-api-access-mlgwk") pod "083851d9-5552-40fb-8b8f-0185a23e1f36" (UID: "083851d9-5552-40fb-8b8f-0185a23e1f36"). InnerVolumeSpecName "kube-api-access-mlgwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.523386 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlgwk\" (UniqueName: \"kubernetes.io/projected/083851d9-5552-40fb-8b8f-0185a23e1f36-kube-api-access-mlgwk\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.523444 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w59rt\" (UniqueName: \"kubernetes.io/projected/752948fa-aec6-4754-baa0-c04dcd8dfb94-kube-api-access-w59rt\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.892757 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zb62r" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.892851 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zb62r" event={"ID":"083851d9-5552-40fb-8b8f-0185a23e1f36","Type":"ContainerDied","Data":"450e9ab5c57c57c0863b9d6a95b4ce5ba7be1ed8c1e378f16254888037f43ec8"} Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.893499 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="450e9ab5c57c57c0863b9d6a95b4ce5ba7be1ed8c1e378f16254888037f43ec8" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.894749 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b982-account-create-update-6p2b9" event={"ID":"752948fa-aec6-4754-baa0-c04dcd8dfb94","Type":"ContainerDied","Data":"3bd02b89727b748484fe4bff769c81a2af67413a41ab89439bcef7f1affd8e31"} Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.894794 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bd02b89727b748484fe4bff769c81a2af67413a41ab89439bcef7f1affd8e31" Feb 02 13:44:50 crc kubenswrapper[4846]: I0202 13:44:50.894872 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b982-account-create-update-6p2b9" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.625767 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-9mpqm"] Feb 02 13:44:52 crc kubenswrapper[4846]: E0202 13:44:52.627505 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="752948fa-aec6-4754-baa0-c04dcd8dfb94" containerName="mariadb-account-create-update" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.627599 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="752948fa-aec6-4754-baa0-c04dcd8dfb94" containerName="mariadb-account-create-update" Feb 02 13:44:52 crc kubenswrapper[4846]: E0202 13:44:52.627704 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="083851d9-5552-40fb-8b8f-0185a23e1f36" containerName="mariadb-database-create" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.627797 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="083851d9-5552-40fb-8b8f-0185a23e1f36" containerName="mariadb-database-create" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.628169 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="752948fa-aec6-4754-baa0-c04dcd8dfb94" containerName="mariadb-account-create-update" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.628269 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="083851d9-5552-40fb-8b8f-0185a23e1f36" containerName="mariadb-database-create" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.629091 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.631561 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.631695 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-v59hg" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.644550 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-9mpqm"] Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.787669 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-config-data\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.787816 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-db-sync-config-data\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.787875 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfl4b\" (UniqueName: \"kubernetes.io/projected/526b1562-1fa5-4ca0-80a5-738f483d37a4-kube-api-access-vfl4b\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.787898 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-combined-ca-bundle\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.889897 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-db-sync-config-data\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.890208 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfl4b\" (UniqueName: \"kubernetes.io/projected/526b1562-1fa5-4ca0-80a5-738f483d37a4-kube-api-access-vfl4b\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.890331 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-combined-ca-bundle\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.890421 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-config-data\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.896170 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-combined-ca-bundle\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.896183 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-config-data\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.896774 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-db-sync-config-data\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.916870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfl4b\" (UniqueName: \"kubernetes.io/projected/526b1562-1fa5-4ca0-80a5-738f483d37a4-kube-api-access-vfl4b\") pod \"glance-db-sync-9mpqm\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:52 crc kubenswrapper[4846]: I0202 13:44:52.965716 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:53 crc kubenswrapper[4846]: I0202 13:44:53.471341 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-9mpqm"] Feb 02 13:44:53 crc kubenswrapper[4846]: W0202 13:44:53.471783 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod526b1562_1fa5_4ca0_80a5_738f483d37a4.slice/crio-2facf51e8d71d2ace94b37b02e0a48639f655de38101919192342363adcd0f0a WatchSource:0}: Error finding container 2facf51e8d71d2ace94b37b02e0a48639f655de38101919192342363adcd0f0a: Status 404 returned error can't find the container with id 2facf51e8d71d2ace94b37b02e0a48639f655de38101919192342363adcd0f0a Feb 02 13:44:53 crc kubenswrapper[4846]: I0202 13:44:53.919198 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9mpqm" event={"ID":"526b1562-1fa5-4ca0-80a5-738f483d37a4","Type":"ContainerStarted","Data":"2facf51e8d71d2ace94b37b02e0a48639f655de38101919192342363adcd0f0a"} Feb 02 13:44:54 crc kubenswrapper[4846]: I0202 13:44:54.926525 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9mpqm" event={"ID":"526b1562-1fa5-4ca0-80a5-738f483d37a4","Type":"ContainerStarted","Data":"cc31acf04483afec48f846a544006a56933563b4e24576cc654cfdd3cb045103"} Feb 02 13:44:54 crc kubenswrapper[4846]: I0202 13:44:54.948347 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-9mpqm" podStartSLOduration=2.948325585 podStartE2EDuration="2.948325585s" podCreationTimestamp="2026-02-02 13:44:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:44:54.93904892 +0000 UTC m=+5726.167635783" watchObservedRunningTime="2026-02-02 13:44:54.948325585 +0000 UTC m=+5726.176912448" Feb 02 13:44:57 crc kubenswrapper[4846]: I0202 13:44:57.427352 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:44:57 crc kubenswrapper[4846]: E0202 13:44:57.427906 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:44:57 crc kubenswrapper[4846]: I0202 13:44:57.951691 4846 generic.go:334] "Generic (PLEG): container finished" podID="526b1562-1fa5-4ca0-80a5-738f483d37a4" containerID="cc31acf04483afec48f846a544006a56933563b4e24576cc654cfdd3cb045103" exitCode=0 Feb 02 13:44:57 crc kubenswrapper[4846]: I0202 13:44:57.951777 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9mpqm" event={"ID":"526b1562-1fa5-4ca0-80a5-738f483d37a4","Type":"ContainerDied","Data":"cc31acf04483afec48f846a544006a56933563b4e24576cc654cfdd3cb045103"} Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.369505 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9mpqm" Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.418148 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfl4b\" (UniqueName: \"kubernetes.io/projected/526b1562-1fa5-4ca0-80a5-738f483d37a4-kube-api-access-vfl4b\") pod \"526b1562-1fa5-4ca0-80a5-738f483d37a4\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.418228 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-db-sync-config-data\") pod \"526b1562-1fa5-4ca0-80a5-738f483d37a4\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.418256 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-config-data\") pod \"526b1562-1fa5-4ca0-80a5-738f483d37a4\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.418289 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-combined-ca-bundle\") pod \"526b1562-1fa5-4ca0-80a5-738f483d37a4\" (UID: \"526b1562-1fa5-4ca0-80a5-738f483d37a4\") " Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.427041 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "526b1562-1fa5-4ca0-80a5-738f483d37a4" (UID: "526b1562-1fa5-4ca0-80a5-738f483d37a4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.428863 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/526b1562-1fa5-4ca0-80a5-738f483d37a4-kube-api-access-vfl4b" (OuterVolumeSpecName: "kube-api-access-vfl4b") pod "526b1562-1fa5-4ca0-80a5-738f483d37a4" (UID: "526b1562-1fa5-4ca0-80a5-738f483d37a4"). InnerVolumeSpecName "kube-api-access-vfl4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.451591 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "526b1562-1fa5-4ca0-80a5-738f483d37a4" (UID: "526b1562-1fa5-4ca0-80a5-738f483d37a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.474691 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-config-data" (OuterVolumeSpecName: "config-data") pod "526b1562-1fa5-4ca0-80a5-738f483d37a4" (UID: "526b1562-1fa5-4ca0-80a5-738f483d37a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.519192 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfl4b\" (UniqueName: \"kubernetes.io/projected/526b1562-1fa5-4ca0-80a5-738f483d37a4-kube-api-access-vfl4b\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.519229 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.519247 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.519260 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526b1562-1fa5-4ca0-80a5-738f483d37a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.970167 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9mpqm" event={"ID":"526b1562-1fa5-4ca0-80a5-738f483d37a4","Type":"ContainerDied","Data":"2facf51e8d71d2ace94b37b02e0a48639f655de38101919192342363adcd0f0a"} Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.970517 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2facf51e8d71d2ace94b37b02e0a48639f655de38101919192342363adcd0f0a" Feb 02 13:44:59 crc kubenswrapper[4846]: I0202 13:44:59.970298 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9mpqm" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.151660 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd"] Feb 02 13:45:00 crc kubenswrapper[4846]: E0202 13:45:00.152366 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="526b1562-1fa5-4ca0-80a5-738f483d37a4" containerName="glance-db-sync" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.152386 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="526b1562-1fa5-4ca0-80a5-738f483d37a4" containerName="glance-db-sync" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.152593 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="526b1562-1fa5-4ca0-80a5-738f483d37a4" containerName="glance-db-sync" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.153189 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.156104 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.156390 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.168523 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd"] Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.344518 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/061cc83a-9da1-4990-bf23-663267b2a816-config-volume\") pod \"collect-profiles-29500665-c82nd\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.344607 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kdgx\" (UniqueName: \"kubernetes.io/projected/061cc83a-9da1-4990-bf23-663267b2a816-kube-api-access-4kdgx\") pod \"collect-profiles-29500665-c82nd\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.347769 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/061cc83a-9da1-4990-bf23-663267b2a816-secret-volume\") pod \"collect-profiles-29500665-c82nd\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.398913 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c8cf44cdc-lt9dl"] Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.400519 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.409223 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c8cf44cdc-lt9dl"] Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.430818 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.440116 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.440408 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.443478 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.443714 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.443869 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-v59hg" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.450508 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/061cc83a-9da1-4990-bf23-663267b2a816-config-volume\") pod \"collect-profiles-29500665-c82nd\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.450571 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kdgx\" (UniqueName: \"kubernetes.io/projected/061cc83a-9da1-4990-bf23-663267b2a816-kube-api-access-4kdgx\") pod \"collect-profiles-29500665-c82nd\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.450687 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/061cc83a-9da1-4990-bf23-663267b2a816-secret-volume\") pod \"collect-profiles-29500665-c82nd\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.451976 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/061cc83a-9da1-4990-bf23-663267b2a816-config-volume\") pod \"collect-profiles-29500665-c82nd\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.460742 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/061cc83a-9da1-4990-bf23-663267b2a816-secret-volume\") pod \"collect-profiles-29500665-c82nd\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.492372 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kdgx\" (UniqueName: \"kubernetes.io/projected/061cc83a-9da1-4990-bf23-663267b2a816-kube-api-access-4kdgx\") pod \"collect-profiles-29500665-c82nd\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.551835 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-scripts\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.551929 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-config\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.551946 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.551964 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.552003 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-config-data\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.552026 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.552098 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.552180 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x4hb\" (UniqueName: \"kubernetes.io/projected/84d4a1d4-7a06-4223-a9aa-efb38da651b3-kube-api-access-7x4hb\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.552209 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7vzj\" (UniqueName: \"kubernetes.io/projected/3dd24092-5dab-4402-b67c-16b3821ccd07-kube-api-access-p7vzj\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.552246 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-dns-svc\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.552270 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-logs\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.603477 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.605341 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.607466 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.618589 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.654475 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.654529 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-config\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.655694 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.655761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-config-data\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.655808 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.655879 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.655914 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.656031 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x4hb\" (UniqueName: \"kubernetes.io/projected/84d4a1d4-7a06-4223-a9aa-efb38da651b3-kube-api-access-7x4hb\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.656087 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7vzj\" (UniqueName: \"kubernetes.io/projected/3dd24092-5dab-4402-b67c-16b3821ccd07-kube-api-access-p7vzj\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.656144 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-dns-svc\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.656181 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-logs\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.656232 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-scripts\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.657495 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-config\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.657499 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.657563 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-logs\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.657786 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.657886 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-dns-svc\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.663427 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-scripts\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.663818 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.678509 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-config-data\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.678726 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7vzj\" (UniqueName: \"kubernetes.io/projected/3dd24092-5dab-4402-b67c-16b3821ccd07-kube-api-access-p7vzj\") pod \"dnsmasq-dns-7c8cf44cdc-lt9dl\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.688911 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x4hb\" (UniqueName: \"kubernetes.io/projected/84d4a1d4-7a06-4223-a9aa-efb38da651b3-kube-api-access-7x4hb\") pod \"glance-default-external-api-0\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.753911 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.757461 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.757528 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x6zt\" (UniqueName: \"kubernetes.io/projected/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-kube-api-access-8x6zt\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.757638 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-logs\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.757664 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.757713 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.757772 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.774847 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.860201 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-logs\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.860236 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.860288 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.860326 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.860360 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.860389 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x6zt\" (UniqueName: \"kubernetes.io/projected/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-kube-api-access-8x6zt\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.861382 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-logs\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.862172 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.864976 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.866412 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.869773 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.878535 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.882802 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x6zt\" (UniqueName: \"kubernetes.io/projected/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-kube-api-access-8x6zt\") pod \"glance-default-internal-api-0\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:00 crc kubenswrapper[4846]: I0202 13:45:00.932677 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:01 crc kubenswrapper[4846]: I0202 13:45:01.351552 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c8cf44cdc-lt9dl"] Feb 02 13:45:01 crc kubenswrapper[4846]: I0202 13:45:01.471688 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd"] Feb 02 13:45:01 crc kubenswrapper[4846]: I0202 13:45:01.748126 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:45:01 crc kubenswrapper[4846]: W0202 13:45:01.759235 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0603ebfc_5a59_43fa_96ab_b8916ba5c3c2.slice/crio-1e879afc2f45f4dc1d1d3c121e93e2d900d2c986a0f48e38d114bbe0c00c160b WatchSource:0}: Error finding container 1e879afc2f45f4dc1d1d3c121e93e2d900d2c986a0f48e38d114bbe0c00c160b: Status 404 returned error can't find the container with id 1e879afc2f45f4dc1d1d3c121e93e2d900d2c986a0f48e38d114bbe0c00c160b Feb 02 13:45:01 crc kubenswrapper[4846]: I0202 13:45:01.934508 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:45:01 crc kubenswrapper[4846]: W0202 13:45:01.960161 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84d4a1d4_7a06_4223_a9aa_efb38da651b3.slice/crio-3527de8d490df0260f3df097d0ac6a6d375b3e0764b2b788e2220284463d34a7 WatchSource:0}: Error finding container 3527de8d490df0260f3df097d0ac6a6d375b3e0764b2b788e2220284463d34a7: Status 404 returned error can't find the container with id 3527de8d490df0260f3df097d0ac6a6d375b3e0764b2b788e2220284463d34a7 Feb 02 13:45:01 crc kubenswrapper[4846]: I0202 13:45:01.994493 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:45:02 crc kubenswrapper[4846]: I0202 13:45:02.052819 4846 generic.go:334] "Generic (PLEG): container finished" podID="3dd24092-5dab-4402-b67c-16b3821ccd07" containerID="bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912" exitCode=0 Feb 02 13:45:02 crc kubenswrapper[4846]: I0202 13:45:02.052943 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" event={"ID":"3dd24092-5dab-4402-b67c-16b3821ccd07","Type":"ContainerDied","Data":"bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912"} Feb 02 13:45:02 crc kubenswrapper[4846]: I0202 13:45:02.052978 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" event={"ID":"3dd24092-5dab-4402-b67c-16b3821ccd07","Type":"ContainerStarted","Data":"2b1905db9b4aa6382841f201ddd366f588bc1bd569dbdc1c4b3f0d44784efcd2"} Feb 02 13:45:02 crc kubenswrapper[4846]: I0202 13:45:02.055486 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"84d4a1d4-7a06-4223-a9aa-efb38da651b3","Type":"ContainerStarted","Data":"3527de8d490df0260f3df097d0ac6a6d375b3e0764b2b788e2220284463d34a7"} Feb 02 13:45:02 crc kubenswrapper[4846]: I0202 13:45:02.066817 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" event={"ID":"061cc83a-9da1-4990-bf23-663267b2a816","Type":"ContainerStarted","Data":"e00862ba6b7cda66a4e976c3aef34084c74f57957a876f2f21606a410a64a845"} Feb 02 13:45:02 crc kubenswrapper[4846]: I0202 13:45:02.066866 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" event={"ID":"061cc83a-9da1-4990-bf23-663267b2a816","Type":"ContainerStarted","Data":"bfc6af938f1dc92f1d8a5a4006b20ebef7f6b6baffe6a01a408c0e903e936e8e"} Feb 02 13:45:02 crc kubenswrapper[4846]: I0202 13:45:02.069246 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2","Type":"ContainerStarted","Data":"1e879afc2f45f4dc1d1d3c121e93e2d900d2c986a0f48e38d114bbe0c00c160b"} Feb 02 13:45:02 crc kubenswrapper[4846]: I0202 13:45:02.836682 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.100953 4846 generic.go:334] "Generic (PLEG): container finished" podID="061cc83a-9da1-4990-bf23-663267b2a816" containerID="e00862ba6b7cda66a4e976c3aef34084c74f57957a876f2f21606a410a64a845" exitCode=0 Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.101407 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" event={"ID":"061cc83a-9da1-4990-bf23-663267b2a816","Type":"ContainerDied","Data":"e00862ba6b7cda66a4e976c3aef34084c74f57957a876f2f21606a410a64a845"} Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.104117 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2","Type":"ContainerStarted","Data":"5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a"} Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.105908 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" event={"ID":"3dd24092-5dab-4402-b67c-16b3821ccd07","Type":"ContainerStarted","Data":"caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327"} Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.107138 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.109203 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"84d4a1d4-7a06-4223-a9aa-efb38da651b3","Type":"ContainerStarted","Data":"b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1"} Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.128412 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" podStartSLOduration=3.128388843 podStartE2EDuration="3.128388843s" podCreationTimestamp="2026-02-02 13:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:45:03.123546196 +0000 UTC m=+5734.352133089" watchObservedRunningTime="2026-02-02 13:45:03.128388843 +0000 UTC m=+5734.356975716" Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.577987 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.669374 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/061cc83a-9da1-4990-bf23-663267b2a816-secret-volume\") pod \"061cc83a-9da1-4990-bf23-663267b2a816\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.669438 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kdgx\" (UniqueName: \"kubernetes.io/projected/061cc83a-9da1-4990-bf23-663267b2a816-kube-api-access-4kdgx\") pod \"061cc83a-9da1-4990-bf23-663267b2a816\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.669487 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/061cc83a-9da1-4990-bf23-663267b2a816-config-volume\") pod \"061cc83a-9da1-4990-bf23-663267b2a816\" (UID: \"061cc83a-9da1-4990-bf23-663267b2a816\") " Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.670439 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/061cc83a-9da1-4990-bf23-663267b2a816-config-volume" (OuterVolumeSpecName: "config-volume") pod "061cc83a-9da1-4990-bf23-663267b2a816" (UID: "061cc83a-9da1-4990-bf23-663267b2a816"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.675904 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/061cc83a-9da1-4990-bf23-663267b2a816-kube-api-access-4kdgx" (OuterVolumeSpecName: "kube-api-access-4kdgx") pod "061cc83a-9da1-4990-bf23-663267b2a816" (UID: "061cc83a-9da1-4990-bf23-663267b2a816"). InnerVolumeSpecName "kube-api-access-4kdgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.688408 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/061cc83a-9da1-4990-bf23-663267b2a816-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "061cc83a-9da1-4990-bf23-663267b2a816" (UID: "061cc83a-9da1-4990-bf23-663267b2a816"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.771277 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/061cc83a-9da1-4990-bf23-663267b2a816-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.771320 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kdgx\" (UniqueName: \"kubernetes.io/projected/061cc83a-9da1-4990-bf23-663267b2a816-kube-api-access-4kdgx\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:03 crc kubenswrapper[4846]: I0202 13:45:03.771336 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/061cc83a-9da1-4990-bf23-663267b2a816-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.121486 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"84d4a1d4-7a06-4223-a9aa-efb38da651b3","Type":"ContainerStarted","Data":"c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b"} Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.121616 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" containerName="glance-log" containerID="cri-o://b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1" gracePeriod=30 Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.121775 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" containerName="glance-httpd" containerID="cri-o://c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b" gracePeriod=30 Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.124248 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2","Type":"ContainerStarted","Data":"2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09"} Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.124400 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" containerName="glance-log" containerID="cri-o://5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a" gracePeriod=30 Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.124482 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" containerName="glance-httpd" containerID="cri-o://2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09" gracePeriod=30 Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.127933 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.128177 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd" event={"ID":"061cc83a-9da1-4990-bf23-663267b2a816","Type":"ContainerDied","Data":"bfc6af938f1dc92f1d8a5a4006b20ebef7f6b6baffe6a01a408c0e903e936e8e"} Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.128227 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfc6af938f1dc92f1d8a5a4006b20ebef7f6b6baffe6a01a408c0e903e936e8e" Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.156297 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.156272146 podStartE2EDuration="4.156272146s" podCreationTimestamp="2026-02-02 13:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:45:04.153452958 +0000 UTC m=+5735.382039841" watchObservedRunningTime="2026-02-02 13:45:04.156272146 +0000 UTC m=+5735.384859009" Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.187223 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.187202187 podStartE2EDuration="4.187202187s" podCreationTimestamp="2026-02-02 13:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:45:04.176415796 +0000 UTC m=+5735.405002669" watchObservedRunningTime="2026-02-02 13:45:04.187202187 +0000 UTC m=+5735.415789050" Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.652875 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f"] Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.658970 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500620-4xq7f"] Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.924418 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.931418 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.997580 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-scripts\") pod \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.997705 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-httpd-run\") pod \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.997746 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x4hb\" (UniqueName: \"kubernetes.io/projected/84d4a1d4-7a06-4223-a9aa-efb38da651b3-kube-api-access-7x4hb\") pod \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.997823 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-logs\") pod \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.997839 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-logs\") pod \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.997860 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-combined-ca-bundle\") pod \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.997877 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-scripts\") pod \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.997903 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-config-data\") pod \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.997942 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x6zt\" (UniqueName: \"kubernetes.io/projected/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-kube-api-access-8x6zt\") pod \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.997998 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-combined-ca-bundle\") pod \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\" (UID: \"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.998044 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-config-data\") pod \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.998062 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-httpd-run\") pod \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\" (UID: \"84d4a1d4-7a06-4223-a9aa-efb38da651b3\") " Feb 02 13:45:04 crc kubenswrapper[4846]: I0202 13:45:04.999035 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "84d4a1d4-7a06-4223-a9aa-efb38da651b3" (UID: "84d4a1d4-7a06-4223-a9aa-efb38da651b3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.002898 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-logs" (OuterVolumeSpecName: "logs") pod "84d4a1d4-7a06-4223-a9aa-efb38da651b3" (UID: "84d4a1d4-7a06-4223-a9aa-efb38da651b3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.003280 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-logs" (OuterVolumeSpecName: "logs") pod "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" (UID: "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.003780 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" (UID: "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.005086 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-kube-api-access-8x6zt" (OuterVolumeSpecName: "kube-api-access-8x6zt") pod "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" (UID: "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2"). InnerVolumeSpecName "kube-api-access-8x6zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.021151 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d4a1d4-7a06-4223-a9aa-efb38da651b3-kube-api-access-7x4hb" (OuterVolumeSpecName: "kube-api-access-7x4hb") pod "84d4a1d4-7a06-4223-a9aa-efb38da651b3" (UID: "84d4a1d4-7a06-4223-a9aa-efb38da651b3"). InnerVolumeSpecName "kube-api-access-7x4hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.024546 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-scripts" (OuterVolumeSpecName: "scripts") pod "84d4a1d4-7a06-4223-a9aa-efb38da651b3" (UID: "84d4a1d4-7a06-4223-a9aa-efb38da651b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.028982 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-scripts" (OuterVolumeSpecName: "scripts") pod "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" (UID: "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.043849 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84d4a1d4-7a06-4223-a9aa-efb38da651b3" (UID: "84d4a1d4-7a06-4223-a9aa-efb38da651b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.047842 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" (UID: "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.062744 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-config-data" (OuterVolumeSpecName: "config-data") pod "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" (UID: "0603ebfc-5a59-43fa-96ab-b8916ba5c3c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.076411 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-config-data" (OuterVolumeSpecName: "config-data") pod "84d4a1d4-7a06-4223-a9aa-efb38da651b3" (UID: "84d4a1d4-7a06-4223-a9aa-efb38da651b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.099710 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.100219 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.100327 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.100430 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.100498 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.100555 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x4hb\" (UniqueName: \"kubernetes.io/projected/84d4a1d4-7a06-4223-a9aa-efb38da651b3-kube-api-access-7x4hb\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.100756 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d4a1d4-7a06-4223-a9aa-efb38da651b3-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.100876 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.100932 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d4a1d4-7a06-4223-a9aa-efb38da651b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.101009 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.101073 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.101139 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x6zt\" (UniqueName: \"kubernetes.io/projected/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2-kube-api-access-8x6zt\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.136176 4846 generic.go:334] "Generic (PLEG): container finished" podID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" containerID="c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b" exitCode=0 Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.137305 4846 generic.go:334] "Generic (PLEG): container finished" podID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" containerID="b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1" exitCode=143 Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.136420 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.136316 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"84d4a1d4-7a06-4223-a9aa-efb38da651b3","Type":"ContainerDied","Data":"c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b"} Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.138545 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"84d4a1d4-7a06-4223-a9aa-efb38da651b3","Type":"ContainerDied","Data":"b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1"} Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.138564 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"84d4a1d4-7a06-4223-a9aa-efb38da651b3","Type":"ContainerDied","Data":"3527de8d490df0260f3df097d0ac6a6d375b3e0764b2b788e2220284463d34a7"} Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.138588 4846 scope.go:117] "RemoveContainer" containerID="c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.141105 4846 generic.go:334] "Generic (PLEG): container finished" podID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" containerID="2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09" exitCode=0 Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.141129 4846 generic.go:334] "Generic (PLEG): container finished" podID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" containerID="5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a" exitCode=143 Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.142815 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.153321 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2","Type":"ContainerDied","Data":"2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09"} Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.153397 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2","Type":"ContainerDied","Data":"5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a"} Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.153414 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0603ebfc-5a59-43fa-96ab-b8916ba5c3c2","Type":"ContainerDied","Data":"1e879afc2f45f4dc1d1d3c121e93e2d900d2c986a0f48e38d114bbe0c00c160b"} Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.163448 4846 scope.go:117] "RemoveContainer" containerID="b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.187884 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.193950 4846 scope.go:117] "RemoveContainer" containerID="c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b" Feb 02 13:45:05 crc kubenswrapper[4846]: E0202 13:45:05.194465 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b\": container with ID starting with c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b not found: ID does not exist" containerID="c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.194528 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b"} err="failed to get container status \"c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b\": rpc error: code = NotFound desc = could not find container \"c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b\": container with ID starting with c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b not found: ID does not exist" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.194558 4846 scope.go:117] "RemoveContainer" containerID="b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1" Feb 02 13:45:05 crc kubenswrapper[4846]: E0202 13:45:05.195078 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1\": container with ID starting with b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1 not found: ID does not exist" containerID="b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.195112 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1"} err="failed to get container status \"b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1\": rpc error: code = NotFound desc = could not find container \"b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1\": container with ID starting with b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1 not found: ID does not exist" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.195138 4846 scope.go:117] "RemoveContainer" containerID="c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.195501 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b"} err="failed to get container status \"c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b\": rpc error: code = NotFound desc = could not find container \"c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b\": container with ID starting with c7ecf1974300e15aa4df09a09a0fb2b681168db665758f570cc888d7f363398b not found: ID does not exist" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.195532 4846 scope.go:117] "RemoveContainer" containerID="b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.195799 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1"} err="failed to get container status \"b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1\": rpc error: code = NotFound desc = could not find container \"b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1\": container with ID starting with b75880bf31c7f8ea4838cc061eb2075bebe9824d8a611bdbccb3bf338ede57f1 not found: ID does not exist" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.195825 4846 scope.go:117] "RemoveContainer" containerID="2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.197323 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.227791 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.235574 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:45:05 crc kubenswrapper[4846]: E0202 13:45:05.236020 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" containerName="glance-httpd" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.236043 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" containerName="glance-httpd" Feb 02 13:45:05 crc kubenswrapper[4846]: E0202 13:45:05.236065 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" containerName="glance-httpd" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.236073 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" containerName="glance-httpd" Feb 02 13:45:05 crc kubenswrapper[4846]: E0202 13:45:05.236094 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" containerName="glance-log" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.236103 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" containerName="glance-log" Feb 02 13:45:05 crc kubenswrapper[4846]: E0202 13:45:05.236134 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" containerName="glance-log" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.236142 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" containerName="glance-log" Feb 02 13:45:05 crc kubenswrapper[4846]: E0202 13:45:05.236155 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="061cc83a-9da1-4990-bf23-663267b2a816" containerName="collect-profiles" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.236164 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="061cc83a-9da1-4990-bf23-663267b2a816" containerName="collect-profiles" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.236368 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" containerName="glance-log" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.236388 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" containerName="glance-log" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.236407 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" containerName="glance-httpd" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.236417 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="061cc83a-9da1-4990-bf23-663267b2a816" containerName="collect-profiles" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.236434 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" containerName="glance-httpd" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.237584 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.241112 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.241433 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.242191 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.242468 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-v59hg" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.246278 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.270359 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.290604 4846 scope.go:117] "RemoveContainer" containerID="5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.308479 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-config-data\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.308893 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.309031 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-logs\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.309163 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.309285 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn9kz\" (UniqueName: \"kubernetes.io/projected/6c8c7249-5482-4258-a3a7-069aa4c087dd-kube-api-access-sn9kz\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.309523 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-scripts\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.309697 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.311928 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.320648 4846 scope.go:117] "RemoveContainer" containerID="2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09" Feb 02 13:45:05 crc kubenswrapper[4846]: E0202 13:45:05.321327 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09\": container with ID starting with 2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09 not found: ID does not exist" containerID="2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.321457 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09"} err="failed to get container status \"2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09\": rpc error: code = NotFound desc = could not find container \"2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09\": container with ID starting with 2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09 not found: ID does not exist" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.321566 4846 scope.go:117] "RemoveContainer" containerID="5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a" Feb 02 13:45:05 crc kubenswrapper[4846]: E0202 13:45:05.326125 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a\": container with ID starting with 5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a not found: ID does not exist" containerID="5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.326454 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a"} err="failed to get container status \"5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a\": rpc error: code = NotFound desc = could not find container \"5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a\": container with ID starting with 5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a not found: ID does not exist" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.326638 4846 scope.go:117] "RemoveContainer" containerID="2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.327345 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09"} err="failed to get container status \"2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09\": rpc error: code = NotFound desc = could not find container \"2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09\": container with ID starting with 2e7ee535c37adb4c72ad8e4bacdd51f0f5a457366c25b8274fa6a911bc8a3f09 not found: ID does not exist" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.327403 4846 scope.go:117] "RemoveContainer" containerID="5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.327778 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a"} err="failed to get container status \"5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a\": rpc error: code = NotFound desc = could not find container \"5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a\": container with ID starting with 5f37e6b5d992d49c60422da5707c02b87fa7427d466ed9fc2e7661c34c48898a not found: ID does not exist" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.354178 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.354312 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.356586 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.356729 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411607 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-scripts\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411685 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411721 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411785 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411807 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411830 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-config-data\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411924 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411946 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-logs\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411967 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.411985 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfp5w\" (UniqueName: \"kubernetes.io/projected/aa2668b1-bd01-4fce-93fd-d7fbf9917247-kube-api-access-lfp5w\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.412005 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn9kz\" (UniqueName: \"kubernetes.io/projected/6c8c7249-5482-4258-a3a7-069aa4c087dd-kube-api-access-sn9kz\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.412034 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.412053 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.412501 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.413954 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-logs\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.416241 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-config-data\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.416309 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-scripts\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.416533 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.419190 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.429728 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn9kz\" (UniqueName: \"kubernetes.io/projected/6c8c7249-5482-4258-a3a7-069aa4c087dd-kube-api-access-sn9kz\") pod \"glance-default-external-api-0\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.438381 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0603ebfc-5a59-43fa-96ab-b8916ba5c3c2" path="/var/lib/kubelet/pods/0603ebfc-5a59-43fa-96ab-b8916ba5c3c2/volumes" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.439161 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d4a1d4-7a06-4223-a9aa-efb38da651b3" path="/var/lib/kubelet/pods/84d4a1d4-7a06-4223-a9aa-efb38da651b3/volumes" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.439917 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d826ccc7-488f-4c36-9427-015882a3cbcd" path="/var/lib/kubelet/pods/d826ccc7-488f-4c36-9427-015882a3cbcd/volumes" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.513902 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.514015 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.514054 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.514084 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.514172 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfp5w\" (UniqueName: \"kubernetes.io/projected/aa2668b1-bd01-4fce-93fd-d7fbf9917247-kube-api-access-lfp5w\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.514219 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.514244 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.514407 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.514702 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.518870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.518916 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.519070 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.520605 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.536384 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfp5w\" (UniqueName: \"kubernetes.io/projected/aa2668b1-bd01-4fce-93fd-d7fbf9917247-kube-api-access-lfp5w\") pod \"glance-default-internal-api-0\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.574020 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 13:45:05 crc kubenswrapper[4846]: I0202 13:45:05.684379 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:06 crc kubenswrapper[4846]: I0202 13:45:06.170031 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:45:06 crc kubenswrapper[4846]: I0202 13:45:06.353959 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:45:06 crc kubenswrapper[4846]: W0202 13:45:06.364983 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa2668b1_bd01_4fce_93fd_d7fbf9917247.slice/crio-f5b4b95ecadde2a0e7f008aa24d9fc7252cf1b192c151c9987603d74a20fa0cd WatchSource:0}: Error finding container f5b4b95ecadde2a0e7f008aa24d9fc7252cf1b192c151c9987603d74a20fa0cd: Status 404 returned error can't find the container with id f5b4b95ecadde2a0e7f008aa24d9fc7252cf1b192c151c9987603d74a20fa0cd Feb 02 13:45:07 crc kubenswrapper[4846]: I0202 13:45:07.215894 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa2668b1-bd01-4fce-93fd-d7fbf9917247","Type":"ContainerStarted","Data":"4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20"} Feb 02 13:45:07 crc kubenswrapper[4846]: I0202 13:45:07.217191 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa2668b1-bd01-4fce-93fd-d7fbf9917247","Type":"ContainerStarted","Data":"f5b4b95ecadde2a0e7f008aa24d9fc7252cf1b192c151c9987603d74a20fa0cd"} Feb 02 13:45:07 crc kubenswrapper[4846]: I0202 13:45:07.220091 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c8c7249-5482-4258-a3a7-069aa4c087dd","Type":"ContainerStarted","Data":"418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71"} Feb 02 13:45:07 crc kubenswrapper[4846]: I0202 13:45:07.220484 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c8c7249-5482-4258-a3a7-069aa4c087dd","Type":"ContainerStarted","Data":"421880557b5059cba6ac3045d8787aa3bd81b82a0120944659f2a0395a6db752"} Feb 02 13:45:08 crc kubenswrapper[4846]: I0202 13:45:08.232679 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa2668b1-bd01-4fce-93fd-d7fbf9917247","Type":"ContainerStarted","Data":"41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5"} Feb 02 13:45:08 crc kubenswrapper[4846]: I0202 13:45:08.235963 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c8c7249-5482-4258-a3a7-069aa4c087dd","Type":"ContainerStarted","Data":"f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013"} Feb 02 13:45:08 crc kubenswrapper[4846]: I0202 13:45:08.261606 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.261588776 podStartE2EDuration="3.261588776s" podCreationTimestamp="2026-02-02 13:45:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:45:08.252726722 +0000 UTC m=+5739.481313595" watchObservedRunningTime="2026-02-02 13:45:08.261588776 +0000 UTC m=+5739.490175639" Feb 02 13:45:08 crc kubenswrapper[4846]: I0202 13:45:08.281956 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.281934301 podStartE2EDuration="3.281934301s" podCreationTimestamp="2026-02-02 13:45:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:45:08.276335285 +0000 UTC m=+5739.504922158" watchObservedRunningTime="2026-02-02 13:45:08.281934301 +0000 UTC m=+5739.510521164" Feb 02 13:45:09 crc kubenswrapper[4846]: I0202 13:45:09.432139 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:45:09 crc kubenswrapper[4846]: E0202 13:45:09.432748 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:45:10 crc kubenswrapper[4846]: I0202 13:45:10.756708 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:10 crc kubenswrapper[4846]: I0202 13:45:10.830233 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58c54f7d8f-kscsh"] Feb 02 13:45:10 crc kubenswrapper[4846]: I0202 13:45:10.830556 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" podUID="efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" containerName="dnsmasq-dns" containerID="cri-o://23fa90dc256fc9e685fbd9ec837bc4860fec55e3efe78e2ea4d3c97dc6bf3559" gracePeriod=10 Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.261284 4846 generic.go:334] "Generic (PLEG): container finished" podID="efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" containerID="23fa90dc256fc9e685fbd9ec837bc4860fec55e3efe78e2ea4d3c97dc6bf3559" exitCode=0 Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.261377 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" event={"ID":"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d","Type":"ContainerDied","Data":"23fa90dc256fc9e685fbd9ec837bc4860fec55e3efe78e2ea4d3c97dc6bf3559"} Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.352804 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.459390 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-dns-svc\") pod \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.459492 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-nb\") pod \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.459548 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-config\") pod \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.459588 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br5t6\" (UniqueName: \"kubernetes.io/projected/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-kube-api-access-br5t6\") pod \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.459688 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-sb\") pod \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\" (UID: \"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d\") " Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.525926 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-kube-api-access-br5t6" (OuterVolumeSpecName: "kube-api-access-br5t6") pod "efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" (UID: "efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d"). InnerVolumeSpecName "kube-api-access-br5t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.547089 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" (UID: "efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.557735 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" (UID: "efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.562028 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.562068 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.562082 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br5t6\" (UniqueName: \"kubernetes.io/projected/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-kube-api-access-br5t6\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.562221 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-config" (OuterVolumeSpecName: "config") pod "efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" (UID: "efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.564263 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" (UID: "efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.664141 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:11 crc kubenswrapper[4846]: I0202 13:45:11.664180 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:12 crc kubenswrapper[4846]: I0202 13:45:12.270941 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" event={"ID":"efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d","Type":"ContainerDied","Data":"7206bc7ec0510f8f6f925298363f0d21eeb8986a586499865a719a5204003b09"} Feb 02 13:45:12 crc kubenswrapper[4846]: I0202 13:45:12.270993 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58c54f7d8f-kscsh" Feb 02 13:45:12 crc kubenswrapper[4846]: I0202 13:45:12.271016 4846 scope.go:117] "RemoveContainer" containerID="23fa90dc256fc9e685fbd9ec837bc4860fec55e3efe78e2ea4d3c97dc6bf3559" Feb 02 13:45:12 crc kubenswrapper[4846]: I0202 13:45:12.292317 4846 scope.go:117] "RemoveContainer" containerID="55cd3b9adbf7d2dc7c46d93d3ea7a77904d06edac6918a9fe761ef3b6d533465" Feb 02 13:45:12 crc kubenswrapper[4846]: I0202 13:45:12.309874 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58c54f7d8f-kscsh"] Feb 02 13:45:12 crc kubenswrapper[4846]: I0202 13:45:12.318858 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58c54f7d8f-kscsh"] Feb 02 13:45:13 crc kubenswrapper[4846]: I0202 13:45:13.434423 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" path="/var/lib/kubelet/pods/efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d/volumes" Feb 02 13:45:15 crc kubenswrapper[4846]: I0202 13:45:15.574999 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 02 13:45:15 crc kubenswrapper[4846]: I0202 13:45:15.575121 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 02 13:45:15 crc kubenswrapper[4846]: I0202 13:45:15.606119 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 02 13:45:15 crc kubenswrapper[4846]: I0202 13:45:15.619530 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 02 13:45:15 crc kubenswrapper[4846]: I0202 13:45:15.684782 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:15 crc kubenswrapper[4846]: I0202 13:45:15.684837 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:15 crc kubenswrapper[4846]: I0202 13:45:15.774330 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:15 crc kubenswrapper[4846]: I0202 13:45:15.867076 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:16 crc kubenswrapper[4846]: I0202 13:45:16.308202 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 02 13:45:16 crc kubenswrapper[4846]: I0202 13:45:16.308250 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:16 crc kubenswrapper[4846]: I0202 13:45:16.308265 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 02 13:45:16 crc kubenswrapper[4846]: I0202 13:45:16.308276 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:18 crc kubenswrapper[4846]: I0202 13:45:18.236123 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 02 13:45:18 crc kubenswrapper[4846]: I0202 13:45:18.294202 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 02 13:45:18 crc kubenswrapper[4846]: I0202 13:45:18.310194 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:18 crc kubenswrapper[4846]: I0202 13:45:18.311996 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.423587 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:45:24 crc kubenswrapper[4846]: E0202 13:45:24.424502 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.760548 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-gvd6t"] Feb 02 13:45:24 crc kubenswrapper[4846]: E0202 13:45:24.761061 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" containerName="init" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.761079 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" containerName="init" Feb 02 13:45:24 crc kubenswrapper[4846]: E0202 13:45:24.761100 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" containerName="dnsmasq-dns" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.761106 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" containerName="dnsmasq-dns" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.761259 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="efb4fcc8-c4de-4fc5-a9c6-9e19be31ed5d" containerName="dnsmasq-dns" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.761851 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gvd6t" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.770709 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-gvd6t"] Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.845769 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvvhj\" (UniqueName: \"kubernetes.io/projected/4008e772-8019-48ec-8fe1-e49445af9f72-kube-api-access-xvvhj\") pod \"placement-db-create-gvd6t\" (UID: \"4008e772-8019-48ec-8fe1-e49445af9f72\") " pod="openstack/placement-db-create-gvd6t" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.846224 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4008e772-8019-48ec-8fe1-e49445af9f72-operator-scripts\") pod \"placement-db-create-gvd6t\" (UID: \"4008e772-8019-48ec-8fe1-e49445af9f72\") " pod="openstack/placement-db-create-gvd6t" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.858719 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c6fe-account-create-update-jbkbk"] Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.860152 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c6fe-account-create-update-jbkbk" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.862472 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.865908 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c6fe-account-create-update-jbkbk"] Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.946926 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lh96\" (UniqueName: \"kubernetes.io/projected/d0fbe008-d516-47f7-ada7-b2e588ae3db4-kube-api-access-9lh96\") pod \"placement-c6fe-account-create-update-jbkbk\" (UID: \"d0fbe008-d516-47f7-ada7-b2e588ae3db4\") " pod="openstack/placement-c6fe-account-create-update-jbkbk" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.947070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvvhj\" (UniqueName: \"kubernetes.io/projected/4008e772-8019-48ec-8fe1-e49445af9f72-kube-api-access-xvvhj\") pod \"placement-db-create-gvd6t\" (UID: \"4008e772-8019-48ec-8fe1-e49445af9f72\") " pod="openstack/placement-db-create-gvd6t" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.947113 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4008e772-8019-48ec-8fe1-e49445af9f72-operator-scripts\") pod \"placement-db-create-gvd6t\" (UID: \"4008e772-8019-48ec-8fe1-e49445af9f72\") " pod="openstack/placement-db-create-gvd6t" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.947160 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0fbe008-d516-47f7-ada7-b2e588ae3db4-operator-scripts\") pod \"placement-c6fe-account-create-update-jbkbk\" (UID: \"d0fbe008-d516-47f7-ada7-b2e588ae3db4\") " pod="openstack/placement-c6fe-account-create-update-jbkbk" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.948133 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4008e772-8019-48ec-8fe1-e49445af9f72-operator-scripts\") pod \"placement-db-create-gvd6t\" (UID: \"4008e772-8019-48ec-8fe1-e49445af9f72\") " pod="openstack/placement-db-create-gvd6t" Feb 02 13:45:24 crc kubenswrapper[4846]: I0202 13:45:24.964660 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvvhj\" (UniqueName: \"kubernetes.io/projected/4008e772-8019-48ec-8fe1-e49445af9f72-kube-api-access-xvvhj\") pod \"placement-db-create-gvd6t\" (UID: \"4008e772-8019-48ec-8fe1-e49445af9f72\") " pod="openstack/placement-db-create-gvd6t" Feb 02 13:45:25 crc kubenswrapper[4846]: I0202 13:45:25.048541 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0fbe008-d516-47f7-ada7-b2e588ae3db4-operator-scripts\") pod \"placement-c6fe-account-create-update-jbkbk\" (UID: \"d0fbe008-d516-47f7-ada7-b2e588ae3db4\") " pod="openstack/placement-c6fe-account-create-update-jbkbk" Feb 02 13:45:25 crc kubenswrapper[4846]: I0202 13:45:25.048600 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lh96\" (UniqueName: \"kubernetes.io/projected/d0fbe008-d516-47f7-ada7-b2e588ae3db4-kube-api-access-9lh96\") pod \"placement-c6fe-account-create-update-jbkbk\" (UID: \"d0fbe008-d516-47f7-ada7-b2e588ae3db4\") " pod="openstack/placement-c6fe-account-create-update-jbkbk" Feb 02 13:45:25 crc kubenswrapper[4846]: I0202 13:45:25.049510 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0fbe008-d516-47f7-ada7-b2e588ae3db4-operator-scripts\") pod \"placement-c6fe-account-create-update-jbkbk\" (UID: \"d0fbe008-d516-47f7-ada7-b2e588ae3db4\") " pod="openstack/placement-c6fe-account-create-update-jbkbk" Feb 02 13:45:25 crc kubenswrapper[4846]: I0202 13:45:25.064315 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lh96\" (UniqueName: \"kubernetes.io/projected/d0fbe008-d516-47f7-ada7-b2e588ae3db4-kube-api-access-9lh96\") pod \"placement-c6fe-account-create-update-jbkbk\" (UID: \"d0fbe008-d516-47f7-ada7-b2e588ae3db4\") " pod="openstack/placement-c6fe-account-create-update-jbkbk" Feb 02 13:45:25 crc kubenswrapper[4846]: I0202 13:45:25.086175 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gvd6t" Feb 02 13:45:25 crc kubenswrapper[4846]: I0202 13:45:25.190555 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c6fe-account-create-update-jbkbk" Feb 02 13:45:25 crc kubenswrapper[4846]: I0202 13:45:25.529463 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-gvd6t"] Feb 02 13:45:25 crc kubenswrapper[4846]: W0202 13:45:25.636607 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0fbe008_d516_47f7_ada7_b2e588ae3db4.slice/crio-51049784bf012c0deced577e55033a260d90e704fcedb5efe65aab460004c8aa WatchSource:0}: Error finding container 51049784bf012c0deced577e55033a260d90e704fcedb5efe65aab460004c8aa: Status 404 returned error can't find the container with id 51049784bf012c0deced577e55033a260d90e704fcedb5efe65aab460004c8aa Feb 02 13:45:25 crc kubenswrapper[4846]: I0202 13:45:25.641821 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c6fe-account-create-update-jbkbk"] Feb 02 13:45:26 crc kubenswrapper[4846]: I0202 13:45:26.408325 4846 generic.go:334] "Generic (PLEG): container finished" podID="d0fbe008-d516-47f7-ada7-b2e588ae3db4" containerID="2aac42116c042c9e7bf73001ac3c3955e6da7b89ddcb55571d22166145af7b33" exitCode=0 Feb 02 13:45:26 crc kubenswrapper[4846]: I0202 13:45:26.408413 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c6fe-account-create-update-jbkbk" event={"ID":"d0fbe008-d516-47f7-ada7-b2e588ae3db4","Type":"ContainerDied","Data":"2aac42116c042c9e7bf73001ac3c3955e6da7b89ddcb55571d22166145af7b33"} Feb 02 13:45:26 crc kubenswrapper[4846]: I0202 13:45:26.408472 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c6fe-account-create-update-jbkbk" event={"ID":"d0fbe008-d516-47f7-ada7-b2e588ae3db4","Type":"ContainerStarted","Data":"51049784bf012c0deced577e55033a260d90e704fcedb5efe65aab460004c8aa"} Feb 02 13:45:26 crc kubenswrapper[4846]: I0202 13:45:26.410339 4846 generic.go:334] "Generic (PLEG): container finished" podID="4008e772-8019-48ec-8fe1-e49445af9f72" containerID="5233047f1e8062b98adf89a4397f34bc1f322c3eb1830130d6478a35aba0da7a" exitCode=0 Feb 02 13:45:26 crc kubenswrapper[4846]: I0202 13:45:26.410384 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gvd6t" event={"ID":"4008e772-8019-48ec-8fe1-e49445af9f72","Type":"ContainerDied","Data":"5233047f1e8062b98adf89a4397f34bc1f322c3eb1830130d6478a35aba0da7a"} Feb 02 13:45:26 crc kubenswrapper[4846]: I0202 13:45:26.410416 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gvd6t" event={"ID":"4008e772-8019-48ec-8fe1-e49445af9f72","Type":"ContainerStarted","Data":"85048c3758d5dcd65b4a3d08719c5e0256b217cd7c831023dfcc70d03e9cfb8d"} Feb 02 13:45:27 crc kubenswrapper[4846]: I0202 13:45:27.987460 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c6fe-account-create-update-jbkbk" Feb 02 13:45:27 crc kubenswrapper[4846]: I0202 13:45:27.992510 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gvd6t" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.104280 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0fbe008-d516-47f7-ada7-b2e588ae3db4-operator-scripts\") pod \"d0fbe008-d516-47f7-ada7-b2e588ae3db4\" (UID: \"d0fbe008-d516-47f7-ada7-b2e588ae3db4\") " Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.104652 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4008e772-8019-48ec-8fe1-e49445af9f72-operator-scripts\") pod \"4008e772-8019-48ec-8fe1-e49445af9f72\" (UID: \"4008e772-8019-48ec-8fe1-e49445af9f72\") " Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.104711 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lh96\" (UniqueName: \"kubernetes.io/projected/d0fbe008-d516-47f7-ada7-b2e588ae3db4-kube-api-access-9lh96\") pod \"d0fbe008-d516-47f7-ada7-b2e588ae3db4\" (UID: \"d0fbe008-d516-47f7-ada7-b2e588ae3db4\") " Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.104791 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvvhj\" (UniqueName: \"kubernetes.io/projected/4008e772-8019-48ec-8fe1-e49445af9f72-kube-api-access-xvvhj\") pod \"4008e772-8019-48ec-8fe1-e49445af9f72\" (UID: \"4008e772-8019-48ec-8fe1-e49445af9f72\") " Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.105245 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4008e772-8019-48ec-8fe1-e49445af9f72-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4008e772-8019-48ec-8fe1-e49445af9f72" (UID: "4008e772-8019-48ec-8fe1-e49445af9f72"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.105706 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0fbe008-d516-47f7-ada7-b2e588ae3db4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d0fbe008-d516-47f7-ada7-b2e588ae3db4" (UID: "d0fbe008-d516-47f7-ada7-b2e588ae3db4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.111935 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4008e772-8019-48ec-8fe1-e49445af9f72-kube-api-access-xvvhj" (OuterVolumeSpecName: "kube-api-access-xvvhj") pod "4008e772-8019-48ec-8fe1-e49445af9f72" (UID: "4008e772-8019-48ec-8fe1-e49445af9f72"). InnerVolumeSpecName "kube-api-access-xvvhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.111993 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0fbe008-d516-47f7-ada7-b2e588ae3db4-kube-api-access-9lh96" (OuterVolumeSpecName: "kube-api-access-9lh96") pod "d0fbe008-d516-47f7-ada7-b2e588ae3db4" (UID: "d0fbe008-d516-47f7-ada7-b2e588ae3db4"). InnerVolumeSpecName "kube-api-access-9lh96". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.206098 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvvhj\" (UniqueName: \"kubernetes.io/projected/4008e772-8019-48ec-8fe1-e49445af9f72-kube-api-access-xvvhj\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.206133 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d0fbe008-d516-47f7-ada7-b2e588ae3db4-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.206145 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4008e772-8019-48ec-8fe1-e49445af9f72-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.206154 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lh96\" (UniqueName: \"kubernetes.io/projected/d0fbe008-d516-47f7-ada7-b2e588ae3db4-kube-api-access-9lh96\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.429522 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c6fe-account-create-update-jbkbk" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.429543 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c6fe-account-create-update-jbkbk" event={"ID":"d0fbe008-d516-47f7-ada7-b2e588ae3db4","Type":"ContainerDied","Data":"51049784bf012c0deced577e55033a260d90e704fcedb5efe65aab460004c8aa"} Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.429585 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51049784bf012c0deced577e55033a260d90e704fcedb5efe65aab460004c8aa" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.431216 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gvd6t" event={"ID":"4008e772-8019-48ec-8fe1-e49445af9f72","Type":"ContainerDied","Data":"85048c3758d5dcd65b4a3d08719c5e0256b217cd7c831023dfcc70d03e9cfb8d"} Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.431238 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85048c3758d5dcd65b4a3d08719c5e0256b217cd7c831023dfcc70d03e9cfb8d" Feb 02 13:45:28 crc kubenswrapper[4846]: I0202 13:45:28.431297 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gvd6t" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.143122 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b75886b77-7px8n"] Feb 02 13:45:30 crc kubenswrapper[4846]: E0202 13:45:30.143804 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4008e772-8019-48ec-8fe1-e49445af9f72" containerName="mariadb-database-create" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.143820 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4008e772-8019-48ec-8fe1-e49445af9f72" containerName="mariadb-database-create" Feb 02 13:45:30 crc kubenswrapper[4846]: E0202 13:45:30.143848 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fbe008-d516-47f7-ada7-b2e588ae3db4" containerName="mariadb-account-create-update" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.143854 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fbe008-d516-47f7-ada7-b2e588ae3db4" containerName="mariadb-account-create-update" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.144042 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0fbe008-d516-47f7-ada7-b2e588ae3db4" containerName="mariadb-account-create-update" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.144062 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4008e772-8019-48ec-8fe1-e49445af9f72" containerName="mariadb-database-create" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.144977 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.162103 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b75886b77-7px8n"] Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.213481 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-5dsdp"] Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.215010 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.217252 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.217539 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-829mf" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.217658 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.221577 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5dsdp"] Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.246158 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-dns-svc\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.246246 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-config\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.246300 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-sb\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.246345 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-nb\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.246363 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbpsb\" (UniqueName: \"kubernetes.io/projected/67d871ee-e010-401c-95e8-9a68e7350c59-kube-api-access-wbpsb\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.347737 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkrp8\" (UniqueName: \"kubernetes.io/projected/59524ffd-8aca-4e7c-bef3-e838e3312129-kube-api-access-vkrp8\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.347824 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-dns-svc\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.347848 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-config-data\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.347892 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-config\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.347927 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-sb\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.347968 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-combined-ca-bundle\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.348001 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59524ffd-8aca-4e7c-bef3-e838e3312129-logs\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.348028 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-scripts\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.348052 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-nb\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.348078 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbpsb\" (UniqueName: \"kubernetes.io/projected/67d871ee-e010-401c-95e8-9a68e7350c59-kube-api-access-wbpsb\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.348878 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-dns-svc\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.349162 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-config\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.349296 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-sb\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.349302 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-nb\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.376227 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbpsb\" (UniqueName: \"kubernetes.io/projected/67d871ee-e010-401c-95e8-9a68e7350c59-kube-api-access-wbpsb\") pod \"dnsmasq-dns-b75886b77-7px8n\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.450652 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-config-data\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.450794 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-combined-ca-bundle\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.450836 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59524ffd-8aca-4e7c-bef3-e838e3312129-logs\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.450863 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-scripts\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.450951 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkrp8\" (UniqueName: \"kubernetes.io/projected/59524ffd-8aca-4e7c-bef3-e838e3312129-kube-api-access-vkrp8\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.452032 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59524ffd-8aca-4e7c-bef3-e838e3312129-logs\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.456933 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-combined-ca-bundle\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.469257 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-scripts\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.469572 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-config-data\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.471834 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.478592 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkrp8\" (UniqueName: \"kubernetes.io/projected/59524ffd-8aca-4e7c-bef3-e838e3312129-kube-api-access-vkrp8\") pod \"placement-db-sync-5dsdp\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:30 crc kubenswrapper[4846]: I0202 13:45:30.530335 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:31 crc kubenswrapper[4846]: I0202 13:45:31.011568 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b75886b77-7px8n"] Feb 02 13:45:31 crc kubenswrapper[4846]: W0202 13:45:31.016825 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67d871ee_e010_401c_95e8_9a68e7350c59.slice/crio-b0b1446a1b02ab44285a14b451534bbd3afe7f53e59f47c2eff3691996f069b1 WatchSource:0}: Error finding container b0b1446a1b02ab44285a14b451534bbd3afe7f53e59f47c2eff3691996f069b1: Status 404 returned error can't find the container with id b0b1446a1b02ab44285a14b451534bbd3afe7f53e59f47c2eff3691996f069b1 Feb 02 13:45:31 crc kubenswrapper[4846]: I0202 13:45:31.196041 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5dsdp"] Feb 02 13:45:31 crc kubenswrapper[4846]: W0202 13:45:31.200776 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59524ffd_8aca_4e7c_bef3_e838e3312129.slice/crio-3b45091a02632eed5d92f51a2efeda7779fb0b9295746f5fb6064a94781d8ed2 WatchSource:0}: Error finding container 3b45091a02632eed5d92f51a2efeda7779fb0b9295746f5fb6064a94781d8ed2: Status 404 returned error can't find the container with id 3b45091a02632eed5d92f51a2efeda7779fb0b9295746f5fb6064a94781d8ed2 Feb 02 13:45:31 crc kubenswrapper[4846]: I0202 13:45:31.501017 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5dsdp" event={"ID":"59524ffd-8aca-4e7c-bef3-e838e3312129","Type":"ContainerStarted","Data":"3b45091a02632eed5d92f51a2efeda7779fb0b9295746f5fb6064a94781d8ed2"} Feb 02 13:45:31 crc kubenswrapper[4846]: I0202 13:45:31.508062 4846 generic.go:334] "Generic (PLEG): container finished" podID="67d871ee-e010-401c-95e8-9a68e7350c59" containerID="bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f" exitCode=0 Feb 02 13:45:31 crc kubenswrapper[4846]: I0202 13:45:31.508140 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b75886b77-7px8n" event={"ID":"67d871ee-e010-401c-95e8-9a68e7350c59","Type":"ContainerDied","Data":"bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f"} Feb 02 13:45:31 crc kubenswrapper[4846]: I0202 13:45:31.508173 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b75886b77-7px8n" event={"ID":"67d871ee-e010-401c-95e8-9a68e7350c59","Type":"ContainerStarted","Data":"b0b1446a1b02ab44285a14b451534bbd3afe7f53e59f47c2eff3691996f069b1"} Feb 02 13:45:32 crc kubenswrapper[4846]: I0202 13:45:32.517272 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5dsdp" event={"ID":"59524ffd-8aca-4e7c-bef3-e838e3312129","Type":"ContainerStarted","Data":"533f3721cb689883a09ff8d5d8bfa088d718faca9befee0385dc08870b204ae0"} Feb 02 13:45:32 crc kubenswrapper[4846]: I0202 13:45:32.521064 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b75886b77-7px8n" event={"ID":"67d871ee-e010-401c-95e8-9a68e7350c59","Type":"ContainerStarted","Data":"cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd"} Feb 02 13:45:32 crc kubenswrapper[4846]: I0202 13:45:32.521812 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:32 crc kubenswrapper[4846]: I0202 13:45:32.561375 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-5dsdp" podStartSLOduration=2.561356182 podStartE2EDuration="2.561356182s" podCreationTimestamp="2026-02-02 13:45:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:45:32.538587689 +0000 UTC m=+5763.767174552" watchObservedRunningTime="2026-02-02 13:45:32.561356182 +0000 UTC m=+5763.789943045" Feb 02 13:45:32 crc kubenswrapper[4846]: I0202 13:45:32.561604 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b75886b77-7px8n" podStartSLOduration=2.561600248 podStartE2EDuration="2.561600248s" podCreationTimestamp="2026-02-02 13:45:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:45:32.558036151 +0000 UTC m=+5763.786623024" watchObservedRunningTime="2026-02-02 13:45:32.561600248 +0000 UTC m=+5763.790187111" Feb 02 13:45:33 crc kubenswrapper[4846]: I0202 13:45:33.531103 4846 generic.go:334] "Generic (PLEG): container finished" podID="59524ffd-8aca-4e7c-bef3-e838e3312129" containerID="533f3721cb689883a09ff8d5d8bfa088d718faca9befee0385dc08870b204ae0" exitCode=0 Feb 02 13:45:33 crc kubenswrapper[4846]: I0202 13:45:33.531153 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5dsdp" event={"ID":"59524ffd-8aca-4e7c-bef3-e838e3312129","Type":"ContainerDied","Data":"533f3721cb689883a09ff8d5d8bfa088d718faca9befee0385dc08870b204ae0"} Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.049465 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.137337 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-config-data\") pod \"59524ffd-8aca-4e7c-bef3-e838e3312129\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.137408 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkrp8\" (UniqueName: \"kubernetes.io/projected/59524ffd-8aca-4e7c-bef3-e838e3312129-kube-api-access-vkrp8\") pod \"59524ffd-8aca-4e7c-bef3-e838e3312129\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.137574 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59524ffd-8aca-4e7c-bef3-e838e3312129-logs\") pod \"59524ffd-8aca-4e7c-bef3-e838e3312129\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.137685 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-combined-ca-bundle\") pod \"59524ffd-8aca-4e7c-bef3-e838e3312129\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.137773 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-scripts\") pod \"59524ffd-8aca-4e7c-bef3-e838e3312129\" (UID: \"59524ffd-8aca-4e7c-bef3-e838e3312129\") " Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.137843 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59524ffd-8aca-4e7c-bef3-e838e3312129-logs" (OuterVolumeSpecName: "logs") pod "59524ffd-8aca-4e7c-bef3-e838e3312129" (UID: "59524ffd-8aca-4e7c-bef3-e838e3312129"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.138262 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59524ffd-8aca-4e7c-bef3-e838e3312129-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.143091 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-scripts" (OuterVolumeSpecName: "scripts") pod "59524ffd-8aca-4e7c-bef3-e838e3312129" (UID: "59524ffd-8aca-4e7c-bef3-e838e3312129"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.143214 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59524ffd-8aca-4e7c-bef3-e838e3312129-kube-api-access-vkrp8" (OuterVolumeSpecName: "kube-api-access-vkrp8") pod "59524ffd-8aca-4e7c-bef3-e838e3312129" (UID: "59524ffd-8aca-4e7c-bef3-e838e3312129"). InnerVolumeSpecName "kube-api-access-vkrp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.179800 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59524ffd-8aca-4e7c-bef3-e838e3312129" (UID: "59524ffd-8aca-4e7c-bef3-e838e3312129"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.183550 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-config-data" (OuterVolumeSpecName: "config-data") pod "59524ffd-8aca-4e7c-bef3-e838e3312129" (UID: "59524ffd-8aca-4e7c-bef3-e838e3312129"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.239968 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.240004 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.240016 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkrp8\" (UniqueName: \"kubernetes.io/projected/59524ffd-8aca-4e7c-bef3-e838e3312129-kube-api-access-vkrp8\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.240027 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59524ffd-8aca-4e7c-bef3-e838e3312129-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.550083 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5dsdp" event={"ID":"59524ffd-8aca-4e7c-bef3-e838e3312129","Type":"ContainerDied","Data":"3b45091a02632eed5d92f51a2efeda7779fb0b9295746f5fb6064a94781d8ed2"} Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.550395 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b45091a02632eed5d92f51a2efeda7779fb0b9295746f5fb6064a94781d8ed2" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.550185 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5dsdp" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.645610 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-68f6f77966-bf8l5"] Feb 02 13:45:35 crc kubenswrapper[4846]: E0202 13:45:35.655246 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59524ffd-8aca-4e7c-bef3-e838e3312129" containerName="placement-db-sync" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.655669 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="59524ffd-8aca-4e7c-bef3-e838e3312129" containerName="placement-db-sync" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.656591 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="59524ffd-8aca-4e7c-bef3-e838e3312129" containerName="placement-db-sync" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.666158 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.669836 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-829mf" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.669992 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.670165 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.670456 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.670649 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.689840 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68f6f77966-bf8l5"] Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.761160 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-scripts\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.761233 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcsng\" (UniqueName: \"kubernetes.io/projected/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-kube-api-access-wcsng\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.761305 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-config-data\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.761356 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-logs\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.761380 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-internal-tls-certs\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.761402 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-combined-ca-bundle\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.761460 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-public-tls-certs\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.863243 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-config-data\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.863316 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-logs\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.863345 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-internal-tls-certs\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.863362 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-combined-ca-bundle\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.863417 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-public-tls-certs\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.863458 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-scripts\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.863495 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcsng\" (UniqueName: \"kubernetes.io/projected/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-kube-api-access-wcsng\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.864105 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-logs\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.867455 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-scripts\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.868208 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-public-tls-certs\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.868497 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-config-data\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.869163 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-combined-ca-bundle\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.870267 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-internal-tls-certs\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.881215 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcsng\" (UniqueName: \"kubernetes.io/projected/f14d64ee-705c-4c20-ab89-12cbfd89f8b8-kube-api-access-wcsng\") pod \"placement-68f6f77966-bf8l5\" (UID: \"f14d64ee-705c-4c20-ab89-12cbfd89f8b8\") " pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:35 crc kubenswrapper[4846]: I0202 13:45:35.996469 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:37 crc kubenswrapper[4846]: I0202 13:45:36.522043 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68f6f77966-bf8l5"] Feb 02 13:45:37 crc kubenswrapper[4846]: I0202 13:45:36.569244 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68f6f77966-bf8l5" event={"ID":"f14d64ee-705c-4c20-ab89-12cbfd89f8b8","Type":"ContainerStarted","Data":"e67256cfcc5ad30ba6d1c2bc97b67cc7be8be0afaee70eab9feb90ed35b310fe"} Feb 02 13:45:37 crc kubenswrapper[4846]: I0202 13:45:37.425659 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:45:37 crc kubenswrapper[4846]: E0202 13:45:37.426251 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:45:37 crc kubenswrapper[4846]: I0202 13:45:37.578582 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68f6f77966-bf8l5" event={"ID":"f14d64ee-705c-4c20-ab89-12cbfd89f8b8","Type":"ContainerStarted","Data":"c1f30cf830fd689c2a822ed2d3c88d09b2375922d97cd122df785e589fa0bc2f"} Feb 02 13:45:37 crc kubenswrapper[4846]: I0202 13:45:37.578640 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68f6f77966-bf8l5" event={"ID":"f14d64ee-705c-4c20-ab89-12cbfd89f8b8","Type":"ContainerStarted","Data":"7b60bb1a70067e7b0549724a62dc6ae2b010f110aebfba7e470051407e33571a"} Feb 02 13:45:37 crc kubenswrapper[4846]: I0202 13:45:37.579856 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:37 crc kubenswrapper[4846]: I0202 13:45:37.579887 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:45:37 crc kubenswrapper[4846]: I0202 13:45:37.604584 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-68f6f77966-bf8l5" podStartSLOduration=2.6045635000000003 podStartE2EDuration="2.6045635s" podCreationTimestamp="2026-02-02 13:45:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:45:37.596284429 +0000 UTC m=+5768.824871302" watchObservedRunningTime="2026-02-02 13:45:37.6045635 +0000 UTC m=+5768.833150363" Feb 02 13:45:38 crc kubenswrapper[4846]: I0202 13:45:38.727212 4846 scope.go:117] "RemoveContainer" containerID="2fb34f76229970f84c5cf931187b40d3199c06f257995dcf7afcafa2919be0c2" Feb 02 13:45:38 crc kubenswrapper[4846]: I0202 13:45:38.749113 4846 scope.go:117] "RemoveContainer" containerID="d666ebbafb8bb66a9e900c734c5679cad2fefe8dca5e81ae52b505726fe93e18" Feb 02 13:45:40 crc kubenswrapper[4846]: I0202 13:45:40.473877 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:45:40 crc kubenswrapper[4846]: I0202 13:45:40.536055 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c8cf44cdc-lt9dl"] Feb 02 13:45:40 crc kubenswrapper[4846]: I0202 13:45:40.536355 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" podUID="3dd24092-5dab-4402-b67c-16b3821ccd07" containerName="dnsmasq-dns" containerID="cri-o://caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327" gracePeriod=10 Feb 02 13:45:40 crc kubenswrapper[4846]: I0202 13:45:40.755515 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" podUID="3dd24092-5dab-4402-b67c-16b3821ccd07" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.61:5353: connect: connection refused" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.093397 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.170171 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-nb\") pod \"3dd24092-5dab-4402-b67c-16b3821ccd07\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.170286 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-config\") pod \"3dd24092-5dab-4402-b67c-16b3821ccd07\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.170394 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-dns-svc\") pod \"3dd24092-5dab-4402-b67c-16b3821ccd07\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.170465 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-sb\") pod \"3dd24092-5dab-4402-b67c-16b3821ccd07\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.170508 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7vzj\" (UniqueName: \"kubernetes.io/projected/3dd24092-5dab-4402-b67c-16b3821ccd07-kube-api-access-p7vzj\") pod \"3dd24092-5dab-4402-b67c-16b3821ccd07\" (UID: \"3dd24092-5dab-4402-b67c-16b3821ccd07\") " Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.178814 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd24092-5dab-4402-b67c-16b3821ccd07-kube-api-access-p7vzj" (OuterVolumeSpecName: "kube-api-access-p7vzj") pod "3dd24092-5dab-4402-b67c-16b3821ccd07" (UID: "3dd24092-5dab-4402-b67c-16b3821ccd07"). InnerVolumeSpecName "kube-api-access-p7vzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.221399 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3dd24092-5dab-4402-b67c-16b3821ccd07" (UID: "3dd24092-5dab-4402-b67c-16b3821ccd07"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.224140 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-config" (OuterVolumeSpecName: "config") pod "3dd24092-5dab-4402-b67c-16b3821ccd07" (UID: "3dd24092-5dab-4402-b67c-16b3821ccd07"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.239283 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3dd24092-5dab-4402-b67c-16b3821ccd07" (UID: "3dd24092-5dab-4402-b67c-16b3821ccd07"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.251761 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3dd24092-5dab-4402-b67c-16b3821ccd07" (UID: "3dd24092-5dab-4402-b67c-16b3821ccd07"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.272719 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.272755 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.272767 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7vzj\" (UniqueName: \"kubernetes.io/projected/3dd24092-5dab-4402-b67c-16b3821ccd07-kube-api-access-p7vzj\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.272776 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.272785 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dd24092-5dab-4402-b67c-16b3821ccd07-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.625699 4846 generic.go:334] "Generic (PLEG): container finished" podID="3dd24092-5dab-4402-b67c-16b3821ccd07" containerID="caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327" exitCode=0 Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.625711 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.625748 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" event={"ID":"3dd24092-5dab-4402-b67c-16b3821ccd07","Type":"ContainerDied","Data":"caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327"} Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.625799 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cf44cdc-lt9dl" event={"ID":"3dd24092-5dab-4402-b67c-16b3821ccd07","Type":"ContainerDied","Data":"2b1905db9b4aa6382841f201ddd366f588bc1bd569dbdc1c4b3f0d44784efcd2"} Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.625822 4846 scope.go:117] "RemoveContainer" containerID="caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.645402 4846 scope.go:117] "RemoveContainer" containerID="bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.654947 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c8cf44cdc-lt9dl"] Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.663751 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c8cf44cdc-lt9dl"] Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.710133 4846 scope.go:117] "RemoveContainer" containerID="caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327" Feb 02 13:45:41 crc kubenswrapper[4846]: E0202 13:45:41.710922 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327\": container with ID starting with caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327 not found: ID does not exist" containerID="caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.710980 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327"} err="failed to get container status \"caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327\": rpc error: code = NotFound desc = could not find container \"caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327\": container with ID starting with caa7892dceb3cb3fdd99fe7e38ff4c75759dbb82d7a55bdef6150239c1bee327 not found: ID does not exist" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.711007 4846 scope.go:117] "RemoveContainer" containerID="bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912" Feb 02 13:45:41 crc kubenswrapper[4846]: E0202 13:45:41.711564 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912\": container with ID starting with bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912 not found: ID does not exist" containerID="bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912" Feb 02 13:45:41 crc kubenswrapper[4846]: I0202 13:45:41.711606 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912"} err="failed to get container status \"bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912\": rpc error: code = NotFound desc = could not find container \"bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912\": container with ID starting with bf8af88594219d2696bb3459fec4800591c8a1df27db9381310af50dfde68912 not found: ID does not exist" Feb 02 13:45:43 crc kubenswrapper[4846]: I0202 13:45:43.434820 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dd24092-5dab-4402-b67c-16b3821ccd07" path="/var/lib/kubelet/pods/3dd24092-5dab-4402-b67c-16b3821ccd07/volumes" Feb 02 13:45:48 crc kubenswrapper[4846]: I0202 13:45:48.423470 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:45:48 crc kubenswrapper[4846]: E0202 13:45:48.423971 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:45:59 crc kubenswrapper[4846]: I0202 13:45:59.433488 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:45:59 crc kubenswrapper[4846]: E0202 13:45:59.434511 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:46:07 crc kubenswrapper[4846]: I0202 13:46:07.016019 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:46:07 crc kubenswrapper[4846]: I0202 13:46:07.082785 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68f6f77966-bf8l5" Feb 02 13:46:11 crc kubenswrapper[4846]: I0202 13:46:11.424537 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:46:11 crc kubenswrapper[4846]: E0202 13:46:11.425055 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:46:23 crc kubenswrapper[4846]: I0202 13:46:23.424210 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:46:23 crc kubenswrapper[4846]: E0202 13:46:23.424976 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.685321 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-lkmcf"] Feb 02 13:46:30 crc kubenswrapper[4846]: E0202 13:46:30.686421 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd24092-5dab-4402-b67c-16b3821ccd07" containerName="init" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.686441 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd24092-5dab-4402-b67c-16b3821ccd07" containerName="init" Feb 02 13:46:30 crc kubenswrapper[4846]: E0202 13:46:30.686480 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd24092-5dab-4402-b67c-16b3821ccd07" containerName="dnsmasq-dns" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.686489 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd24092-5dab-4402-b67c-16b3821ccd07" containerName="dnsmasq-dns" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.686714 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd24092-5dab-4402-b67c-16b3821ccd07" containerName="dnsmasq-dns" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.687419 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lkmcf" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.699732 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-2751-account-create-update-5dsmk"] Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.700968 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2751-account-create-update-5dsmk" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.704968 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.726270 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lkmcf"] Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.730276 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2751-account-create-update-5dsmk"] Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.799789 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-6l9gs"] Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.801001 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6l9gs" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.807406 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6l9gs"] Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.876108 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb99b\" (UniqueName: \"kubernetes.io/projected/f2cf8f76-47bb-4770-908b-723059ba5e13-kube-api-access-zb99b\") pod \"nova-api-db-create-lkmcf\" (UID: \"f2cf8f76-47bb-4770-908b-723059ba5e13\") " pod="openstack/nova-api-db-create-lkmcf" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.876184 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54e66fd9-2451-40c4-b317-783d545e1db9-operator-scripts\") pod \"nova-api-2751-account-create-update-5dsmk\" (UID: \"54e66fd9-2451-40c4-b317-783d545e1db9\") " pod="openstack/nova-api-2751-account-create-update-5dsmk" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.876230 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmbzt\" (UniqueName: \"kubernetes.io/projected/54e66fd9-2451-40c4-b317-783d545e1db9-kube-api-access-gmbzt\") pod \"nova-api-2751-account-create-update-5dsmk\" (UID: \"54e66fd9-2451-40c4-b317-783d545e1db9\") " pod="openstack/nova-api-2751-account-create-update-5dsmk" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.876337 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2cf8f76-47bb-4770-908b-723059ba5e13-operator-scripts\") pod \"nova-api-db-create-lkmcf\" (UID: \"f2cf8f76-47bb-4770-908b-723059ba5e13\") " pod="openstack/nova-api-db-create-lkmcf" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.900259 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-t6jvf"] Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.901665 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-t6jvf" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.912428 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-0340-account-create-update-rcslh"] Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.918803 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0340-account-create-update-rcslh" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.920766 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.927022 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-t6jvf"] Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.941018 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0340-account-create-update-rcslh"] Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.978333 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wdcz\" (UniqueName: \"kubernetes.io/projected/6d98e27a-a15d-49d4-a2ba-06accdf7770c-kube-api-access-8wdcz\") pod \"nova-cell0-db-create-6l9gs\" (UID: \"6d98e27a-a15d-49d4-a2ba-06accdf7770c\") " pod="openstack/nova-cell0-db-create-6l9gs" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.978383 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb99b\" (UniqueName: \"kubernetes.io/projected/f2cf8f76-47bb-4770-908b-723059ba5e13-kube-api-access-zb99b\") pod \"nova-api-db-create-lkmcf\" (UID: \"f2cf8f76-47bb-4770-908b-723059ba5e13\") " pod="openstack/nova-api-db-create-lkmcf" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.978418 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54e66fd9-2451-40c4-b317-783d545e1db9-operator-scripts\") pod \"nova-api-2751-account-create-update-5dsmk\" (UID: \"54e66fd9-2451-40c4-b317-783d545e1db9\") " pod="openstack/nova-api-2751-account-create-update-5dsmk" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.978446 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmbzt\" (UniqueName: \"kubernetes.io/projected/54e66fd9-2451-40c4-b317-783d545e1db9-kube-api-access-gmbzt\") pod \"nova-api-2751-account-create-update-5dsmk\" (UID: \"54e66fd9-2451-40c4-b317-783d545e1db9\") " pod="openstack/nova-api-2751-account-create-update-5dsmk" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.978473 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d98e27a-a15d-49d4-a2ba-06accdf7770c-operator-scripts\") pod \"nova-cell0-db-create-6l9gs\" (UID: \"6d98e27a-a15d-49d4-a2ba-06accdf7770c\") " pod="openstack/nova-cell0-db-create-6l9gs" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.978502 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2cf8f76-47bb-4770-908b-723059ba5e13-operator-scripts\") pod \"nova-api-db-create-lkmcf\" (UID: \"f2cf8f76-47bb-4770-908b-723059ba5e13\") " pod="openstack/nova-api-db-create-lkmcf" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.979241 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2cf8f76-47bb-4770-908b-723059ba5e13-operator-scripts\") pod \"nova-api-db-create-lkmcf\" (UID: \"f2cf8f76-47bb-4770-908b-723059ba5e13\") " pod="openstack/nova-api-db-create-lkmcf" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.979405 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54e66fd9-2451-40c4-b317-783d545e1db9-operator-scripts\") pod \"nova-api-2751-account-create-update-5dsmk\" (UID: \"54e66fd9-2451-40c4-b317-783d545e1db9\") " pod="openstack/nova-api-2751-account-create-update-5dsmk" Feb 02 13:46:30 crc kubenswrapper[4846]: I0202 13:46:30.998355 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb99b\" (UniqueName: \"kubernetes.io/projected/f2cf8f76-47bb-4770-908b-723059ba5e13-kube-api-access-zb99b\") pod \"nova-api-db-create-lkmcf\" (UID: \"f2cf8f76-47bb-4770-908b-723059ba5e13\") " pod="openstack/nova-api-db-create-lkmcf" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.000126 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmbzt\" (UniqueName: \"kubernetes.io/projected/54e66fd9-2451-40c4-b317-783d545e1db9-kube-api-access-gmbzt\") pod \"nova-api-2751-account-create-update-5dsmk\" (UID: \"54e66fd9-2451-40c4-b317-783d545e1db9\") " pod="openstack/nova-api-2751-account-create-update-5dsmk" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.027938 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lkmcf" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.081319 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2751-account-create-update-5dsmk" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.083130 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07556157-fc87-4ff5-9fba-2ee61250b2a5-operator-scripts\") pod \"nova-cell0-0340-account-create-update-rcslh\" (UID: \"07556157-fc87-4ff5-9fba-2ee61250b2a5\") " pod="openstack/nova-cell0-0340-account-create-update-rcslh" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.083185 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ea7209b-7df1-4b9a-a666-cd8257cf8088-operator-scripts\") pod \"nova-cell1-db-create-t6jvf\" (UID: \"7ea7209b-7df1-4b9a-a666-cd8257cf8088\") " pod="openstack/nova-cell1-db-create-t6jvf" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.083224 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wdcz\" (UniqueName: \"kubernetes.io/projected/6d98e27a-a15d-49d4-a2ba-06accdf7770c-kube-api-access-8wdcz\") pod \"nova-cell0-db-create-6l9gs\" (UID: \"6d98e27a-a15d-49d4-a2ba-06accdf7770c\") " pod="openstack/nova-cell0-db-create-6l9gs" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.083268 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d98e27a-a15d-49d4-a2ba-06accdf7770c-operator-scripts\") pod \"nova-cell0-db-create-6l9gs\" (UID: \"6d98e27a-a15d-49d4-a2ba-06accdf7770c\") " pod="openstack/nova-cell0-db-create-6l9gs" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.083298 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26skt\" (UniqueName: \"kubernetes.io/projected/07556157-fc87-4ff5-9fba-2ee61250b2a5-kube-api-access-26skt\") pod \"nova-cell0-0340-account-create-update-rcslh\" (UID: \"07556157-fc87-4ff5-9fba-2ee61250b2a5\") " pod="openstack/nova-cell0-0340-account-create-update-rcslh" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.083325 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bb45\" (UniqueName: \"kubernetes.io/projected/7ea7209b-7df1-4b9a-a666-cd8257cf8088-kube-api-access-7bb45\") pod \"nova-cell1-db-create-t6jvf\" (UID: \"7ea7209b-7df1-4b9a-a666-cd8257cf8088\") " pod="openstack/nova-cell1-db-create-t6jvf" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.084171 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d98e27a-a15d-49d4-a2ba-06accdf7770c-operator-scripts\") pod \"nova-cell0-db-create-6l9gs\" (UID: \"6d98e27a-a15d-49d4-a2ba-06accdf7770c\") " pod="openstack/nova-cell0-db-create-6l9gs" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.093764 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-34fb-account-create-update-sdf2h"] Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.095517 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.101853 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.105735 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wdcz\" (UniqueName: \"kubernetes.io/projected/6d98e27a-a15d-49d4-a2ba-06accdf7770c-kube-api-access-8wdcz\") pod \"nova-cell0-db-create-6l9gs\" (UID: \"6d98e27a-a15d-49d4-a2ba-06accdf7770c\") " pod="openstack/nova-cell0-db-create-6l9gs" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.115304 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-34fb-account-create-update-sdf2h"] Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.119336 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6l9gs" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.184576 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26skt\" (UniqueName: \"kubernetes.io/projected/07556157-fc87-4ff5-9fba-2ee61250b2a5-kube-api-access-26skt\") pod \"nova-cell0-0340-account-create-update-rcslh\" (UID: \"07556157-fc87-4ff5-9fba-2ee61250b2a5\") " pod="openstack/nova-cell0-0340-account-create-update-rcslh" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.184635 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bb45\" (UniqueName: \"kubernetes.io/projected/7ea7209b-7df1-4b9a-a666-cd8257cf8088-kube-api-access-7bb45\") pod \"nova-cell1-db-create-t6jvf\" (UID: \"7ea7209b-7df1-4b9a-a666-cd8257cf8088\") " pod="openstack/nova-cell1-db-create-t6jvf" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.184729 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07556157-fc87-4ff5-9fba-2ee61250b2a5-operator-scripts\") pod \"nova-cell0-0340-account-create-update-rcslh\" (UID: \"07556157-fc87-4ff5-9fba-2ee61250b2a5\") " pod="openstack/nova-cell0-0340-account-create-update-rcslh" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.184779 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ea7209b-7df1-4b9a-a666-cd8257cf8088-operator-scripts\") pod \"nova-cell1-db-create-t6jvf\" (UID: \"7ea7209b-7df1-4b9a-a666-cd8257cf8088\") " pod="openstack/nova-cell1-db-create-t6jvf" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.185642 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ea7209b-7df1-4b9a-a666-cd8257cf8088-operator-scripts\") pod \"nova-cell1-db-create-t6jvf\" (UID: \"7ea7209b-7df1-4b9a-a666-cd8257cf8088\") " pod="openstack/nova-cell1-db-create-t6jvf" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.185850 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07556157-fc87-4ff5-9fba-2ee61250b2a5-operator-scripts\") pod \"nova-cell0-0340-account-create-update-rcslh\" (UID: \"07556157-fc87-4ff5-9fba-2ee61250b2a5\") " pod="openstack/nova-cell0-0340-account-create-update-rcslh" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.204641 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bb45\" (UniqueName: \"kubernetes.io/projected/7ea7209b-7df1-4b9a-a666-cd8257cf8088-kube-api-access-7bb45\") pod \"nova-cell1-db-create-t6jvf\" (UID: \"7ea7209b-7df1-4b9a-a666-cd8257cf8088\") " pod="openstack/nova-cell1-db-create-t6jvf" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.216356 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26skt\" (UniqueName: \"kubernetes.io/projected/07556157-fc87-4ff5-9fba-2ee61250b2a5-kube-api-access-26skt\") pod \"nova-cell0-0340-account-create-update-rcslh\" (UID: \"07556157-fc87-4ff5-9fba-2ee61250b2a5\") " pod="openstack/nova-cell0-0340-account-create-update-rcslh" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.227445 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-t6jvf" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.243889 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0340-account-create-update-rcslh" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.288452 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5073120c-a1eb-4cf7-8b4c-a2b807027589-operator-scripts\") pod \"nova-cell1-34fb-account-create-update-sdf2h\" (UID: \"5073120c-a1eb-4cf7-8b4c-a2b807027589\") " pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.288612 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrf5n\" (UniqueName: \"kubernetes.io/projected/5073120c-a1eb-4cf7-8b4c-a2b807027589-kube-api-access-jrf5n\") pod \"nova-cell1-34fb-account-create-update-sdf2h\" (UID: \"5073120c-a1eb-4cf7-8b4c-a2b807027589\") " pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.390325 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrf5n\" (UniqueName: \"kubernetes.io/projected/5073120c-a1eb-4cf7-8b4c-a2b807027589-kube-api-access-jrf5n\") pod \"nova-cell1-34fb-account-create-update-sdf2h\" (UID: \"5073120c-a1eb-4cf7-8b4c-a2b807027589\") " pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.390465 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5073120c-a1eb-4cf7-8b4c-a2b807027589-operator-scripts\") pod \"nova-cell1-34fb-account-create-update-sdf2h\" (UID: \"5073120c-a1eb-4cf7-8b4c-a2b807027589\") " pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.391560 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5073120c-a1eb-4cf7-8b4c-a2b807027589-operator-scripts\") pod \"nova-cell1-34fb-account-create-update-sdf2h\" (UID: \"5073120c-a1eb-4cf7-8b4c-a2b807027589\") " pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.407122 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrf5n\" (UniqueName: \"kubernetes.io/projected/5073120c-a1eb-4cf7-8b4c-a2b807027589-kube-api-access-jrf5n\") pod \"nova-cell1-34fb-account-create-update-sdf2h\" (UID: \"5073120c-a1eb-4cf7-8b4c-a2b807027589\") " pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.499447 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" Feb 02 13:46:31 crc kubenswrapper[4846]: W0202 13:46:31.563524 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2cf8f76_47bb_4770_908b_723059ba5e13.slice/crio-3951419cab56c3d86cf88d9ee0e57d1f3fbb82c637c827b33ac9dcf8c87ec75a WatchSource:0}: Error finding container 3951419cab56c3d86cf88d9ee0e57d1f3fbb82c637c827b33ac9dcf8c87ec75a: Status 404 returned error can't find the container with id 3951419cab56c3d86cf88d9ee0e57d1f3fbb82c637c827b33ac9dcf8c87ec75a Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.571653 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lkmcf"] Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.658263 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2751-account-create-update-5dsmk"] Feb 02 13:46:31 crc kubenswrapper[4846]: W0202 13:46:31.665184 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54e66fd9_2451_40c4_b317_783d545e1db9.slice/crio-b2ff16d0e09dbdd769131146ae14dd8faf34610e3ce917f55137e13f348568fb WatchSource:0}: Error finding container b2ff16d0e09dbdd769131146ae14dd8faf34610e3ce917f55137e13f348568fb: Status 404 returned error can't find the container with id b2ff16d0e09dbdd769131146ae14dd8faf34610e3ce917f55137e13f348568fb Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.772639 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6l9gs"] Feb 02 13:46:31 crc kubenswrapper[4846]: W0202 13:46:31.782694 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d98e27a_a15d_49d4_a2ba_06accdf7770c.slice/crio-57cbc064fa6bf52db3b469732aa500367cb178e3ab1836ba275bd0075101d7d6 WatchSource:0}: Error finding container 57cbc064fa6bf52db3b469732aa500367cb178e3ab1836ba275bd0075101d7d6: Status 404 returned error can't find the container with id 57cbc064fa6bf52db3b469732aa500367cb178e3ab1836ba275bd0075101d7d6 Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.836568 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-t6jvf"] Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.868609 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0340-account-create-update-rcslh"] Feb 02 13:46:31 crc kubenswrapper[4846]: I0202 13:46:31.968917 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-34fb-account-create-update-sdf2h"] Feb 02 13:46:32 crc kubenswrapper[4846]: W0202 13:46:32.016317 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5073120c_a1eb_4cf7_8b4c_a2b807027589.slice/crio-a10deab1b15a241261a006c190f2bd61b16a5d9e7be398921cfbd7ca5fa19547 WatchSource:0}: Error finding container a10deab1b15a241261a006c190f2bd61b16a5d9e7be398921cfbd7ca5fa19547: Status 404 returned error can't find the container with id a10deab1b15a241261a006c190f2bd61b16a5d9e7be398921cfbd7ca5fa19547 Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.107827 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" event={"ID":"5073120c-a1eb-4cf7-8b4c-a2b807027589","Type":"ContainerStarted","Data":"a10deab1b15a241261a006c190f2bd61b16a5d9e7be398921cfbd7ca5fa19547"} Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.112570 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-t6jvf" event={"ID":"7ea7209b-7df1-4b9a-a666-cd8257cf8088","Type":"ContainerStarted","Data":"8893cbeaad2d8b9be4717112cb07b6e1b8a3d21172730b7a2da7125d168c166e"} Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.112617 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-t6jvf" event={"ID":"7ea7209b-7df1-4b9a-a666-cd8257cf8088","Type":"ContainerStarted","Data":"e0263f1f67e095fbcb3b43f433b746ac4a6d2a34dbbc0a83fe7a608a9e22fb32"} Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.114888 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2751-account-create-update-5dsmk" event={"ID":"54e66fd9-2451-40c4-b317-783d545e1db9","Type":"ContainerStarted","Data":"37020039b2fa96897ec7fba8404021b59483a98cc8dc8eda772c3424a2bd6350"} Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.114950 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2751-account-create-update-5dsmk" event={"ID":"54e66fd9-2451-40c4-b317-783d545e1db9","Type":"ContainerStarted","Data":"b2ff16d0e09dbdd769131146ae14dd8faf34610e3ce917f55137e13f348568fb"} Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.117783 4846 generic.go:334] "Generic (PLEG): container finished" podID="f2cf8f76-47bb-4770-908b-723059ba5e13" containerID="c5e00b3a85089e49d54b4fcbb634f4a71d45c3cadc78cc4584d3f28f246ef73e" exitCode=0 Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.117877 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lkmcf" event={"ID":"f2cf8f76-47bb-4770-908b-723059ba5e13","Type":"ContainerDied","Data":"c5e00b3a85089e49d54b4fcbb634f4a71d45c3cadc78cc4584d3f28f246ef73e"} Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.117910 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lkmcf" event={"ID":"f2cf8f76-47bb-4770-908b-723059ba5e13","Type":"ContainerStarted","Data":"3951419cab56c3d86cf88d9ee0e57d1f3fbb82c637c827b33ac9dcf8c87ec75a"} Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.119728 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0340-account-create-update-rcslh" event={"ID":"07556157-fc87-4ff5-9fba-2ee61250b2a5","Type":"ContainerStarted","Data":"bde9ed7b44432a961213246d09167ed3f93e4af598387c455fe0acd27369465e"} Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.124577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6l9gs" event={"ID":"6d98e27a-a15d-49d4-a2ba-06accdf7770c","Type":"ContainerStarted","Data":"f129d363d1dbdddac20c21c32b3245bac34e95f3d272ad4d1e2ff8d825e642fb"} Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.124634 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6l9gs" event={"ID":"6d98e27a-a15d-49d4-a2ba-06accdf7770c","Type":"ContainerStarted","Data":"57cbc064fa6bf52db3b469732aa500367cb178e3ab1836ba275bd0075101d7d6"} Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.129968 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-t6jvf" podStartSLOduration=2.129954085 podStartE2EDuration="2.129954085s" podCreationTimestamp="2026-02-02 13:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:32.126229295 +0000 UTC m=+5823.354816158" watchObservedRunningTime="2026-02-02 13:46:32.129954085 +0000 UTC m=+5823.358540948" Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.147165 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-2751-account-create-update-5dsmk" podStartSLOduration=2.147143193 podStartE2EDuration="2.147143193s" podCreationTimestamp="2026-02-02 13:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:32.143353151 +0000 UTC m=+5823.371940004" watchObservedRunningTime="2026-02-02 13:46:32.147143193 +0000 UTC m=+5823.375730056" Feb 02 13:46:32 crc kubenswrapper[4846]: I0202 13:46:32.160355 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-6l9gs" podStartSLOduration=2.160337753 podStartE2EDuration="2.160337753s" podCreationTimestamp="2026-02-02 13:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:32.155776393 +0000 UTC m=+5823.384363266" watchObservedRunningTime="2026-02-02 13:46:32.160337753 +0000 UTC m=+5823.388924616" Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.141639 4846 generic.go:334] "Generic (PLEG): container finished" podID="07556157-fc87-4ff5-9fba-2ee61250b2a5" containerID="567d80943db2be70626e34b3ff797b1eae044b0b8ede6c647397a824c261a48e" exitCode=0 Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.141796 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0340-account-create-update-rcslh" event={"ID":"07556157-fc87-4ff5-9fba-2ee61250b2a5","Type":"ContainerDied","Data":"567d80943db2be70626e34b3ff797b1eae044b0b8ede6c647397a824c261a48e"} Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.144584 4846 generic.go:334] "Generic (PLEG): container finished" podID="6d98e27a-a15d-49d4-a2ba-06accdf7770c" containerID="f129d363d1dbdddac20c21c32b3245bac34e95f3d272ad4d1e2ff8d825e642fb" exitCode=0 Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.144699 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6l9gs" event={"ID":"6d98e27a-a15d-49d4-a2ba-06accdf7770c","Type":"ContainerDied","Data":"f129d363d1dbdddac20c21c32b3245bac34e95f3d272ad4d1e2ff8d825e642fb"} Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.156664 4846 generic.go:334] "Generic (PLEG): container finished" podID="5073120c-a1eb-4cf7-8b4c-a2b807027589" containerID="764ebfdfa1d02b6a8ea75afc9431bf348d397e476a0d32ce5e0d2e93cd5c8748" exitCode=0 Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.156838 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" event={"ID":"5073120c-a1eb-4cf7-8b4c-a2b807027589","Type":"ContainerDied","Data":"764ebfdfa1d02b6a8ea75afc9431bf348d397e476a0d32ce5e0d2e93cd5c8748"} Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.171309 4846 generic.go:334] "Generic (PLEG): container finished" podID="7ea7209b-7df1-4b9a-a666-cd8257cf8088" containerID="8893cbeaad2d8b9be4717112cb07b6e1b8a3d21172730b7a2da7125d168c166e" exitCode=0 Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.171508 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-t6jvf" event={"ID":"7ea7209b-7df1-4b9a-a666-cd8257cf8088","Type":"ContainerDied","Data":"8893cbeaad2d8b9be4717112cb07b6e1b8a3d21172730b7a2da7125d168c166e"} Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.173494 4846 generic.go:334] "Generic (PLEG): container finished" podID="54e66fd9-2451-40c4-b317-783d545e1db9" containerID="37020039b2fa96897ec7fba8404021b59483a98cc8dc8eda772c3424a2bd6350" exitCode=0 Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.173728 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2751-account-create-update-5dsmk" event={"ID":"54e66fd9-2451-40c4-b317-783d545e1db9","Type":"ContainerDied","Data":"37020039b2fa96897ec7fba8404021b59483a98cc8dc8eda772c3424a2bd6350"} Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.572474 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lkmcf" Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.748658 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb99b\" (UniqueName: \"kubernetes.io/projected/f2cf8f76-47bb-4770-908b-723059ba5e13-kube-api-access-zb99b\") pod \"f2cf8f76-47bb-4770-908b-723059ba5e13\" (UID: \"f2cf8f76-47bb-4770-908b-723059ba5e13\") " Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.748810 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2cf8f76-47bb-4770-908b-723059ba5e13-operator-scripts\") pod \"f2cf8f76-47bb-4770-908b-723059ba5e13\" (UID: \"f2cf8f76-47bb-4770-908b-723059ba5e13\") " Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.749460 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2cf8f76-47bb-4770-908b-723059ba5e13-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f2cf8f76-47bb-4770-908b-723059ba5e13" (UID: "f2cf8f76-47bb-4770-908b-723059ba5e13"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.753758 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2cf8f76-47bb-4770-908b-723059ba5e13-kube-api-access-zb99b" (OuterVolumeSpecName: "kube-api-access-zb99b") pod "f2cf8f76-47bb-4770-908b-723059ba5e13" (UID: "f2cf8f76-47bb-4770-908b-723059ba5e13"). InnerVolumeSpecName "kube-api-access-zb99b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.851328 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb99b\" (UniqueName: \"kubernetes.io/projected/f2cf8f76-47bb-4770-908b-723059ba5e13-kube-api-access-zb99b\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:33 crc kubenswrapper[4846]: I0202 13:46:33.851365 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2cf8f76-47bb-4770-908b-723059ba5e13-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.182282 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lkmcf" event={"ID":"f2cf8f76-47bb-4770-908b-723059ba5e13","Type":"ContainerDied","Data":"3951419cab56c3d86cf88d9ee0e57d1f3fbb82c637c827b33ac9dcf8c87ec75a"} Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.182320 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3951419cab56c3d86cf88d9ee0e57d1f3fbb82c637c827b33ac9dcf8c87ec75a" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.182354 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lkmcf" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.538578 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2751-account-create-update-5dsmk" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.671192 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-t6jvf" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.674460 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmbzt\" (UniqueName: \"kubernetes.io/projected/54e66fd9-2451-40c4-b317-783d545e1db9-kube-api-access-gmbzt\") pod \"54e66fd9-2451-40c4-b317-783d545e1db9\" (UID: \"54e66fd9-2451-40c4-b317-783d545e1db9\") " Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.675178 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54e66fd9-2451-40c4-b317-783d545e1db9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54e66fd9-2451-40c4-b317-783d545e1db9" (UID: "54e66fd9-2451-40c4-b317-783d545e1db9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.675733 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54e66fd9-2451-40c4-b317-783d545e1db9-operator-scripts\") pod \"54e66fd9-2451-40c4-b317-783d545e1db9\" (UID: \"54e66fd9-2451-40c4-b317-783d545e1db9\") " Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.675930 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bb45\" (UniqueName: \"kubernetes.io/projected/7ea7209b-7df1-4b9a-a666-cd8257cf8088-kube-api-access-7bb45\") pod \"7ea7209b-7df1-4b9a-a666-cd8257cf8088\" (UID: \"7ea7209b-7df1-4b9a-a666-cd8257cf8088\") " Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.676554 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54e66fd9-2451-40c4-b317-783d545e1db9-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.677890 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54e66fd9-2451-40c4-b317-783d545e1db9-kube-api-access-gmbzt" (OuterVolumeSpecName: "kube-api-access-gmbzt") pod "54e66fd9-2451-40c4-b317-783d545e1db9" (UID: "54e66fd9-2451-40c4-b317-783d545e1db9"). InnerVolumeSpecName "kube-api-access-gmbzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.679391 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ea7209b-7df1-4b9a-a666-cd8257cf8088-kube-api-access-7bb45" (OuterVolumeSpecName: "kube-api-access-7bb45") pod "7ea7209b-7df1-4b9a-a666-cd8257cf8088" (UID: "7ea7209b-7df1-4b9a-a666-cd8257cf8088"). InnerVolumeSpecName "kube-api-access-7bb45". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.686743 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6l9gs" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.752727 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0340-account-create-update-rcslh" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.768906 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.777041 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ea7209b-7df1-4b9a-a666-cd8257cf8088-operator-scripts\") pod \"7ea7209b-7df1-4b9a-a666-cd8257cf8088\" (UID: \"7ea7209b-7df1-4b9a-a666-cd8257cf8088\") " Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.777160 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07556157-fc87-4ff5-9fba-2ee61250b2a5-operator-scripts\") pod \"07556157-fc87-4ff5-9fba-2ee61250b2a5\" (UID: \"07556157-fc87-4ff5-9fba-2ee61250b2a5\") " Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.777236 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrf5n\" (UniqueName: \"kubernetes.io/projected/5073120c-a1eb-4cf7-8b4c-a2b807027589-kube-api-access-jrf5n\") pod \"5073120c-a1eb-4cf7-8b4c-a2b807027589\" (UID: \"5073120c-a1eb-4cf7-8b4c-a2b807027589\") " Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.777340 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wdcz\" (UniqueName: \"kubernetes.io/projected/6d98e27a-a15d-49d4-a2ba-06accdf7770c-kube-api-access-8wdcz\") pod \"6d98e27a-a15d-49d4-a2ba-06accdf7770c\" (UID: \"6d98e27a-a15d-49d4-a2ba-06accdf7770c\") " Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.777394 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5073120c-a1eb-4cf7-8b4c-a2b807027589-operator-scripts\") pod \"5073120c-a1eb-4cf7-8b4c-a2b807027589\" (UID: \"5073120c-a1eb-4cf7-8b4c-a2b807027589\") " Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.778007 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5073120c-a1eb-4cf7-8b4c-a2b807027589-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5073120c-a1eb-4cf7-8b4c-a2b807027589" (UID: "5073120c-a1eb-4cf7-8b4c-a2b807027589"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.778106 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26skt\" (UniqueName: \"kubernetes.io/projected/07556157-fc87-4ff5-9fba-2ee61250b2a5-kube-api-access-26skt\") pod \"07556157-fc87-4ff5-9fba-2ee61250b2a5\" (UID: \"07556157-fc87-4ff5-9fba-2ee61250b2a5\") " Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.778145 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d98e27a-a15d-49d4-a2ba-06accdf7770c-operator-scripts\") pod \"6d98e27a-a15d-49d4-a2ba-06accdf7770c\" (UID: \"6d98e27a-a15d-49d4-a2ba-06accdf7770c\") " Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.778497 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07556157-fc87-4ff5-9fba-2ee61250b2a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "07556157-fc87-4ff5-9fba-2ee61250b2a5" (UID: "07556157-fc87-4ff5-9fba-2ee61250b2a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.778564 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmbzt\" (UniqueName: \"kubernetes.io/projected/54e66fd9-2451-40c4-b317-783d545e1db9-kube-api-access-gmbzt\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.778609 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5073120c-a1eb-4cf7-8b4c-a2b807027589-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.778637 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bb45\" (UniqueName: \"kubernetes.io/projected/7ea7209b-7df1-4b9a-a666-cd8257cf8088-kube-api-access-7bb45\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.778561 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d98e27a-a15d-49d4-a2ba-06accdf7770c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6d98e27a-a15d-49d4-a2ba-06accdf7770c" (UID: "6d98e27a-a15d-49d4-a2ba-06accdf7770c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.777965 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ea7209b-7df1-4b9a-a666-cd8257cf8088-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ea7209b-7df1-4b9a-a666-cd8257cf8088" (UID: "7ea7209b-7df1-4b9a-a666-cd8257cf8088"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.782176 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5073120c-a1eb-4cf7-8b4c-a2b807027589-kube-api-access-jrf5n" (OuterVolumeSpecName: "kube-api-access-jrf5n") pod "5073120c-a1eb-4cf7-8b4c-a2b807027589" (UID: "5073120c-a1eb-4cf7-8b4c-a2b807027589"). InnerVolumeSpecName "kube-api-access-jrf5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.782996 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d98e27a-a15d-49d4-a2ba-06accdf7770c-kube-api-access-8wdcz" (OuterVolumeSpecName: "kube-api-access-8wdcz") pod "6d98e27a-a15d-49d4-a2ba-06accdf7770c" (UID: "6d98e27a-a15d-49d4-a2ba-06accdf7770c"). InnerVolumeSpecName "kube-api-access-8wdcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.789460 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07556157-fc87-4ff5-9fba-2ee61250b2a5-kube-api-access-26skt" (OuterVolumeSpecName: "kube-api-access-26skt") pod "07556157-fc87-4ff5-9fba-2ee61250b2a5" (UID: "07556157-fc87-4ff5-9fba-2ee61250b2a5"). InnerVolumeSpecName "kube-api-access-26skt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.879289 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wdcz\" (UniqueName: \"kubernetes.io/projected/6d98e27a-a15d-49d4-a2ba-06accdf7770c-kube-api-access-8wdcz\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.879340 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26skt\" (UniqueName: \"kubernetes.io/projected/07556157-fc87-4ff5-9fba-2ee61250b2a5-kube-api-access-26skt\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.879350 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d98e27a-a15d-49d4-a2ba-06accdf7770c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.879359 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ea7209b-7df1-4b9a-a666-cd8257cf8088-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.879367 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07556157-fc87-4ff5-9fba-2ee61250b2a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:34 crc kubenswrapper[4846]: I0202 13:46:34.879375 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrf5n\" (UniqueName: \"kubernetes.io/projected/5073120c-a1eb-4cf7-8b4c-a2b807027589-kube-api-access-jrf5n\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.205145 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2751-account-create-update-5dsmk" event={"ID":"54e66fd9-2451-40c4-b317-783d545e1db9","Type":"ContainerDied","Data":"b2ff16d0e09dbdd769131146ae14dd8faf34610e3ce917f55137e13f348568fb"} Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.205214 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2ff16d0e09dbdd769131146ae14dd8faf34610e3ce917f55137e13f348568fb" Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.205367 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2751-account-create-update-5dsmk" Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.209395 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0340-account-create-update-rcslh" event={"ID":"07556157-fc87-4ff5-9fba-2ee61250b2a5","Type":"ContainerDied","Data":"bde9ed7b44432a961213246d09167ed3f93e4af598387c455fe0acd27369465e"} Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.209458 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bde9ed7b44432a961213246d09167ed3f93e4af598387c455fe0acd27369465e" Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.209595 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0340-account-create-update-rcslh" Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.211430 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6l9gs" event={"ID":"6d98e27a-a15d-49d4-a2ba-06accdf7770c","Type":"ContainerDied","Data":"57cbc064fa6bf52db3b469732aa500367cb178e3ab1836ba275bd0075101d7d6"} Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.211471 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57cbc064fa6bf52db3b469732aa500367cb178e3ab1836ba275bd0075101d7d6" Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.211571 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6l9gs" Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.213820 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" event={"ID":"5073120c-a1eb-4cf7-8b4c-a2b807027589","Type":"ContainerDied","Data":"a10deab1b15a241261a006c190f2bd61b16a5d9e7be398921cfbd7ca5fa19547"} Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.213864 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a10deab1b15a241261a006c190f2bd61b16a5d9e7be398921cfbd7ca5fa19547" Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.213960 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-34fb-account-create-update-sdf2h" Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.216011 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-t6jvf" event={"ID":"7ea7209b-7df1-4b9a-a666-cd8257cf8088","Type":"ContainerDied","Data":"e0263f1f67e095fbcb3b43f433b746ac4a6d2a34dbbc0a83fe7a608a9e22fb32"} Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.216059 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0263f1f67e095fbcb3b43f433b746ac4a6d2a34dbbc0a83fe7a608a9e22fb32" Feb 02 13:46:35 crc kubenswrapper[4846]: I0202 13:46:35.216137 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-t6jvf" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.269907 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ck6nm"] Feb 02 13:46:36 crc kubenswrapper[4846]: E0202 13:46:36.270675 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07556157-fc87-4ff5-9fba-2ee61250b2a5" containerName="mariadb-account-create-update" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.270692 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="07556157-fc87-4ff5-9fba-2ee61250b2a5" containerName="mariadb-account-create-update" Feb 02 13:46:36 crc kubenswrapper[4846]: E0202 13:46:36.270709 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea7209b-7df1-4b9a-a666-cd8257cf8088" containerName="mariadb-database-create" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.270717 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea7209b-7df1-4b9a-a666-cd8257cf8088" containerName="mariadb-database-create" Feb 02 13:46:36 crc kubenswrapper[4846]: E0202 13:46:36.270748 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54e66fd9-2451-40c4-b317-783d545e1db9" containerName="mariadb-account-create-update" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.270756 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="54e66fd9-2451-40c4-b317-783d545e1db9" containerName="mariadb-account-create-update" Feb 02 13:46:36 crc kubenswrapper[4846]: E0202 13:46:36.270773 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d98e27a-a15d-49d4-a2ba-06accdf7770c" containerName="mariadb-database-create" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.270783 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d98e27a-a15d-49d4-a2ba-06accdf7770c" containerName="mariadb-database-create" Feb 02 13:46:36 crc kubenswrapper[4846]: E0202 13:46:36.270818 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2cf8f76-47bb-4770-908b-723059ba5e13" containerName="mariadb-database-create" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.270826 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2cf8f76-47bb-4770-908b-723059ba5e13" containerName="mariadb-database-create" Feb 02 13:46:36 crc kubenswrapper[4846]: E0202 13:46:36.270842 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5073120c-a1eb-4cf7-8b4c-a2b807027589" containerName="mariadb-account-create-update" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.270850 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5073120c-a1eb-4cf7-8b4c-a2b807027589" containerName="mariadb-account-create-update" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.271062 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="07556157-fc87-4ff5-9fba-2ee61250b2a5" containerName="mariadb-account-create-update" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.271082 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea7209b-7df1-4b9a-a666-cd8257cf8088" containerName="mariadb-database-create" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.271098 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="54e66fd9-2451-40c4-b317-783d545e1db9" containerName="mariadb-account-create-update" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.271110 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2cf8f76-47bb-4770-908b-723059ba5e13" containerName="mariadb-database-create" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.271129 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d98e27a-a15d-49d4-a2ba-06accdf7770c" containerName="mariadb-database-create" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.271194 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5073120c-a1eb-4cf7-8b4c-a2b807027589" containerName="mariadb-account-create-update" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.271916 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.274275 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.274527 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tvv56" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.274819 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.282315 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ck6nm"] Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.401850 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.401912 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-config-data\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.402051 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b66qs\" (UniqueName: \"kubernetes.io/projected/2e8fb480-8530-42db-86ce-35326f86b8da-kube-api-access-b66qs\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.402093 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-scripts\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.503436 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.503485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-config-data\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.503557 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b66qs\" (UniqueName: \"kubernetes.io/projected/2e8fb480-8530-42db-86ce-35326f86b8da-kube-api-access-b66qs\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.503585 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-scripts\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.510228 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-config-data\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.515246 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-scripts\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.515332 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.529070 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b66qs\" (UniqueName: \"kubernetes.io/projected/2e8fb480-8530-42db-86ce-35326f86b8da-kube-api-access-b66qs\") pod \"nova-cell0-conductor-db-sync-ck6nm\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:36 crc kubenswrapper[4846]: I0202 13:46:36.600293 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:37 crc kubenswrapper[4846]: I0202 13:46:37.037373 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ck6nm"] Feb 02 13:46:37 crc kubenswrapper[4846]: W0202 13:46:37.045524 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e8fb480_8530_42db_86ce_35326f86b8da.slice/crio-df2b49b2d5472c9e0e44004f218b06bf6f236e8933f8a677d412c25d4fed9dbc WatchSource:0}: Error finding container df2b49b2d5472c9e0e44004f218b06bf6f236e8933f8a677d412c25d4fed9dbc: Status 404 returned error can't find the container with id df2b49b2d5472c9e0e44004f218b06bf6f236e8933f8a677d412c25d4fed9dbc Feb 02 13:46:37 crc kubenswrapper[4846]: I0202 13:46:37.244466 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ck6nm" event={"ID":"2e8fb480-8530-42db-86ce-35326f86b8da","Type":"ContainerStarted","Data":"7fd22820549e8a0ab310981d03f95edd3178f01653a9cc5680b5fbffd915d8b2"} Feb 02 13:46:37 crc kubenswrapper[4846]: I0202 13:46:37.247337 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ck6nm" event={"ID":"2e8fb480-8530-42db-86ce-35326f86b8da","Type":"ContainerStarted","Data":"df2b49b2d5472c9e0e44004f218b06bf6f236e8933f8a677d412c25d4fed9dbc"} Feb 02 13:46:37 crc kubenswrapper[4846]: I0202 13:46:37.261712 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-ck6nm" podStartSLOduration=1.261692394 podStartE2EDuration="1.261692394s" podCreationTimestamp="2026-02-02 13:46:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:37.257744908 +0000 UTC m=+5828.486331791" watchObservedRunningTime="2026-02-02 13:46:37.261692394 +0000 UTC m=+5828.490279257" Feb 02 13:46:38 crc kubenswrapper[4846]: I0202 13:46:38.423990 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:46:38 crc kubenswrapper[4846]: E0202 13:46:38.424231 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:46:38 crc kubenswrapper[4846]: I0202 13:46:38.869338 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-68fc8c869-kkbb8" podUID="9c5ac153-1fd9-491b-942d-2b0b3e3e9da5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.90:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 13:46:42 crc kubenswrapper[4846]: I0202 13:46:42.287384 4846 generic.go:334] "Generic (PLEG): container finished" podID="2e8fb480-8530-42db-86ce-35326f86b8da" containerID="7fd22820549e8a0ab310981d03f95edd3178f01653a9cc5680b5fbffd915d8b2" exitCode=0 Feb 02 13:46:42 crc kubenswrapper[4846]: I0202 13:46:42.287470 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ck6nm" event={"ID":"2e8fb480-8530-42db-86ce-35326f86b8da","Type":"ContainerDied","Data":"7fd22820549e8a0ab310981d03f95edd3178f01653a9cc5680b5fbffd915d8b2"} Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.655336 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.841627 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-scripts\") pod \"2e8fb480-8530-42db-86ce-35326f86b8da\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.841814 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-config-data\") pod \"2e8fb480-8530-42db-86ce-35326f86b8da\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.841883 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b66qs\" (UniqueName: \"kubernetes.io/projected/2e8fb480-8530-42db-86ce-35326f86b8da-kube-api-access-b66qs\") pod \"2e8fb480-8530-42db-86ce-35326f86b8da\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.842822 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-combined-ca-bundle\") pod \"2e8fb480-8530-42db-86ce-35326f86b8da\" (UID: \"2e8fb480-8530-42db-86ce-35326f86b8da\") " Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.848177 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-scripts" (OuterVolumeSpecName: "scripts") pod "2e8fb480-8530-42db-86ce-35326f86b8da" (UID: "2e8fb480-8530-42db-86ce-35326f86b8da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.852951 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e8fb480-8530-42db-86ce-35326f86b8da-kube-api-access-b66qs" (OuterVolumeSpecName: "kube-api-access-b66qs") pod "2e8fb480-8530-42db-86ce-35326f86b8da" (UID: "2e8fb480-8530-42db-86ce-35326f86b8da"). InnerVolumeSpecName "kube-api-access-b66qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.873807 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e8fb480-8530-42db-86ce-35326f86b8da" (UID: "2e8fb480-8530-42db-86ce-35326f86b8da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.874537 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-config-data" (OuterVolumeSpecName: "config-data") pod "2e8fb480-8530-42db-86ce-35326f86b8da" (UID: "2e8fb480-8530-42db-86ce-35326f86b8da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.945839 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.945887 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b66qs\" (UniqueName: \"kubernetes.io/projected/2e8fb480-8530-42db-86ce-35326f86b8da-kube-api-access-b66qs\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.945903 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:43 crc kubenswrapper[4846]: I0202 13:46:43.945919 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e8fb480-8530-42db-86ce-35326f86b8da-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.304075 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ck6nm" event={"ID":"2e8fb480-8530-42db-86ce-35326f86b8da","Type":"ContainerDied","Data":"df2b49b2d5472c9e0e44004f218b06bf6f236e8933f8a677d412c25d4fed9dbc"} Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.304112 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df2b49b2d5472c9e0e44004f218b06bf6f236e8933f8a677d412c25d4fed9dbc" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.304137 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ck6nm" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.371940 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 13:46:44 crc kubenswrapper[4846]: E0202 13:46:44.372276 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e8fb480-8530-42db-86ce-35326f86b8da" containerName="nova-cell0-conductor-db-sync" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.372291 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e8fb480-8530-42db-86ce-35326f86b8da" containerName="nova-cell0-conductor-db-sync" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.372457 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e8fb480-8530-42db-86ce-35326f86b8da" containerName="nova-cell0-conductor-db-sync" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.373000 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.375650 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.375807 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tvv56" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.396958 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.455610 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.455710 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96n7m\" (UniqueName: \"kubernetes.io/projected/4b608189-194f-4aae-bd70-9c6ea33f2d44-kube-api-access-96n7m\") pod \"nova-cell0-conductor-0\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.455739 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.557269 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.557327 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96n7m\" (UniqueName: \"kubernetes.io/projected/4b608189-194f-4aae-bd70-9c6ea33f2d44-kube-api-access-96n7m\") pod \"nova-cell0-conductor-0\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.557354 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.561192 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.566491 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.573322 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96n7m\" (UniqueName: \"kubernetes.io/projected/4b608189-194f-4aae-bd70-9c6ea33f2d44-kube-api-access-96n7m\") pod \"nova-cell0-conductor-0\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:44 crc kubenswrapper[4846]: I0202 13:46:44.694367 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:45 crc kubenswrapper[4846]: I0202 13:46:45.127685 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 13:46:45 crc kubenswrapper[4846]: I0202 13:46:45.313822 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4b608189-194f-4aae-bd70-9c6ea33f2d44","Type":"ContainerStarted","Data":"143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb"} Feb 02 13:46:45 crc kubenswrapper[4846]: I0202 13:46:45.313866 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4b608189-194f-4aae-bd70-9c6ea33f2d44","Type":"ContainerStarted","Data":"16383ed8055255e688a36e31acf8ed4f437f6c216c4564fb304a79a9c67c39e2"} Feb 02 13:46:45 crc kubenswrapper[4846]: I0202 13:46:45.313900 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:45 crc kubenswrapper[4846]: I0202 13:46:45.339875 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.339856356 podStartE2EDuration="1.339856356s" podCreationTimestamp="2026-02-02 13:46:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:45.330180621 +0000 UTC m=+5836.558767534" watchObservedRunningTime="2026-02-02 13:46:45.339856356 +0000 UTC m=+5836.568443219" Feb 02 13:46:50 crc kubenswrapper[4846]: I0202 13:46:50.424460 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:46:50 crc kubenswrapper[4846]: E0202 13:46:50.425798 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:46:54 crc kubenswrapper[4846]: I0202 13:46:54.727840 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.205667 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-hgf4w"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.207024 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.215100 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.215140 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.224590 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-hgf4w"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.354208 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.355616 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.357767 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.366107 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcmpt\" (UniqueName: \"kubernetes.io/projected/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-kube-api-access-zcmpt\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.366177 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.366552 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-config-data\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.366775 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-scripts\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.378541 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.379715 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.386847 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.398507 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.412845 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.457720 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.459293 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.461816 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.469207 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-scripts\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.469267 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.469303 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcmpt\" (UniqueName: \"kubernetes.io/projected/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-kube-api-access-zcmpt\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.469337 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.469365 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhxn8\" (UniqueName: \"kubernetes.io/projected/dd46fafd-285b-43a0-a507-99825152e830-kube-api-access-zhxn8\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.469409 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.469445 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-config-data\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.475313 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.476105 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-scripts\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.479279 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.480047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-config-data\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.512147 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcmpt\" (UniqueName: \"kubernetes.io/projected/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-kube-api-access-zcmpt\") pod \"nova-cell0-cell-mapping-hgf4w\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.529138 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.556066 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.558167 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.561849 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.571992 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " pod="openstack/nova-scheduler-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.572043 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77ztj\" (UniqueName: \"kubernetes.io/projected/7f189345-f382-474e-b3bb-721f7cdbcfab-kube-api-access-77ztj\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.572093 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s8s5\" (UniqueName: \"kubernetes.io/projected/ce616c5c-5881-4cfc-934d-b1c789d76e46-kube-api-access-9s8s5\") pod \"nova-scheduler-0\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " pod="openstack/nova-scheduler-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.572112 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-config-data\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.572139 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.572195 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhxn8\" (UniqueName: \"kubernetes.io/projected/dd46fafd-285b-43a0-a507-99825152e830-kube-api-access-zhxn8\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.572233 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.572251 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-config-data\") pod \"nova-scheduler-0\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " pod="openstack/nova-scheduler-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.572274 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f189345-f382-474e-b3bb-721f7cdbcfab-logs\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.572301 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.579440 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.579950 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.597776 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.608359 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhxn8\" (UniqueName: \"kubernetes.io/projected/dd46fafd-285b-43a0-a507-99825152e830-kube-api-access-zhxn8\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.613260 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8555c84f6c-52q4p"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.619658 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.650441 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8555c84f6c-52q4p"] Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.671284 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.674290 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmssg\" (UniqueName: \"kubernetes.io/projected/a3b9618d-8751-4770-817d-9db761074535-kube-api-access-jmssg\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.674336 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3b9618d-8751-4770-817d-9db761074535-logs\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.674371 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s8s5\" (UniqueName: \"kubernetes.io/projected/ce616c5c-5881-4cfc-934d-b1c789d76e46-kube-api-access-9s8s5\") pod \"nova-scheduler-0\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " pod="openstack/nova-scheduler-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.674391 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-config-data\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.674851 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-config-data\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.678306 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-config-data\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.681179 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.681261 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-config-data\") pod \"nova-scheduler-0\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " pod="openstack/nova-scheduler-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.681321 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f189345-f382-474e-b3bb-721f7cdbcfab-logs\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.681370 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.681409 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " pod="openstack/nova-scheduler-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.681447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77ztj\" (UniqueName: \"kubernetes.io/projected/7f189345-f382-474e-b3bb-721f7cdbcfab-kube-api-access-77ztj\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.688223 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f189345-f382-474e-b3bb-721f7cdbcfab-logs\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.688903 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-config-data\") pod \"nova-scheduler-0\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " pod="openstack/nova-scheduler-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.690400 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " pod="openstack/nova-scheduler-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.691301 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.698199 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s8s5\" (UniqueName: \"kubernetes.io/projected/ce616c5c-5881-4cfc-934d-b1c789d76e46-kube-api-access-9s8s5\") pod \"nova-scheduler-0\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " pod="openstack/nova-scheduler-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.712024 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77ztj\" (UniqueName: \"kubernetes.io/projected/7f189345-f382-474e-b3bb-721f7cdbcfab-kube-api-access-77ztj\") pod \"nova-metadata-0\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " pod="openstack/nova-metadata-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.783893 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-nb\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.783968 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-config\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.784157 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-config-data\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.784244 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.784275 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-dns-svc\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.784317 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8ltz\" (UniqueName: \"kubernetes.io/projected/db6600ab-c487-4e40-ace0-f51e528c46af-kube-api-access-j8ltz\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.784380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmssg\" (UniqueName: \"kubernetes.io/projected/a3b9618d-8751-4770-817d-9db761074535-kube-api-access-jmssg\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.784398 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-sb\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.784416 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3b9618d-8751-4770-817d-9db761074535-logs\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.784838 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3b9618d-8751-4770-817d-9db761074535-logs\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.792269 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.792598 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-config-data\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.809079 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmssg\" (UniqueName: \"kubernetes.io/projected/a3b9618d-8751-4770-817d-9db761074535-kube-api-access-jmssg\") pod \"nova-api-0\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " pod="openstack/nova-api-0" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.886691 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-dns-svc\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.886791 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8ltz\" (UniqueName: \"kubernetes.io/projected/db6600ab-c487-4e40-ace0-f51e528c46af-kube-api-access-j8ltz\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.886862 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-sb\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.886908 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-nb\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.886949 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-config\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.887859 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-dns-svc\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.888019 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-sb\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.888114 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-config\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.888145 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-nb\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:55 crc kubenswrapper[4846]: I0202 13:46:55.908557 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8ltz\" (UniqueName: \"kubernetes.io/projected/db6600ab-c487-4e40-ace0-f51e528c46af-kube-api-access-j8ltz\") pod \"dnsmasq-dns-8555c84f6c-52q4p\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.009853 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.010166 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.013093 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.030718 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.073570 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-hgf4w"] Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.164537 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-t45x7"] Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.170847 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.174996 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.190979 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 02 13:46:56 crc kubenswrapper[4846]: W0202 13:46:56.210801 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd46fafd_285b_43a0_a507_99825152e830.slice/crio-334e4251ec2e7b4a841f0479eb318ac92a39438f7ca7af4964e8e1b78fb44391 WatchSource:0}: Error finding container 334e4251ec2e7b4a841f0479eb318ac92a39438f7ca7af4964e8e1b78fb44391: Status 404 returned error can't find the container with id 334e4251ec2e7b4a841f0479eb318ac92a39438f7ca7af4964e8e1b78fb44391 Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.212130 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-t45x7"] Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.214610 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.214785 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-scripts\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.214828 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zfkc\" (UniqueName: \"kubernetes.io/projected/587b6a35-3ed8-4a91-8c38-daefcb8a244e-kube-api-access-4zfkc\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.214868 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-config-data\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.223178 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.316485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-scripts\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.316541 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zfkc\" (UniqueName: \"kubernetes.io/projected/587b6a35-3ed8-4a91-8c38-daefcb8a244e-kube-api-access-4zfkc\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.316572 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-config-data\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.316650 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.322616 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-config-data\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.323870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.325489 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-scripts\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.334951 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zfkc\" (UniqueName: \"kubernetes.io/projected/587b6a35-3ed8-4a91-8c38-daefcb8a244e-kube-api-access-4zfkc\") pod \"nova-cell1-conductor-db-sync-t45x7\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.451513 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hgf4w" event={"ID":"8904d23d-2e73-4eb8-b6f6-8d559d0365e6","Type":"ContainerStarted","Data":"8d1edba65159d50673e4e593a3f7a608fd860b9d5b8b72df9c25acb668c97010"} Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.451568 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hgf4w" event={"ID":"8904d23d-2e73-4eb8-b6f6-8d559d0365e6","Type":"ContainerStarted","Data":"4df91dc4ad44655a8bc6950564dc9cdbe6f66807edccf4e44edf49aed0dc1439"} Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.459037 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd46fafd-285b-43a0-a507-99825152e830","Type":"ContainerStarted","Data":"6dc65fd6eb95788dfcdc64c8e640e201ac85911b7c67a230691d8f0e6824e81f"} Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.459080 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd46fafd-285b-43a0-a507-99825152e830","Type":"ContainerStarted","Data":"334e4251ec2e7b4a841f0479eb318ac92a39438f7ca7af4964e8e1b78fb44391"} Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.479988 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-hgf4w" podStartSLOduration=1.479968821 podStartE2EDuration="1.479968821s" podCreationTimestamp="2026-02-02 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:56.474645552 +0000 UTC m=+5847.703232425" watchObservedRunningTime="2026-02-02 13:46:56.479968821 +0000 UTC m=+5847.708555684" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.506113 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.523413 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.5233930359999999 podStartE2EDuration="1.523393036s" podCreationTimestamp="2026-02-02 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:56.49807305 +0000 UTC m=+5847.726659913" watchObservedRunningTime="2026-02-02 13:46:56.523393036 +0000 UTC m=+5847.751979899" Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.594034 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.713695 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.723024 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:46:56 crc kubenswrapper[4846]: I0202 13:46:56.730407 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8555c84f6c-52q4p"] Feb 02 13:46:56 crc kubenswrapper[4846]: W0202 13:46:56.755848 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3b9618d_8751_4770_817d_9db761074535.slice/crio-414f330ff5797f6d145669d1168e60a056f8dfed937421ac39267ff06735e2b4 WatchSource:0}: Error finding container 414f330ff5797f6d145669d1168e60a056f8dfed937421ac39267ff06735e2b4: Status 404 returned error can't find the container with id 414f330ff5797f6d145669d1168e60a056f8dfed937421ac39267ff06735e2b4 Feb 02 13:46:56 crc kubenswrapper[4846]: W0202 13:46:56.758131 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce616c5c_5881_4cfc_934d_b1c789d76e46.slice/crio-1791a921e29367f6e09c5f6cf09ec5fb5bb6a2b2128a6248b4f1d90c65ea2abb WatchSource:0}: Error finding container 1791a921e29367f6e09c5f6cf09ec5fb5bb6a2b2128a6248b4f1d90c65ea2abb: Status 404 returned error can't find the container with id 1791a921e29367f6e09c5f6cf09ec5fb5bb6a2b2128a6248b4f1d90c65ea2abb Feb 02 13:46:56 crc kubenswrapper[4846]: W0202 13:46:56.778691 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb6600ab_c487_4e40_ace0_f51e528c46af.slice/crio-8d8cc5e042e621946ff9993859f09a2ce32692e95f79ee89d4081c08b09acc69 WatchSource:0}: Error finding container 8d8cc5e042e621946ff9993859f09a2ce32692e95f79ee89d4081c08b09acc69: Status 404 returned error can't find the container with id 8d8cc5e042e621946ff9993859f09a2ce32692e95f79ee89d4081c08b09acc69 Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.083236 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-t45x7"] Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.473975 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7f189345-f382-474e-b3bb-721f7cdbcfab","Type":"ContainerStarted","Data":"04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.474215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7f189345-f382-474e-b3bb-721f7cdbcfab","Type":"ContainerStarted","Data":"7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.474225 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7f189345-f382-474e-b3bb-721f7cdbcfab","Type":"ContainerStarted","Data":"64714a833fb1f932884b1d3effa02645fad6a433aac2857380f2b90004078cb1"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.491011 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce616c5c-5881-4cfc-934d-b1c789d76e46","Type":"ContainerStarted","Data":"57e7b6c791e1736f567e02efac024b4da6aebcc1959db3ab46ca55884c16cded"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.491071 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce616c5c-5881-4cfc-934d-b1c789d76e46","Type":"ContainerStarted","Data":"1791a921e29367f6e09c5f6cf09ec5fb5bb6a2b2128a6248b4f1d90c65ea2abb"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.496178 4846 generic.go:334] "Generic (PLEG): container finished" podID="db6600ab-c487-4e40-ace0-f51e528c46af" containerID="0cb523451a149729826966a1d66a091f3787fc72dad8c7e597763e0070bb447c" exitCode=0 Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.496269 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" event={"ID":"db6600ab-c487-4e40-ace0-f51e528c46af","Type":"ContainerDied","Data":"0cb523451a149729826966a1d66a091f3787fc72dad8c7e597763e0070bb447c"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.496295 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" event={"ID":"db6600ab-c487-4e40-ace0-f51e528c46af","Type":"ContainerStarted","Data":"8d8cc5e042e621946ff9993859f09a2ce32692e95f79ee89d4081c08b09acc69"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.511024 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-t45x7" event={"ID":"587b6a35-3ed8-4a91-8c38-daefcb8a244e","Type":"ContainerStarted","Data":"f15402751c443c63c555370f43747b6463389e32f6f111ab8230b1c238f9f2e2"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.511090 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-t45x7" event={"ID":"587b6a35-3ed8-4a91-8c38-daefcb8a244e","Type":"ContainerStarted","Data":"dc723a5b95d2b53c724b6990a9d8f5123ddeefb5ae2af0c9930dcabeb6b47712"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.520430 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.520412669 podStartE2EDuration="2.520412669s" podCreationTimestamp="2026-02-02 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:57.503733314 +0000 UTC m=+5848.732320207" watchObservedRunningTime="2026-02-02 13:46:57.520412669 +0000 UTC m=+5848.748999532" Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.534350 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.534326817 podStartE2EDuration="2.534326817s" podCreationTimestamp="2026-02-02 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:57.526593979 +0000 UTC m=+5848.755180852" watchObservedRunningTime="2026-02-02 13:46:57.534326817 +0000 UTC m=+5848.762913680" Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.550455 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3b9618d-8751-4770-817d-9db761074535","Type":"ContainerStarted","Data":"3fb804d856b1de1cc7cb165564d4afc3349b7de29a7f2728c659fc710010e2a8"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.550504 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3b9618d-8751-4770-817d-9db761074535","Type":"ContainerStarted","Data":"ebb01c224316779e1edabea50d37cc50fd49dc66d7798e3b8c4e9680f8fb51fa"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.550514 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3b9618d-8751-4770-817d-9db761074535","Type":"ContainerStarted","Data":"414f330ff5797f6d145669d1168e60a056f8dfed937421ac39267ff06735e2b4"} Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.577864 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-t45x7" podStartSLOduration=1.577840964 podStartE2EDuration="1.577840964s" podCreationTimestamp="2026-02-02 13:46:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:57.571850227 +0000 UTC m=+5848.800437100" watchObservedRunningTime="2026-02-02 13:46:57.577840964 +0000 UTC m=+5848.806427827" Feb 02 13:46:57 crc kubenswrapper[4846]: I0202 13:46:57.607498 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.607470303 podStartE2EDuration="2.607470303s" podCreationTimestamp="2026-02-02 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:57.599077999 +0000 UTC m=+5848.827664872" watchObservedRunningTime="2026-02-02 13:46:57.607470303 +0000 UTC m=+5848.836057186" Feb 02 13:46:58 crc kubenswrapper[4846]: I0202 13:46:58.560888 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" event={"ID":"db6600ab-c487-4e40-ace0-f51e528c46af","Type":"ContainerStarted","Data":"f9c86bdd27617ceea4f5372425c2905824471788c126325fedd4cf403b48495a"} Feb 02 13:46:58 crc kubenswrapper[4846]: I0202 13:46:58.587245 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" podStartSLOduration=3.587223317 podStartE2EDuration="3.587223317s" podCreationTimestamp="2026-02-02 13:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:46:58.584120621 +0000 UTC m=+5849.812707484" watchObservedRunningTime="2026-02-02 13:46:58.587223317 +0000 UTC m=+5849.815810190" Feb 02 13:46:59 crc kubenswrapper[4846]: I0202 13:46:59.569426 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:46:59 crc kubenswrapper[4846]: I0202 13:46:59.876027 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:46:59 crc kubenswrapper[4846]: I0202 13:46:59.876290 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7f189345-f382-474e-b3bb-721f7cdbcfab" containerName="nova-metadata-log" containerID="cri-o://7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b" gracePeriod=30 Feb 02 13:46:59 crc kubenswrapper[4846]: I0202 13:46:59.876390 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7f189345-f382-474e-b3bb-721f7cdbcfab" containerName="nova-metadata-metadata" containerID="cri-o://04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b" gracePeriod=30 Feb 02 13:46:59 crc kubenswrapper[4846]: I0202 13:46:59.899708 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 13:46:59 crc kubenswrapper[4846]: I0202 13:46:59.900033 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="dd46fafd-285b-43a0-a507-99825152e830" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://6dc65fd6eb95788dfcdc64c8e640e201ac85911b7c67a230691d8f0e6824e81f" gracePeriod=30 Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.548245 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.587218 4846 generic.go:334] "Generic (PLEG): container finished" podID="7f189345-f382-474e-b3bb-721f7cdbcfab" containerID="04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b" exitCode=0 Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.587247 4846 generic.go:334] "Generic (PLEG): container finished" podID="7f189345-f382-474e-b3bb-721f7cdbcfab" containerID="7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b" exitCode=143 Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.587320 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.589099 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7f189345-f382-474e-b3bb-721f7cdbcfab","Type":"ContainerDied","Data":"04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b"} Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.590245 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7f189345-f382-474e-b3bb-721f7cdbcfab","Type":"ContainerDied","Data":"7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b"} Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.603173 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7f189345-f382-474e-b3bb-721f7cdbcfab","Type":"ContainerDied","Data":"64714a833fb1f932884b1d3effa02645fad6a433aac2857380f2b90004078cb1"} Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.603287 4846 scope.go:117] "RemoveContainer" containerID="04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.605973 4846 generic.go:334] "Generic (PLEG): container finished" podID="587b6a35-3ed8-4a91-8c38-daefcb8a244e" containerID="f15402751c443c63c555370f43747b6463389e32f6f111ab8230b1c238f9f2e2" exitCode=0 Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.606047 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-t45x7" event={"ID":"587b6a35-3ed8-4a91-8c38-daefcb8a244e","Type":"ContainerDied","Data":"f15402751c443c63c555370f43747b6463389e32f6f111ab8230b1c238f9f2e2"} Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.613773 4846 generic.go:334] "Generic (PLEG): container finished" podID="dd46fafd-285b-43a0-a507-99825152e830" containerID="6dc65fd6eb95788dfcdc64c8e640e201ac85911b7c67a230691d8f0e6824e81f" exitCode=0 Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.614856 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd46fafd-285b-43a0-a507-99825152e830","Type":"ContainerDied","Data":"6dc65fd6eb95788dfcdc64c8e640e201ac85911b7c67a230691d8f0e6824e81f"} Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.633966 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-combined-ca-bundle\") pod \"7f189345-f382-474e-b3bb-721f7cdbcfab\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.634067 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-config-data\") pod \"7f189345-f382-474e-b3bb-721f7cdbcfab\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.634159 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77ztj\" (UniqueName: \"kubernetes.io/projected/7f189345-f382-474e-b3bb-721f7cdbcfab-kube-api-access-77ztj\") pod \"7f189345-f382-474e-b3bb-721f7cdbcfab\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.634236 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f189345-f382-474e-b3bb-721f7cdbcfab-logs\") pod \"7f189345-f382-474e-b3bb-721f7cdbcfab\" (UID: \"7f189345-f382-474e-b3bb-721f7cdbcfab\") " Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.646004 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f189345-f382-474e-b3bb-721f7cdbcfab-kube-api-access-77ztj" (OuterVolumeSpecName: "kube-api-access-77ztj") pod "7f189345-f382-474e-b3bb-721f7cdbcfab" (UID: "7f189345-f382-474e-b3bb-721f7cdbcfab"). InnerVolumeSpecName "kube-api-access-77ztj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.650405 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f189345-f382-474e-b3bb-721f7cdbcfab-logs" (OuterVolumeSpecName: "logs") pod "7f189345-f382-474e-b3bb-721f7cdbcfab" (UID: "7f189345-f382-474e-b3bb-721f7cdbcfab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.669101 4846 scope.go:117] "RemoveContainer" containerID="7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.669174 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f189345-f382-474e-b3bb-721f7cdbcfab" (UID: "7f189345-f382-474e-b3bb-721f7cdbcfab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.669256 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-config-data" (OuterVolumeSpecName: "config-data") pod "7f189345-f382-474e-b3bb-721f7cdbcfab" (UID: "7f189345-f382-474e-b3bb-721f7cdbcfab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.672367 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.735769 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.735820 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f189345-f382-474e-b3bb-721f7cdbcfab-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.735829 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77ztj\" (UniqueName: \"kubernetes.io/projected/7f189345-f382-474e-b3bb-721f7cdbcfab-kube-api-access-77ztj\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.735839 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f189345-f382-474e-b3bb-721f7cdbcfab-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.769266 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.784788 4846 scope.go:117] "RemoveContainer" containerID="04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b" Feb 02 13:47:00 crc kubenswrapper[4846]: E0202 13:47:00.788732 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b\": container with ID starting with 04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b not found: ID does not exist" containerID="04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.788781 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b"} err="failed to get container status \"04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b\": rpc error: code = NotFound desc = could not find container \"04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b\": container with ID starting with 04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b not found: ID does not exist" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.788811 4846 scope.go:117] "RemoveContainer" containerID="7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b" Feb 02 13:47:00 crc kubenswrapper[4846]: E0202 13:47:00.789444 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b\": container with ID starting with 7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b not found: ID does not exist" containerID="7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.789505 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b"} err="failed to get container status \"7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b\": rpc error: code = NotFound desc = could not find container \"7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b\": container with ID starting with 7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b not found: ID does not exist" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.789539 4846 scope.go:117] "RemoveContainer" containerID="04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.790083 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b"} err="failed to get container status \"04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b\": rpc error: code = NotFound desc = could not find container \"04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b\": container with ID starting with 04c64864fa39e47e3cdf55a57c61be341c3283a2c24e87ba9bffb5357134948b not found: ID does not exist" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.790114 4846 scope.go:117] "RemoveContainer" containerID="7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.790487 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b"} err="failed to get container status \"7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b\": rpc error: code = NotFound desc = could not find container \"7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b\": container with ID starting with 7bd63f0a59754495812b53f75fe625a9ee278c1c5e6acff2cf638041627cb76b not found: ID does not exist" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.836936 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhxn8\" (UniqueName: \"kubernetes.io/projected/dd46fafd-285b-43a0-a507-99825152e830-kube-api-access-zhxn8\") pod \"dd46fafd-285b-43a0-a507-99825152e830\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.837200 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-combined-ca-bundle\") pod \"dd46fafd-285b-43a0-a507-99825152e830\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.837309 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-config-data\") pod \"dd46fafd-285b-43a0-a507-99825152e830\" (UID: \"dd46fafd-285b-43a0-a507-99825152e830\") " Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.841685 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd46fafd-285b-43a0-a507-99825152e830-kube-api-access-zhxn8" (OuterVolumeSpecName: "kube-api-access-zhxn8") pod "dd46fafd-285b-43a0-a507-99825152e830" (UID: "dd46fafd-285b-43a0-a507-99825152e830"). InnerVolumeSpecName "kube-api-access-zhxn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.865874 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-config-data" (OuterVolumeSpecName: "config-data") pod "dd46fafd-285b-43a0-a507-99825152e830" (UID: "dd46fafd-285b-43a0-a507-99825152e830"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.865906 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd46fafd-285b-43a0-a507-99825152e830" (UID: "dd46fafd-285b-43a0-a507-99825152e830"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.932748 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.939325 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.939524 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd46fafd-285b-43a0-a507-99825152e830-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.939654 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhxn8\" (UniqueName: \"kubernetes.io/projected/dd46fafd-285b-43a0-a507-99825152e830-kube-api-access-zhxn8\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.947836 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.963235 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:00 crc kubenswrapper[4846]: E0202 13:47:00.963675 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f189345-f382-474e-b3bb-721f7cdbcfab" containerName="nova-metadata-metadata" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.963691 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f189345-f382-474e-b3bb-721f7cdbcfab" containerName="nova-metadata-metadata" Feb 02 13:47:00 crc kubenswrapper[4846]: E0202 13:47:00.963708 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd46fafd-285b-43a0-a507-99825152e830" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.963715 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd46fafd-285b-43a0-a507-99825152e830" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 13:47:00 crc kubenswrapper[4846]: E0202 13:47:00.963730 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f189345-f382-474e-b3bb-721f7cdbcfab" containerName="nova-metadata-log" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.963736 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f189345-f382-474e-b3bb-721f7cdbcfab" containerName="nova-metadata-log" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.963911 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f189345-f382-474e-b3bb-721f7cdbcfab" containerName="nova-metadata-metadata" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.963926 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f189345-f382-474e-b3bb-721f7cdbcfab" containerName="nova-metadata-log" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.963944 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd46fafd-285b-43a0-a507-99825152e830" containerName="nova-cell1-novncproxy-novncproxy" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.965070 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.973091 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.978198 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 02 13:47:00 crc kubenswrapper[4846]: I0202 13:47:00.978503 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.010849 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.041057 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-config-data\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.041344 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-logs\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.041509 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.041728 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.041932 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mn9b\" (UniqueName: \"kubernetes.io/projected/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-kube-api-access-4mn9b\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.142812 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.142851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.142914 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mn9b\" (UniqueName: \"kubernetes.io/projected/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-kube-api-access-4mn9b\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.142974 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-config-data\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.143007 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-logs\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.143422 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-logs\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.147374 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.148484 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.149858 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-config-data\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.162474 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mn9b\" (UniqueName: \"kubernetes.io/projected/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-kube-api-access-4mn9b\") pod \"nova-metadata-0\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.352506 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.452639 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f189345-f382-474e-b3bb-721f7cdbcfab" path="/var/lib/kubelet/pods/7f189345-f382-474e-b3bb-721f7cdbcfab/volumes" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.626493 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.626502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd46fafd-285b-43a0-a507-99825152e830","Type":"ContainerDied","Data":"334e4251ec2e7b4a841f0479eb318ac92a39438f7ca7af4964e8e1b78fb44391"} Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.626587 4846 scope.go:117] "RemoveContainer" containerID="6dc65fd6eb95788dfcdc64c8e640e201ac85911b7c67a230691d8f0e6824e81f" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.666362 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.685330 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.699591 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.700711 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.706908 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.707110 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.707319 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.707370 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.756949 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.757002 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.757051 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4vsc\" (UniqueName: \"kubernetes.io/projected/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-kube-api-access-m4vsc\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.757172 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.757204 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.807614 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:01 crc kubenswrapper[4846]: W0202 13:47:01.832574 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa7a808a_a347_4cd1_8f12_c8a1d9e7e87e.slice/crio-e1ded89aeaa541bf79155d68459e0fc55f94edd47352ad0f72204eb839a1e810 WatchSource:0}: Error finding container e1ded89aeaa541bf79155d68459e0fc55f94edd47352ad0f72204eb839a1e810: Status 404 returned error can't find the container with id e1ded89aeaa541bf79155d68459e0fc55f94edd47352ad0f72204eb839a1e810 Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.860372 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.860430 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.860560 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.860593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.860653 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4vsc\" (UniqueName: \"kubernetes.io/projected/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-kube-api-access-m4vsc\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.868706 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.868774 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.869553 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.878905 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.888977 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4vsc\" (UniqueName: \"kubernetes.io/projected/e611f1f1-65e6-4ce4-898b-a16e4bf0ee61-kube-api-access-m4vsc\") pod \"nova-cell1-novncproxy-0\" (UID: \"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61\") " pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:01 crc kubenswrapper[4846]: I0202 13:47:01.952418 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.019206 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.065122 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-combined-ca-bundle\") pod \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.065226 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-config-data\") pod \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.065246 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-scripts\") pod \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.065275 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zfkc\" (UniqueName: \"kubernetes.io/projected/587b6a35-3ed8-4a91-8c38-daefcb8a244e-kube-api-access-4zfkc\") pod \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\" (UID: \"587b6a35-3ed8-4a91-8c38-daefcb8a244e\") " Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.069581 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/587b6a35-3ed8-4a91-8c38-daefcb8a244e-kube-api-access-4zfkc" (OuterVolumeSpecName: "kube-api-access-4zfkc") pod "587b6a35-3ed8-4a91-8c38-daefcb8a244e" (UID: "587b6a35-3ed8-4a91-8c38-daefcb8a244e"). InnerVolumeSpecName "kube-api-access-4zfkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.070084 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-scripts" (OuterVolumeSpecName: "scripts") pod "587b6a35-3ed8-4a91-8c38-daefcb8a244e" (UID: "587b6a35-3ed8-4a91-8c38-daefcb8a244e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.097573 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "587b6a35-3ed8-4a91-8c38-daefcb8a244e" (UID: "587b6a35-3ed8-4a91-8c38-daefcb8a244e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.098014 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-config-data" (OuterVolumeSpecName: "config-data") pod "587b6a35-3ed8-4a91-8c38-daefcb8a244e" (UID: "587b6a35-3ed8-4a91-8c38-daefcb8a244e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.167706 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.167740 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.167750 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/587b6a35-3ed8-4a91-8c38-daefcb8a244e-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.167762 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zfkc\" (UniqueName: \"kubernetes.io/projected/587b6a35-3ed8-4a91-8c38-daefcb8a244e-kube-api-access-4zfkc\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.447970 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.645833 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61","Type":"ContainerStarted","Data":"e2c30f0d80c668de05cc791c9a1eeb36a27f56a73817d8c284014f5a021642f2"} Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.649151 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e","Type":"ContainerStarted","Data":"ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24"} Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.649264 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e","Type":"ContainerStarted","Data":"553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3"} Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.649282 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e","Type":"ContainerStarted","Data":"e1ded89aeaa541bf79155d68459e0fc55f94edd47352ad0f72204eb839a1e810"} Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.652084 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-t45x7" event={"ID":"587b6a35-3ed8-4a91-8c38-daefcb8a244e","Type":"ContainerDied","Data":"dc723a5b95d2b53c724b6990a9d8f5123ddeefb5ae2af0c9930dcabeb6b47712"} Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.652128 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc723a5b95d2b53c724b6990a9d8f5123ddeefb5ae2af0c9930dcabeb6b47712" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.652214 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-t45x7" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.654537 4846 generic.go:334] "Generic (PLEG): container finished" podID="8904d23d-2e73-4eb8-b6f6-8d559d0365e6" containerID="8d1edba65159d50673e4e593a3f7a608fd860b9d5b8b72df9c25acb668c97010" exitCode=0 Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.654645 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hgf4w" event={"ID":"8904d23d-2e73-4eb8-b6f6-8d559d0365e6","Type":"ContainerDied","Data":"8d1edba65159d50673e4e593a3f7a608fd860b9d5b8b72df9c25acb668c97010"} Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.707879 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.707839409 podStartE2EDuration="2.707839409s" podCreationTimestamp="2026-02-02 13:47:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:02.675116074 +0000 UTC m=+5853.903702957" watchObservedRunningTime="2026-02-02 13:47:02.707839409 +0000 UTC m=+5853.936426272" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.719795 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 13:47:02 crc kubenswrapper[4846]: E0202 13:47:02.720649 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="587b6a35-3ed8-4a91-8c38-daefcb8a244e" containerName="nova-cell1-conductor-db-sync" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.720675 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="587b6a35-3ed8-4a91-8c38-daefcb8a244e" containerName="nova-cell1-conductor-db-sync" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.721063 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="587b6a35-3ed8-4a91-8c38-daefcb8a244e" containerName="nova-cell1-conductor-db-sync" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.722197 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.740690 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.768927 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.785397 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c785t\" (UniqueName: \"kubernetes.io/projected/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-kube-api-access-c785t\") pod \"nova-cell1-conductor-0\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.785459 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.785703 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.887743 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.887835 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c785t\" (UniqueName: \"kubernetes.io/projected/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-kube-api-access-c785t\") pod \"nova-cell1-conductor-0\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.887861 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.892419 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.892497 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:02 crc kubenswrapper[4846]: I0202 13:47:02.904788 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c785t\" (UniqueName: \"kubernetes.io/projected/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-kube-api-access-c785t\") pod \"nova-cell1-conductor-0\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:03 crc kubenswrapper[4846]: I0202 13:47:03.065948 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:03 crc kubenswrapper[4846]: I0202 13:47:03.446349 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd46fafd-285b-43a0-a507-99825152e830" path="/var/lib/kubelet/pods/dd46fafd-285b-43a0-a507-99825152e830/volumes" Feb 02 13:47:03 crc kubenswrapper[4846]: I0202 13:47:03.522088 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 13:47:03 crc kubenswrapper[4846]: I0202 13:47:03.682851 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e611f1f1-65e6-4ce4-898b-a16e4bf0ee61","Type":"ContainerStarted","Data":"41786546b2db1cb71e6a941e2d46da712c50d31eb118b4180058d9406db81a03"} Feb 02 13:47:03 crc kubenswrapper[4846]: I0202 13:47:03.689953 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"83c7cd73-a72d-4f4f-9b36-b7677290ecfa","Type":"ContainerStarted","Data":"af7ddccd8d7a27c0629f75c22f8d8fb745372e2a4ede5cfb1b8b51a9c2764859"} Feb 02 13:47:03 crc kubenswrapper[4846]: I0202 13:47:03.717493 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.717469549 podStartE2EDuration="2.717469549s" podCreationTimestamp="2026-02-02 13:47:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:03.706721377 +0000 UTC m=+5854.935308240" watchObservedRunningTime="2026-02-02 13:47:03.717469549 +0000 UTC m=+5854.946056412" Feb 02 13:47:03 crc kubenswrapper[4846]: I0202 13:47:03.980204 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.123857 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-config-data\") pod \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.124481 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-combined-ca-bundle\") pod \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.124688 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-scripts\") pod \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.124908 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcmpt\" (UniqueName: \"kubernetes.io/projected/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-kube-api-access-zcmpt\") pod \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\" (UID: \"8904d23d-2e73-4eb8-b6f6-8d559d0365e6\") " Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.130225 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-scripts" (OuterVolumeSpecName: "scripts") pod "8904d23d-2e73-4eb8-b6f6-8d559d0365e6" (UID: "8904d23d-2e73-4eb8-b6f6-8d559d0365e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.143797 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-kube-api-access-zcmpt" (OuterVolumeSpecName: "kube-api-access-zcmpt") pod "8904d23d-2e73-4eb8-b6f6-8d559d0365e6" (UID: "8904d23d-2e73-4eb8-b6f6-8d559d0365e6"). InnerVolumeSpecName "kube-api-access-zcmpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.158019 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-config-data" (OuterVolumeSpecName: "config-data") pod "8904d23d-2e73-4eb8-b6f6-8d559d0365e6" (UID: "8904d23d-2e73-4eb8-b6f6-8d559d0365e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.158918 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8904d23d-2e73-4eb8-b6f6-8d559d0365e6" (UID: "8904d23d-2e73-4eb8-b6f6-8d559d0365e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.226707 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcmpt\" (UniqueName: \"kubernetes.io/projected/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-kube-api-access-zcmpt\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.226757 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.226769 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.226780 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8904d23d-2e73-4eb8-b6f6-8d559d0365e6-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.703324 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hgf4w" event={"ID":"8904d23d-2e73-4eb8-b6f6-8d559d0365e6","Type":"ContainerDied","Data":"4df91dc4ad44655a8bc6950564dc9cdbe6f66807edccf4e44edf49aed0dc1439"} Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.703361 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4df91dc4ad44655a8bc6950564dc9cdbe6f66807edccf4e44edf49aed0dc1439" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.703332 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hgf4w" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.705930 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"83c7cd73-a72d-4f4f-9b36-b7677290ecfa","Type":"ContainerStarted","Data":"e515de17b99aabcf65b27fd30dc2a3f94580b7ba2de58e83857ffaf644ca0ad8"} Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.705991 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.766206 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.766183547 podStartE2EDuration="2.766183547s" podCreationTimestamp="2026-02-02 13:47:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:04.738409723 +0000 UTC m=+5855.966996596" watchObservedRunningTime="2026-02-02 13:47:04.766183547 +0000 UTC m=+5855.994770440" Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.889538 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.889772 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a3b9618d-8751-4770-817d-9db761074535" containerName="nova-api-log" containerID="cri-o://ebb01c224316779e1edabea50d37cc50fd49dc66d7798e3b8c4e9680f8fb51fa" gracePeriod=30 Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.890112 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a3b9618d-8751-4770-817d-9db761074535" containerName="nova-api-api" containerID="cri-o://3fb804d856b1de1cc7cb165564d4afc3349b7de29a7f2728c659fc710010e2a8" gracePeriod=30 Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.925642 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.925897 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ce616c5c-5881-4cfc-934d-b1c789d76e46" containerName="nova-scheduler-scheduler" containerID="cri-o://57e7b6c791e1736f567e02efac024b4da6aebcc1959db3ab46ca55884c16cded" gracePeriod=30 Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.935381 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.935656 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" containerName="nova-metadata-log" containerID="cri-o://553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3" gracePeriod=30 Feb 02 13:47:04 crc kubenswrapper[4846]: I0202 13:47:04.935748 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" containerName="nova-metadata-metadata" containerID="cri-o://ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24" gracePeriod=30 Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.424334 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:47:05 crc kubenswrapper[4846]: E0202 13:47:05.424591 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.663993 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.724832 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" containerID="ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24" exitCode=0 Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.724865 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" containerID="553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3" exitCode=143 Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.724944 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e","Type":"ContainerDied","Data":"ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24"} Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.724981 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e","Type":"ContainerDied","Data":"553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3"} Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.724993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e","Type":"ContainerDied","Data":"e1ded89aeaa541bf79155d68459e0fc55f94edd47352ad0f72204eb839a1e810"} Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.725012 4846 scope.go:117] "RemoveContainer" containerID="ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.725080 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.729984 4846 generic.go:334] "Generic (PLEG): container finished" podID="a3b9618d-8751-4770-817d-9db761074535" containerID="3fb804d856b1de1cc7cb165564d4afc3349b7de29a7f2728c659fc710010e2a8" exitCode=0 Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.730013 4846 generic.go:334] "Generic (PLEG): container finished" podID="a3b9618d-8751-4770-817d-9db761074535" containerID="ebb01c224316779e1edabea50d37cc50fd49dc66d7798e3b8c4e9680f8fb51fa" exitCode=143 Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.730062 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3b9618d-8751-4770-817d-9db761074535","Type":"ContainerDied","Data":"3fb804d856b1de1cc7cb165564d4afc3349b7de29a7f2728c659fc710010e2a8"} Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.730098 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3b9618d-8751-4770-817d-9db761074535","Type":"ContainerDied","Data":"ebb01c224316779e1edabea50d37cc50fd49dc66d7798e3b8c4e9680f8fb51fa"} Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.759974 4846 scope.go:117] "RemoveContainer" containerID="553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.764990 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mn9b\" (UniqueName: \"kubernetes.io/projected/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-kube-api-access-4mn9b\") pod \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.765065 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-nova-metadata-tls-certs\") pod \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.765415 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-config-data\") pod \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.765480 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-logs\") pod \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.765660 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-combined-ca-bundle\") pod \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\" (UID: \"aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e\") " Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.766139 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-logs" (OuterVolumeSpecName: "logs") pod "aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" (UID: "aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.766545 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.770646 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-kube-api-access-4mn9b" (OuterVolumeSpecName: "kube-api-access-4mn9b") pod "aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" (UID: "aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e"). InnerVolumeSpecName "kube-api-access-4mn9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.794671 4846 scope.go:117] "RemoveContainer" containerID="ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24" Feb 02 13:47:05 crc kubenswrapper[4846]: E0202 13:47:05.795117 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24\": container with ID starting with ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24 not found: ID does not exist" containerID="ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.795162 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24"} err="failed to get container status \"ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24\": rpc error: code = NotFound desc = could not find container \"ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24\": container with ID starting with ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24 not found: ID does not exist" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.798413 4846 scope.go:117] "RemoveContainer" containerID="553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.799284 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" (UID: "aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:05 crc kubenswrapper[4846]: E0202 13:47:05.799037 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3\": container with ID starting with 553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3 not found: ID does not exist" containerID="553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.799911 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3"} err="failed to get container status \"553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3\": rpc error: code = NotFound desc = could not find container \"553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3\": container with ID starting with 553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3 not found: ID does not exist" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.799949 4846 scope.go:117] "RemoveContainer" containerID="ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.800384 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24"} err="failed to get container status \"ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24\": rpc error: code = NotFound desc = could not find container \"ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24\": container with ID starting with ea3bc851b619700a285b193617778667c30f5e80995acaaec57c9ff1e1943e24 not found: ID does not exist" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.800414 4846 scope.go:117] "RemoveContainer" containerID="553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.802528 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3"} err="failed to get container status \"553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3\": rpc error: code = NotFound desc = could not find container \"553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3\": container with ID starting with 553cfeea8945a601b460f7adf06df8a78cb6574f032cbf9304934a1941281ac3 not found: ID does not exist" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.804004 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-config-data" (OuterVolumeSpecName: "config-data") pod "aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" (UID: "aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.842734 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" (UID: "aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.868417 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mn9b\" (UniqueName: \"kubernetes.io/projected/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-kube-api-access-4mn9b\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.868449 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.868462 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.868473 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:05 crc kubenswrapper[4846]: I0202 13:47:05.905518 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.026133 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.070697 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-combined-ca-bundle\") pod \"a3b9618d-8751-4770-817d-9db761074535\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.070751 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3b9618d-8751-4770-817d-9db761074535-logs\") pod \"a3b9618d-8751-4770-817d-9db761074535\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.070866 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmssg\" (UniqueName: \"kubernetes.io/projected/a3b9618d-8751-4770-817d-9db761074535-kube-api-access-jmssg\") pod \"a3b9618d-8751-4770-817d-9db761074535\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.070939 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-config-data\") pod \"a3b9618d-8751-4770-817d-9db761074535\" (UID: \"a3b9618d-8751-4770-817d-9db761074535\") " Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.072012 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3b9618d-8751-4770-817d-9db761074535-logs" (OuterVolumeSpecName: "logs") pod "a3b9618d-8751-4770-817d-9db761074535" (UID: "a3b9618d-8751-4770-817d-9db761074535"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.079079 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3b9618d-8751-4770-817d-9db761074535-kube-api-access-jmssg" (OuterVolumeSpecName: "kube-api-access-jmssg") pod "a3b9618d-8751-4770-817d-9db761074535" (UID: "a3b9618d-8751-4770-817d-9db761074535"). InnerVolumeSpecName "kube-api-access-jmssg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.102786 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b75886b77-7px8n"] Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.103052 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b75886b77-7px8n" podUID="67d871ee-e010-401c-95e8-9a68e7350c59" containerName="dnsmasq-dns" containerID="cri-o://cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd" gracePeriod=10 Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.141825 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3b9618d-8751-4770-817d-9db761074535" (UID: "a3b9618d-8751-4770-817d-9db761074535"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.167875 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-config-data" (OuterVolumeSpecName: "config-data") pod "a3b9618d-8751-4770-817d-9db761074535" (UID: "a3b9618d-8751-4770-817d-9db761074535"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.176982 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.177044 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3b9618d-8751-4770-817d-9db761074535-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.177057 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmssg\" (UniqueName: \"kubernetes.io/projected/a3b9618d-8751-4770-817d-9db761074535-kube-api-access-jmssg\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.177071 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3b9618d-8751-4770-817d-9db761074535-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.377938 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.412068 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.425934 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:06 crc kubenswrapper[4846]: E0202 13:47:06.426409 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8904d23d-2e73-4eb8-b6f6-8d559d0365e6" containerName="nova-manage" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.426430 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8904d23d-2e73-4eb8-b6f6-8d559d0365e6" containerName="nova-manage" Feb 02 13:47:06 crc kubenswrapper[4846]: E0202 13:47:06.426444 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3b9618d-8751-4770-817d-9db761074535" containerName="nova-api-api" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.426453 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3b9618d-8751-4770-817d-9db761074535" containerName="nova-api-api" Feb 02 13:47:06 crc kubenswrapper[4846]: E0202 13:47:06.426470 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" containerName="nova-metadata-log" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.426479 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" containerName="nova-metadata-log" Feb 02 13:47:06 crc kubenswrapper[4846]: E0202 13:47:06.426496 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" containerName="nova-metadata-metadata" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.426503 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" containerName="nova-metadata-metadata" Feb 02 13:47:06 crc kubenswrapper[4846]: E0202 13:47:06.426541 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3b9618d-8751-4770-817d-9db761074535" containerName="nova-api-log" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.426547 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3b9618d-8751-4770-817d-9db761074535" containerName="nova-api-log" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.426756 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3b9618d-8751-4770-817d-9db761074535" containerName="nova-api-log" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.426775 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3b9618d-8751-4770-817d-9db761074535" containerName="nova-api-api" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.426790 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8904d23d-2e73-4eb8-b6f6-8d559d0365e6" containerName="nova-manage" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.426798 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" containerName="nova-metadata-metadata" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.426813 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" containerName="nova-metadata-log" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.428029 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.431504 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.431708 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.446966 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.585348 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-config-data\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.585394 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5hd4\" (UniqueName: \"kubernetes.io/projected/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-kube-api-access-m5hd4\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.585420 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.585589 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-logs\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.585654 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.614413 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.687575 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-config-data\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.687637 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5hd4\" (UniqueName: \"kubernetes.io/projected/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-kube-api-access-m5hd4\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.687674 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.687742 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-logs\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.687792 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.688190 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-logs\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.691607 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.691923 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.692092 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-config-data\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.703544 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5hd4\" (UniqueName: \"kubernetes.io/projected/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-kube-api-access-m5hd4\") pod \"nova-metadata-0\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.746488 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.746489 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a3b9618d-8751-4770-817d-9db761074535","Type":"ContainerDied","Data":"414f330ff5797f6d145669d1168e60a056f8dfed937421ac39267ff06735e2b4"} Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.746585 4846 scope.go:117] "RemoveContainer" containerID="3fb804d856b1de1cc7cb165564d4afc3349b7de29a7f2728c659fc710010e2a8" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.749424 4846 generic.go:334] "Generic (PLEG): container finished" podID="67d871ee-e010-401c-95e8-9a68e7350c59" containerID="cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd" exitCode=0 Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.749500 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b75886b77-7px8n" event={"ID":"67d871ee-e010-401c-95e8-9a68e7350c59","Type":"ContainerDied","Data":"cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd"} Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.749525 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b75886b77-7px8n" event={"ID":"67d871ee-e010-401c-95e8-9a68e7350c59","Type":"ContainerDied","Data":"b0b1446a1b02ab44285a14b451534bbd3afe7f53e59f47c2eff3691996f069b1"} Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.749596 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b75886b77-7px8n" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.770980 4846 scope.go:117] "RemoveContainer" containerID="ebb01c224316779e1edabea50d37cc50fd49dc66d7798e3b8c4e9680f8fb51fa" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.772730 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.785887 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.789168 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-nb\") pod \"67d871ee-e010-401c-95e8-9a68e7350c59\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.789423 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbpsb\" (UniqueName: \"kubernetes.io/projected/67d871ee-e010-401c-95e8-9a68e7350c59-kube-api-access-wbpsb\") pod \"67d871ee-e010-401c-95e8-9a68e7350c59\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.789605 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-sb\") pod \"67d871ee-e010-401c-95e8-9a68e7350c59\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.789798 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-dns-svc\") pod \"67d871ee-e010-401c-95e8-9a68e7350c59\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.789907 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-config\") pod \"67d871ee-e010-401c-95e8-9a68e7350c59\" (UID: \"67d871ee-e010-401c-95e8-9a68e7350c59\") " Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.798369 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d871ee-e010-401c-95e8-9a68e7350c59-kube-api-access-wbpsb" (OuterVolumeSpecName: "kube-api-access-wbpsb") pod "67d871ee-e010-401c-95e8-9a68e7350c59" (UID: "67d871ee-e010-401c-95e8-9a68e7350c59"). InnerVolumeSpecName "kube-api-access-wbpsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.815241 4846 scope.go:117] "RemoveContainer" containerID="cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.824941 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.850401 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:06 crc kubenswrapper[4846]: E0202 13:47:06.851395 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d871ee-e010-401c-95e8-9a68e7350c59" containerName="dnsmasq-dns" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.851420 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d871ee-e010-401c-95e8-9a68e7350c59" containerName="dnsmasq-dns" Feb 02 13:47:06 crc kubenswrapper[4846]: E0202 13:47:06.851458 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d871ee-e010-401c-95e8-9a68e7350c59" containerName="init" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.851465 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d871ee-e010-401c-95e8-9a68e7350c59" containerName="init" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.851938 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d871ee-e010-401c-95e8-9a68e7350c59" containerName="dnsmasq-dns" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.851992 4846 scope.go:117] "RemoveContainer" containerID="bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.854101 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.857814 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.857909 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "67d871ee-e010-401c-95e8-9a68e7350c59" (UID: "67d871ee-e010-401c-95e8-9a68e7350c59"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.863311 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.864157 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-config" (OuterVolumeSpecName: "config") pod "67d871ee-e010-401c-95e8-9a68e7350c59" (UID: "67d871ee-e010-401c-95e8-9a68e7350c59"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.883605 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "67d871ee-e010-401c-95e8-9a68e7350c59" (UID: "67d871ee-e010-401c-95e8-9a68e7350c59"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.884526 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "67d871ee-e010-401c-95e8-9a68e7350c59" (UID: "67d871ee-e010-401c-95e8-9a68e7350c59"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.887188 4846 scope.go:117] "RemoveContainer" containerID="cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd" Feb 02 13:47:06 crc kubenswrapper[4846]: E0202 13:47:06.887701 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd\": container with ID starting with cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd not found: ID does not exist" containerID="cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.887747 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd"} err="failed to get container status \"cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd\": rpc error: code = NotFound desc = could not find container \"cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd\": container with ID starting with cb80dcaf6665fd6f7ac59436f7dbaf984696234ecedd8882479013cc3ba559fd not found: ID does not exist" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.887774 4846 scope.go:117] "RemoveContainer" containerID="bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f" Feb 02 13:47:06 crc kubenswrapper[4846]: E0202 13:47:06.888088 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f\": container with ID starting with bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f not found: ID does not exist" containerID="bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.888110 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f"} err="failed to get container status \"bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f\": rpc error: code = NotFound desc = could not find container \"bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f\": container with ID starting with bd188400a6d565839ab0b528705af929416c96d4629126000252bfb21917ce9f not found: ID does not exist" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.892372 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.892406 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.892419 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.892431 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67d871ee-e010-401c-95e8-9a68e7350c59-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.892443 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbpsb\" (UniqueName: \"kubernetes.io/projected/67d871ee-e010-401c-95e8-9a68e7350c59-kube-api-access-wbpsb\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.994349 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p892r\" (UniqueName: \"kubernetes.io/projected/42edffd8-83cd-4697-af2f-d8cbca35d1a5-kube-api-access-p892r\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.994435 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-config-data\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.994594 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:06 crc kubenswrapper[4846]: I0202 13:47:06.994799 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42edffd8-83cd-4697-af2f-d8cbca35d1a5-logs\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.020374 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.096357 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.096448 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42edffd8-83cd-4697-af2f-d8cbca35d1a5-logs\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.096510 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p892r\" (UniqueName: \"kubernetes.io/projected/42edffd8-83cd-4697-af2f-d8cbca35d1a5-kube-api-access-p892r\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.096551 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-config-data\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.096873 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42edffd8-83cd-4697-af2f-d8cbca35d1a5-logs\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.097802 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b75886b77-7px8n"] Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.105280 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-config-data\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.105923 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.108216 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b75886b77-7px8n"] Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.123252 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p892r\" (UniqueName: \"kubernetes.io/projected/42edffd8-83cd-4697-af2f-d8cbca35d1a5-kube-api-access-p892r\") pod \"nova-api-0\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " pod="openstack/nova-api-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.181649 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.353229 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:07 crc kubenswrapper[4846]: W0202 13:47:07.361988 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e0f03ea_71c3_48cd_b80b_e2fef2e308b9.slice/crio-63bce56c7b1473d5646a2283abaeccad25d2287c50ae5668c64e8fba6a5aab21 WatchSource:0}: Error finding container 63bce56c7b1473d5646a2283abaeccad25d2287c50ae5668c64e8fba6a5aab21: Status 404 returned error can't find the container with id 63bce56c7b1473d5646a2283abaeccad25d2287c50ae5668c64e8fba6a5aab21 Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.471747 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d871ee-e010-401c-95e8-9a68e7350c59" path="/var/lib/kubelet/pods/67d871ee-e010-401c-95e8-9a68e7350c59/volumes" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.472332 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3b9618d-8751-4770-817d-9db761074535" path="/var/lib/kubelet/pods/a3b9618d-8751-4770-817d-9db761074535/volumes" Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.472913 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e" path="/var/lib/kubelet/pods/aa7a808a-a347-4cd1-8f12-c8a1d9e7e87e/volumes" Feb 02 13:47:07 crc kubenswrapper[4846]: W0202 13:47:07.678925 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42edffd8_83cd_4697_af2f_d8cbca35d1a5.slice/crio-87fd30337eb45e4cdc707fd0327f17d44f376a63d0cb5ade63f278a40e6742fc WatchSource:0}: Error finding container 87fd30337eb45e4cdc707fd0327f17d44f376a63d0cb5ade63f278a40e6742fc: Status 404 returned error can't find the container with id 87fd30337eb45e4cdc707fd0327f17d44f376a63d0cb5ade63f278a40e6742fc Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.679406 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.777568 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42edffd8-83cd-4697-af2f-d8cbca35d1a5","Type":"ContainerStarted","Data":"87fd30337eb45e4cdc707fd0327f17d44f376a63d0cb5ade63f278a40e6742fc"} Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.784579 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9","Type":"ContainerStarted","Data":"15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586"} Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.784641 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9","Type":"ContainerStarted","Data":"93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe"} Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.784655 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9","Type":"ContainerStarted","Data":"63bce56c7b1473d5646a2283abaeccad25d2287c50ae5668c64e8fba6a5aab21"} Feb 02 13:47:07 crc kubenswrapper[4846]: I0202 13:47:07.813944 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.813921583 podStartE2EDuration="1.813921583s" podCreationTimestamp="2026-02-02 13:47:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:07.80430377 +0000 UTC m=+5859.032890663" watchObservedRunningTime="2026-02-02 13:47:07.813921583 +0000 UTC m=+5859.042508446" Feb 02 13:47:08 crc kubenswrapper[4846]: I0202 13:47:08.093827 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 02 13:47:08 crc kubenswrapper[4846]: I0202 13:47:08.794360 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42edffd8-83cd-4697-af2f-d8cbca35d1a5","Type":"ContainerStarted","Data":"648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95"} Feb 02 13:47:08 crc kubenswrapper[4846]: I0202 13:47:08.794405 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42edffd8-83cd-4697-af2f-d8cbca35d1a5","Type":"ContainerStarted","Data":"91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7"} Feb 02 13:47:08 crc kubenswrapper[4846]: I0202 13:47:08.815047 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.815028996 podStartE2EDuration="2.815028996s" podCreationTimestamp="2026-02-02 13:47:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:08.812175446 +0000 UTC m=+5860.040762309" watchObservedRunningTime="2026-02-02 13:47:08.815028996 +0000 UTC m=+5860.043615859" Feb 02 13:47:11 crc kubenswrapper[4846]: I0202 13:47:11.773797 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 13:47:11 crc kubenswrapper[4846]: I0202 13:47:11.775790 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 13:47:12 crc kubenswrapper[4846]: I0202 13:47:12.020328 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:12 crc kubenswrapper[4846]: I0202 13:47:12.036778 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:12 crc kubenswrapper[4846]: I0202 13:47:12.852759 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.047669 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-6b2jw"] Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.048819 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.051109 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.051640 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.065841 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6b2jw"] Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.211338 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-config-data\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.211501 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.211529 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-scripts\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.211596 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rnsv\" (UniqueName: \"kubernetes.io/projected/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-kube-api-access-6rnsv\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.313756 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-config-data\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.313859 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.313890 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-scripts\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.313945 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rnsv\" (UniqueName: \"kubernetes.io/projected/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-kube-api-access-6rnsv\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.321336 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-scripts\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.322014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.322277 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-config-data\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.331411 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rnsv\" (UniqueName: \"kubernetes.io/projected/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-kube-api-access-6rnsv\") pod \"nova-cell1-cell-mapping-6b2jw\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.381850 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.814973 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6b2jw"] Feb 02 13:47:13 crc kubenswrapper[4846]: W0202 13:47:13.818026 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfddbbaf0_81be_43a4_8441_dcdf4d9b83c0.slice/crio-22c9ef0ca4ca270ff4392031a0fa6a8da9045e68a0c1f236560632e565180b05 WatchSource:0}: Error finding container 22c9ef0ca4ca270ff4392031a0fa6a8da9045e68a0c1f236560632e565180b05: Status 404 returned error can't find the container with id 22c9ef0ca4ca270ff4392031a0fa6a8da9045e68a0c1f236560632e565180b05 Feb 02 13:47:13 crc kubenswrapper[4846]: I0202 13:47:13.846234 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6b2jw" event={"ID":"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0","Type":"ContainerStarted","Data":"22c9ef0ca4ca270ff4392031a0fa6a8da9045e68a0c1f236560632e565180b05"} Feb 02 13:47:14 crc kubenswrapper[4846]: I0202 13:47:14.858133 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6b2jw" event={"ID":"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0","Type":"ContainerStarted","Data":"c771d54ac604b5cf688d25d00e2566e75cdb141e7feca723b8a949b372c3f03a"} Feb 02 13:47:14 crc kubenswrapper[4846]: I0202 13:47:14.880799 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-6b2jw" podStartSLOduration=1.880776687 podStartE2EDuration="1.880776687s" podCreationTimestamp="2026-02-02 13:47:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:14.873607003 +0000 UTC m=+5866.102193866" watchObservedRunningTime="2026-02-02 13:47:14.880776687 +0000 UTC m=+5866.109363550" Feb 02 13:47:16 crc kubenswrapper[4846]: I0202 13:47:16.773907 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 13:47:16 crc kubenswrapper[4846]: I0202 13:47:16.774223 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 13:47:17 crc kubenswrapper[4846]: I0202 13:47:17.182823 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 13:47:17 crc kubenswrapper[4846]: I0202 13:47:17.183362 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 13:47:17 crc kubenswrapper[4846]: I0202 13:47:17.786881 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.89:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 13:47:17 crc kubenswrapper[4846]: I0202 13:47:17.786893 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.89:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 13:47:18 crc kubenswrapper[4846]: I0202 13:47:18.267062 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 13:47:18 crc kubenswrapper[4846]: I0202 13:47:18.267067 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 13:47:18 crc kubenswrapper[4846]: I0202 13:47:18.900372 4846 generic.go:334] "Generic (PLEG): container finished" podID="fddbbaf0-81be-43a4-8441-dcdf4d9b83c0" containerID="c771d54ac604b5cf688d25d00e2566e75cdb141e7feca723b8a949b372c3f03a" exitCode=0 Feb 02 13:47:18 crc kubenswrapper[4846]: I0202 13:47:18.900409 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6b2jw" event={"ID":"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0","Type":"ContainerDied","Data":"c771d54ac604b5cf688d25d00e2566e75cdb141e7feca723b8a949b372c3f03a"} Feb 02 13:47:19 crc kubenswrapper[4846]: I0202 13:47:19.436082 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:47:19 crc kubenswrapper[4846]: E0202 13:47:19.436391 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.241700 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.364486 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-config-data\") pod \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.364749 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-scripts\") pod \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.364936 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rnsv\" (UniqueName: \"kubernetes.io/projected/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-kube-api-access-6rnsv\") pod \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.365051 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-combined-ca-bundle\") pod \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\" (UID: \"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0\") " Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.377766 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-scripts" (OuterVolumeSpecName: "scripts") pod "fddbbaf0-81be-43a4-8441-dcdf4d9b83c0" (UID: "fddbbaf0-81be-43a4-8441-dcdf4d9b83c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.382133 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-kube-api-access-6rnsv" (OuterVolumeSpecName: "kube-api-access-6rnsv") pod "fddbbaf0-81be-43a4-8441-dcdf4d9b83c0" (UID: "fddbbaf0-81be-43a4-8441-dcdf4d9b83c0"). InnerVolumeSpecName "kube-api-access-6rnsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.401374 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fddbbaf0-81be-43a4-8441-dcdf4d9b83c0" (UID: "fddbbaf0-81be-43a4-8441-dcdf4d9b83c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.418465 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-config-data" (OuterVolumeSpecName: "config-data") pod "fddbbaf0-81be-43a4-8441-dcdf4d9b83c0" (UID: "fddbbaf0-81be-43a4-8441-dcdf4d9b83c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.469401 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.469744 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.469758 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.469770 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rnsv\" (UniqueName: \"kubernetes.io/projected/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0-kube-api-access-6rnsv\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.918266 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6b2jw" event={"ID":"fddbbaf0-81be-43a4-8441-dcdf4d9b83c0","Type":"ContainerDied","Data":"22c9ef0ca4ca270ff4392031a0fa6a8da9045e68a0c1f236560632e565180b05"} Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.918308 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22c9ef0ca4ca270ff4392031a0fa6a8da9045e68a0c1f236560632e565180b05" Feb 02 13:47:20 crc kubenswrapper[4846]: I0202 13:47:20.918390 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6b2jw" Feb 02 13:47:21 crc kubenswrapper[4846]: I0202 13:47:21.106299 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:21 crc kubenswrapper[4846]: I0202 13:47:21.106546 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerName="nova-api-log" containerID="cri-o://91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7" gracePeriod=30 Feb 02 13:47:21 crc kubenswrapper[4846]: I0202 13:47:21.106618 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerName="nova-api-api" containerID="cri-o://648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95" gracePeriod=30 Feb 02 13:47:21 crc kubenswrapper[4846]: I0202 13:47:21.134808 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:21 crc kubenswrapper[4846]: I0202 13:47:21.135065 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerName="nova-metadata-log" containerID="cri-o://93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe" gracePeriod=30 Feb 02 13:47:21 crc kubenswrapper[4846]: I0202 13:47:21.135190 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerName="nova-metadata-metadata" containerID="cri-o://15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586" gracePeriod=30 Feb 02 13:47:21 crc kubenswrapper[4846]: I0202 13:47:21.927245 4846 generic.go:334] "Generic (PLEG): container finished" podID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerID="93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe" exitCode=143 Feb 02 13:47:21 crc kubenswrapper[4846]: I0202 13:47:21.927316 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9","Type":"ContainerDied","Data":"93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe"} Feb 02 13:47:21 crc kubenswrapper[4846]: I0202 13:47:21.930096 4846 generic.go:334] "Generic (PLEG): container finished" podID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerID="91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7" exitCode=143 Feb 02 13:47:21 crc kubenswrapper[4846]: I0202 13:47:21.930166 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42edffd8-83cd-4697-af2f-d8cbca35d1a5","Type":"ContainerDied","Data":"91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7"} Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.285556 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6mxjq"] Feb 02 13:47:23 crc kubenswrapper[4846]: E0202 13:47:23.286559 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddbbaf0-81be-43a4-8441-dcdf4d9b83c0" containerName="nova-manage" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.286576 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddbbaf0-81be-43a4-8441-dcdf4d9b83c0" containerName="nova-manage" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.286829 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fddbbaf0-81be-43a4-8441-dcdf4d9b83c0" containerName="nova-manage" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.292223 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.303010 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6mxjq"] Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.424360 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-utilities\") pod \"certified-operators-6mxjq\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.424673 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t78ff\" (UniqueName: \"kubernetes.io/projected/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-kube-api-access-t78ff\") pod \"certified-operators-6mxjq\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.424771 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-catalog-content\") pod \"certified-operators-6mxjq\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.526922 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-utilities\") pod \"certified-operators-6mxjq\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.527259 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t78ff\" (UniqueName: \"kubernetes.io/projected/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-kube-api-access-t78ff\") pod \"certified-operators-6mxjq\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.527371 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-catalog-content\") pod \"certified-operators-6mxjq\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.527576 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-utilities\") pod \"certified-operators-6mxjq\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.527731 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-catalog-content\") pod \"certified-operators-6mxjq\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.551513 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t78ff\" (UniqueName: \"kubernetes.io/projected/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-kube-api-access-t78ff\") pod \"certified-operators-6mxjq\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:23 crc kubenswrapper[4846]: I0202 13:47:23.613188 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.085099 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6mxjq"] Feb 02 13:47:24 crc kubenswrapper[4846]: W0202 13:47:24.091523 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60e91bc7_6a91_4abd_9ea3_e5e462a682cd.slice/crio-1680972c833ccddd850a04f09172b36af291b967963b7170643d08adf3528b6c WatchSource:0}: Error finding container 1680972c833ccddd850a04f09172b36af291b967963b7170643d08adf3528b6c: Status 404 returned error can't find the container with id 1680972c833ccddd850a04f09172b36af291b967963b7170643d08adf3528b6c Feb 02 13:47:24 crc kubenswrapper[4846]: E0202 13:47:24.363244 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42edffd8_83cd_4697_af2f_d8cbca35d1a5.slice/crio-conmon-648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42edffd8_83cd_4697_af2f_d8cbca35d1a5.slice/crio-648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e0f03ea_71c3_48cd_b80b_e2fef2e308b9.slice/crio-conmon-15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586.scope\": RecentStats: unable to find data in memory cache]" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.748444 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.756875 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.857903 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-config-data\") pod \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.857971 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-nova-metadata-tls-certs\") pod \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.858017 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5hd4\" (UniqueName: \"kubernetes.io/projected/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-kube-api-access-m5hd4\") pod \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.858049 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-config-data\") pod \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.858066 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p892r\" (UniqueName: \"kubernetes.io/projected/42edffd8-83cd-4697-af2f-d8cbca35d1a5-kube-api-access-p892r\") pod \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.858103 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-combined-ca-bundle\") pod \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.858119 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-logs\") pod \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\" (UID: \"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9\") " Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.858169 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42edffd8-83cd-4697-af2f-d8cbca35d1a5-logs\") pod \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.858204 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-combined-ca-bundle\") pod \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\" (UID: \"42edffd8-83cd-4697-af2f-d8cbca35d1a5\") " Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.860287 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-logs" (OuterVolumeSpecName: "logs") pod "1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" (UID: "1e0f03ea-71c3-48cd-b80b-e2fef2e308b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.861465 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42edffd8-83cd-4697-af2f-d8cbca35d1a5-logs" (OuterVolumeSpecName: "logs") pod "42edffd8-83cd-4697-af2f-d8cbca35d1a5" (UID: "42edffd8-83cd-4697-af2f-d8cbca35d1a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.864042 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42edffd8-83cd-4697-af2f-d8cbca35d1a5-kube-api-access-p892r" (OuterVolumeSpecName: "kube-api-access-p892r") pod "42edffd8-83cd-4697-af2f-d8cbca35d1a5" (UID: "42edffd8-83cd-4697-af2f-d8cbca35d1a5"). InnerVolumeSpecName "kube-api-access-p892r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.864312 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-kube-api-access-m5hd4" (OuterVolumeSpecName: "kube-api-access-m5hd4") pod "1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" (UID: "1e0f03ea-71c3-48cd-b80b-e2fef2e308b9"). InnerVolumeSpecName "kube-api-access-m5hd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.886399 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-config-data" (OuterVolumeSpecName: "config-data") pod "42edffd8-83cd-4697-af2f-d8cbca35d1a5" (UID: "42edffd8-83cd-4697-af2f-d8cbca35d1a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.887014 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42edffd8-83cd-4697-af2f-d8cbca35d1a5" (UID: "42edffd8-83cd-4697-af2f-d8cbca35d1a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.887984 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" (UID: "1e0f03ea-71c3-48cd-b80b-e2fef2e308b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.888803 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-config-data" (OuterVolumeSpecName: "config-data") pod "1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" (UID: "1e0f03ea-71c3-48cd-b80b-e2fef2e308b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.922489 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" (UID: "1e0f03ea-71c3-48cd-b80b-e2fef2e308b9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.959726 4846 generic.go:334] "Generic (PLEG): container finished" podID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerID="15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586" exitCode=0 Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.959793 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.959804 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9","Type":"ContainerDied","Data":"15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586"} Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.959852 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e0f03ea-71c3-48cd-b80b-e2fef2e308b9","Type":"ContainerDied","Data":"63bce56c7b1473d5646a2283abaeccad25d2287c50ae5668c64e8fba6a5aab21"} Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.959873 4846 scope.go:117] "RemoveContainer" containerID="15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.960138 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.961234 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5hd4\" (UniqueName: \"kubernetes.io/projected/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-kube-api-access-m5hd4\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.961261 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p892r\" (UniqueName: \"kubernetes.io/projected/42edffd8-83cd-4697-af2f-d8cbca35d1a5-kube-api-access-p892r\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.961272 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.961283 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.961295 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.961306 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42edffd8-83cd-4697-af2f-d8cbca35d1a5-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.961314 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42edffd8-83cd-4697-af2f-d8cbca35d1a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.961323 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.962480 4846 generic.go:334] "Generic (PLEG): container finished" podID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerID="648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95" exitCode=0 Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.962543 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42edffd8-83cd-4697-af2f-d8cbca35d1a5","Type":"ContainerDied","Data":"648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95"} Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.962569 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42edffd8-83cd-4697-af2f-d8cbca35d1a5","Type":"ContainerDied","Data":"87fd30337eb45e4cdc707fd0327f17d44f376a63d0cb5ade63f278a40e6742fc"} Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.962546 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.968515 4846 generic.go:334] "Generic (PLEG): container finished" podID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerID="46c1842de9d2086d9301be32aff44bc2987c51c4aed2eb870061fdd7b86bf861" exitCode=0 Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.968565 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mxjq" event={"ID":"60e91bc7-6a91-4abd-9ea3-e5e462a682cd","Type":"ContainerDied","Data":"46c1842de9d2086d9301be32aff44bc2987c51c4aed2eb870061fdd7b86bf861"} Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.968599 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mxjq" event={"ID":"60e91bc7-6a91-4abd-9ea3-e5e462a682cd","Type":"ContainerStarted","Data":"1680972c833ccddd850a04f09172b36af291b967963b7170643d08adf3528b6c"} Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.983419 4846 scope.go:117] "RemoveContainer" containerID="93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe" Feb 02 13:47:24 crc kubenswrapper[4846]: I0202 13:47:24.983890 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.013193 4846 scope.go:117] "RemoveContainer" containerID="15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586" Feb 02 13:47:25 crc kubenswrapper[4846]: E0202 13:47:25.013687 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586\": container with ID starting with 15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586 not found: ID does not exist" containerID="15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.013736 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586"} err="failed to get container status \"15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586\": rpc error: code = NotFound desc = could not find container \"15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586\": container with ID starting with 15848a280cde2027843bdda727d059adbd6fa099c3cddf4e4116323606d02586 not found: ID does not exist" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.013765 4846 scope.go:117] "RemoveContainer" containerID="93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe" Feb 02 13:47:25 crc kubenswrapper[4846]: E0202 13:47:25.014045 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe\": container with ID starting with 93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe not found: ID does not exist" containerID="93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.014069 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe"} err="failed to get container status \"93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe\": rpc error: code = NotFound desc = could not find container \"93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe\": container with ID starting with 93dad29d27e3fe7e6e616654e4033d074c30d783f18c826325592e8aab9f1abe not found: ID does not exist" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.014080 4846 scope.go:117] "RemoveContainer" containerID="648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.014571 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.027389 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.046776 4846 scope.go:117] "RemoveContainer" containerID="91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.049878 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.071907 4846 scope.go:117] "RemoveContainer" containerID="648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95" Feb 02 13:47:25 crc kubenswrapper[4846]: E0202 13:47:25.073856 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95\": container with ID starting with 648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95 not found: ID does not exist" containerID="648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.073895 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95"} err="failed to get container status \"648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95\": rpc error: code = NotFound desc = could not find container \"648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95\": container with ID starting with 648b1360c4e7e698e9dbb71e59f13a76405c6b67fd79adecad6219f8bcbb6a95 not found: ID does not exist" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.073927 4846 scope.go:117] "RemoveContainer" containerID="91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7" Feb 02 13:47:25 crc kubenswrapper[4846]: E0202 13:47:25.074167 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7\": container with ID starting with 91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7 not found: ID does not exist" containerID="91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.074202 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7"} err="failed to get container status \"91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7\": rpc error: code = NotFound desc = could not find container \"91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7\": container with ID starting with 91e6b903e23edd232b7db319f3fc4bb2fab768403d442a0bd471c2ea92b723f7 not found: ID does not exist" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.077909 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.088400 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:25 crc kubenswrapper[4846]: E0202 13:47:25.088858 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerName="nova-api-log" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.088871 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerName="nova-api-log" Feb 02 13:47:25 crc kubenswrapper[4846]: E0202 13:47:25.088894 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerName="nova-metadata-metadata" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.088901 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerName="nova-metadata-metadata" Feb 02 13:47:25 crc kubenswrapper[4846]: E0202 13:47:25.088925 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerName="nova-metadata-log" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.088931 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerName="nova-metadata-log" Feb 02 13:47:25 crc kubenswrapper[4846]: E0202 13:47:25.088944 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerName="nova-api-api" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.088950 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerName="nova-api-api" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.089126 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerName="nova-api-log" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.089145 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerName="nova-metadata-metadata" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.089156 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" containerName="nova-api-api" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.089170 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" containerName="nova-metadata-log" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.090162 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.092574 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.092798 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.097263 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.098799 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.100479 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.107998 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.118596 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.266459 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/496f99aa-8e97-489f-8957-7deaba86dc1a-logs\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.266538 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpm5f\" (UniqueName: \"kubernetes.io/projected/23d3a089-c3dc-411d-a099-3ff80dd40d2d-kube-api-access-fpm5f\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.266564 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh5zw\" (UniqueName: \"kubernetes.io/projected/496f99aa-8e97-489f-8957-7deaba86dc1a-kube-api-access-gh5zw\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.266608 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-config-data\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.266794 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d3a089-c3dc-411d-a099-3ff80dd40d2d-logs\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.266872 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.266912 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.266947 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.267379 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-config-data\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.369571 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.369676 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.369756 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-config-data\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.369778 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/496f99aa-8e97-489f-8957-7deaba86dc1a-logs\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.369837 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpm5f\" (UniqueName: \"kubernetes.io/projected/23d3a089-c3dc-411d-a099-3ff80dd40d2d-kube-api-access-fpm5f\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.369872 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh5zw\" (UniqueName: \"kubernetes.io/projected/496f99aa-8e97-489f-8957-7deaba86dc1a-kube-api-access-gh5zw\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.369927 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-config-data\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.369978 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d3a089-c3dc-411d-a099-3ff80dd40d2d-logs\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.369999 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.371080 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d3a089-c3dc-411d-a099-3ff80dd40d2d-logs\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.371274 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/496f99aa-8e97-489f-8957-7deaba86dc1a-logs\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.374137 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.374683 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-config-data\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.374920 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-config-data\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.375069 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.375850 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.386472 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh5zw\" (UniqueName: \"kubernetes.io/projected/496f99aa-8e97-489f-8957-7deaba86dc1a-kube-api-access-gh5zw\") pod \"nova-api-0\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.387854 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpm5f\" (UniqueName: \"kubernetes.io/projected/23d3a089-c3dc-411d-a099-3ff80dd40d2d-kube-api-access-fpm5f\") pod \"nova-metadata-0\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.412855 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.420950 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.436034 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e0f03ea-71c3-48cd-b80b-e2fef2e308b9" path="/var/lib/kubelet/pods/1e0f03ea-71c3-48cd-b80b-e2fef2e308b9/volumes" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.437533 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42edffd8-83cd-4697-af2f-d8cbca35d1a5" path="/var/lib/kubelet/pods/42edffd8-83cd-4697-af2f-d8cbca35d1a5/volumes" Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.921761 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 13:47:25 crc kubenswrapper[4846]: W0202 13:47:25.924611 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23d3a089_c3dc_411d_a099_3ff80dd40d2d.slice/crio-389c37c72a1c48a690c38947ca2eae59fc7924f19c28c4edd35e18d3cb4d678f WatchSource:0}: Error finding container 389c37c72a1c48a690c38947ca2eae59fc7924f19c28c4edd35e18d3cb4d678f: Status 404 returned error can't find the container with id 389c37c72a1c48a690c38947ca2eae59fc7924f19c28c4edd35e18d3cb4d678f Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.987712 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mxjq" event={"ID":"60e91bc7-6a91-4abd-9ea3-e5e462a682cd","Type":"ContainerStarted","Data":"b613a7f558aff9f9c293db50d57b1e92483c5da503087919be60368dadf69889"} Feb 02 13:47:25 crc kubenswrapper[4846]: I0202 13:47:25.991352 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23d3a089-c3dc-411d-a099-3ff80dd40d2d","Type":"ContainerStarted","Data":"389c37c72a1c48a690c38947ca2eae59fc7924f19c28c4edd35e18d3cb4d678f"} Feb 02 13:47:26 crc kubenswrapper[4846]: W0202 13:47:26.013937 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod496f99aa_8e97_489f_8957_7deaba86dc1a.slice/crio-9d9d802f6fecf6ce509abbb42f3bf13819eb700b3cbe6c88253abf006cde10a7 WatchSource:0}: Error finding container 9d9d802f6fecf6ce509abbb42f3bf13819eb700b3cbe6c88253abf006cde10a7: Status 404 returned error can't find the container with id 9d9d802f6fecf6ce509abbb42f3bf13819eb700b3cbe6c88253abf006cde10a7 Feb 02 13:47:26 crc kubenswrapper[4846]: I0202 13:47:26.023723 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.006439 4846 generic.go:334] "Generic (PLEG): container finished" podID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerID="b613a7f558aff9f9c293db50d57b1e92483c5da503087919be60368dadf69889" exitCode=0 Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.006548 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mxjq" event={"ID":"60e91bc7-6a91-4abd-9ea3-e5e462a682cd","Type":"ContainerDied","Data":"b613a7f558aff9f9c293db50d57b1e92483c5da503087919be60368dadf69889"} Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.007173 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mxjq" event={"ID":"60e91bc7-6a91-4abd-9ea3-e5e462a682cd","Type":"ContainerStarted","Data":"fa15995595e1388630572ec7a422986f43a6367ab7ec6952e80248bb3b6532b9"} Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.009160 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23d3a089-c3dc-411d-a099-3ff80dd40d2d","Type":"ContainerStarted","Data":"34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb"} Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.009240 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23d3a089-c3dc-411d-a099-3ff80dd40d2d","Type":"ContainerStarted","Data":"211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490"} Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.011569 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"496f99aa-8e97-489f-8957-7deaba86dc1a","Type":"ContainerStarted","Data":"25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199"} Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.011595 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"496f99aa-8e97-489f-8957-7deaba86dc1a","Type":"ContainerStarted","Data":"dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b"} Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.011606 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"496f99aa-8e97-489f-8957-7deaba86dc1a","Type":"ContainerStarted","Data":"9d9d802f6fecf6ce509abbb42f3bf13819eb700b3cbe6c88253abf006cde10a7"} Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.039586 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6mxjq" podStartSLOduration=2.543309852 podStartE2EDuration="4.039562059s" podCreationTimestamp="2026-02-02 13:47:23 +0000 UTC" firstStartedPulling="2026-02-02 13:47:24.98364973 +0000 UTC m=+5876.212236583" lastFinishedPulling="2026-02-02 13:47:26.479901927 +0000 UTC m=+5877.708488790" observedRunningTime="2026-02-02 13:47:27.026350208 +0000 UTC m=+5878.254937121" watchObservedRunningTime="2026-02-02 13:47:27.039562059 +0000 UTC m=+5878.268148922" Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.050537 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.050508544 podStartE2EDuration="2.050508544s" podCreationTimestamp="2026-02-02 13:47:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:27.046883697 +0000 UTC m=+5878.275470570" watchObservedRunningTime="2026-02-02 13:47:27.050508544 +0000 UTC m=+5878.279095407" Feb 02 13:47:27 crc kubenswrapper[4846]: I0202 13:47:27.076939 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.076915096 podStartE2EDuration="2.076915096s" podCreationTimestamp="2026-02-02 13:47:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:27.070929271 +0000 UTC m=+5878.299516144" watchObservedRunningTime="2026-02-02 13:47:27.076915096 +0000 UTC m=+5878.305501959" Feb 02 13:47:30 crc kubenswrapper[4846]: I0202 13:47:30.413102 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 13:47:30 crc kubenswrapper[4846]: I0202 13:47:30.413891 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 13:47:32 crc kubenswrapper[4846]: I0202 13:47:32.424516 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:47:32 crc kubenswrapper[4846]: E0202 13:47:32.426126 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:47:33 crc kubenswrapper[4846]: I0202 13:47:33.613839 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:33 crc kubenswrapper[4846]: I0202 13:47:33.613899 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:33 crc kubenswrapper[4846]: I0202 13:47:33.659351 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:34 crc kubenswrapper[4846]: I0202 13:47:34.130105 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:34 crc kubenswrapper[4846]: I0202 13:47:34.176063 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6mxjq"] Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.094177 4846 generic.go:334] "Generic (PLEG): container finished" podID="ce616c5c-5881-4cfc-934d-b1c789d76e46" containerID="57e7b6c791e1736f567e02efac024b4da6aebcc1959db3ab46ca55884c16cded" exitCode=137 Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.094959 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce616c5c-5881-4cfc-934d-b1c789d76e46","Type":"ContainerDied","Data":"57e7b6c791e1736f567e02efac024b4da6aebcc1959db3ab46ca55884c16cded"} Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.318120 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.413807 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.413984 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.421707 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.422909 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.474276 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-combined-ca-bundle\") pod \"ce616c5c-5881-4cfc-934d-b1c789d76e46\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.474346 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s8s5\" (UniqueName: \"kubernetes.io/projected/ce616c5c-5881-4cfc-934d-b1c789d76e46-kube-api-access-9s8s5\") pod \"ce616c5c-5881-4cfc-934d-b1c789d76e46\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.474366 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-config-data\") pod \"ce616c5c-5881-4cfc-934d-b1c789d76e46\" (UID: \"ce616c5c-5881-4cfc-934d-b1c789d76e46\") " Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.479953 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce616c5c-5881-4cfc-934d-b1c789d76e46-kube-api-access-9s8s5" (OuterVolumeSpecName: "kube-api-access-9s8s5") pod "ce616c5c-5881-4cfc-934d-b1c789d76e46" (UID: "ce616c5c-5881-4cfc-934d-b1c789d76e46"). InnerVolumeSpecName "kube-api-access-9s8s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.504635 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce616c5c-5881-4cfc-934d-b1c789d76e46" (UID: "ce616c5c-5881-4cfc-934d-b1c789d76e46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.506161 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-config-data" (OuterVolumeSpecName: "config-data") pod "ce616c5c-5881-4cfc-934d-b1c789d76e46" (UID: "ce616c5c-5881-4cfc-934d-b1c789d76e46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.578357 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s8s5\" (UniqueName: \"kubernetes.io/projected/ce616c5c-5881-4cfc-934d-b1c789d76e46-kube-api-access-9s8s5\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.578551 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:35 crc kubenswrapper[4846]: I0202 13:47:35.578681 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce616c5c-5881-4cfc-934d-b1c789d76e46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.106451 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6mxjq" podUID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerName="registry-server" containerID="cri-o://fa15995595e1388630572ec7a422986f43a6367ab7ec6952e80248bb3b6532b9" gracePeriod=2 Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.107882 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.123761 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce616c5c-5881-4cfc-934d-b1c789d76e46","Type":"ContainerDied","Data":"1791a921e29367f6e09c5f6cf09ec5fb5bb6a2b2128a6248b4f1d90c65ea2abb"} Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.123836 4846 scope.go:117] "RemoveContainer" containerID="57e7b6c791e1736f567e02efac024b4da6aebcc1959db3ab46ca55884c16cded" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.156131 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.179904 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.205084 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 13:47:36 crc kubenswrapper[4846]: E0202 13:47:36.205576 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce616c5c-5881-4cfc-934d-b1c789d76e46" containerName="nova-scheduler-scheduler" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.205610 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce616c5c-5881-4cfc-934d-b1c789d76e46" containerName="nova-scheduler-scheduler" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.205810 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce616c5c-5881-4cfc-934d-b1c789d76e46" containerName="nova-scheduler-scheduler" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.206470 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.211076 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.215012 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.292778 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.292836 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22qxf\" (UniqueName: \"kubernetes.io/projected/1977e95e-eac6-462d-b7fd-caacae087fee-kube-api-access-22qxf\") pod \"nova-scheduler-0\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.292924 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-config-data\") pod \"nova-scheduler-0\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.394525 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.394565 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22qxf\" (UniqueName: \"kubernetes.io/projected/1977e95e-eac6-462d-b7fd-caacae087fee-kube-api-access-22qxf\") pod \"nova-scheduler-0\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.394601 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-config-data\") pod \"nova-scheduler-0\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.411393 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-config-data\") pod \"nova-scheduler-0\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.414578 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22qxf\" (UniqueName: \"kubernetes.io/projected/1977e95e-eac6-462d-b7fd-caacae087fee-kube-api-access-22qxf\") pod \"nova-scheduler-0\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.417666 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " pod="openstack/nova-scheduler-0" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.426957 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.93:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.427013 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.93:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.509935 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.94:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.509963 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.94:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 13:47:36 crc kubenswrapper[4846]: I0202 13:47:36.643202 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.124524 4846 generic.go:334] "Generic (PLEG): container finished" podID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerID="fa15995595e1388630572ec7a422986f43a6367ab7ec6952e80248bb3b6532b9" exitCode=0 Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.125186 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mxjq" event={"ID":"60e91bc7-6a91-4abd-9ea3-e5e462a682cd","Type":"ContainerDied","Data":"fa15995595e1388630572ec7a422986f43a6367ab7ec6952e80248bb3b6532b9"} Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.142112 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.202235 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 13:47:37 crc kubenswrapper[4846]: W0202 13:47:37.205225 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1977e95e_eac6_462d_b7fd_caacae087fee.slice/crio-ef7013589d21129eb3f31931366d245b7cb70e18a51784ea043b1ad249742e49 WatchSource:0}: Error finding container ef7013589d21129eb3f31931366d245b7cb70e18a51784ea043b1ad249742e49: Status 404 returned error can't find the container with id ef7013589d21129eb3f31931366d245b7cb70e18a51784ea043b1ad249742e49 Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.216791 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t78ff\" (UniqueName: \"kubernetes.io/projected/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-kube-api-access-t78ff\") pod \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.216913 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-catalog-content\") pod \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.217023 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-utilities\") pod \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\" (UID: \"60e91bc7-6a91-4abd-9ea3-e5e462a682cd\") " Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.218154 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-utilities" (OuterVolumeSpecName: "utilities") pod "60e91bc7-6a91-4abd-9ea3-e5e462a682cd" (UID: "60e91bc7-6a91-4abd-9ea3-e5e462a682cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.223305 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-kube-api-access-t78ff" (OuterVolumeSpecName: "kube-api-access-t78ff") pod "60e91bc7-6a91-4abd-9ea3-e5e462a682cd" (UID: "60e91bc7-6a91-4abd-9ea3-e5e462a682cd"). InnerVolumeSpecName "kube-api-access-t78ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.288339 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60e91bc7-6a91-4abd-9ea3-e5e462a682cd" (UID: "60e91bc7-6a91-4abd-9ea3-e5e462a682cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.319733 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.319775 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.319789 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t78ff\" (UniqueName: \"kubernetes.io/projected/60e91bc7-6a91-4abd-9ea3-e5e462a682cd-kube-api-access-t78ff\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:37 crc kubenswrapper[4846]: I0202 13:47:37.438986 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce616c5c-5881-4cfc-934d-b1c789d76e46" path="/var/lib/kubelet/pods/ce616c5c-5881-4cfc-934d-b1c789d76e46/volumes" Feb 02 13:47:38 crc kubenswrapper[4846]: I0202 13:47:38.139602 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mxjq" event={"ID":"60e91bc7-6a91-4abd-9ea3-e5e462a682cd","Type":"ContainerDied","Data":"1680972c833ccddd850a04f09172b36af291b967963b7170643d08adf3528b6c"} Feb 02 13:47:38 crc kubenswrapper[4846]: I0202 13:47:38.139689 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mxjq" Feb 02 13:47:38 crc kubenswrapper[4846]: I0202 13:47:38.139922 4846 scope.go:117] "RemoveContainer" containerID="fa15995595e1388630572ec7a422986f43a6367ab7ec6952e80248bb3b6532b9" Feb 02 13:47:38 crc kubenswrapper[4846]: I0202 13:47:38.142476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1977e95e-eac6-462d-b7fd-caacae087fee","Type":"ContainerStarted","Data":"861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923"} Feb 02 13:47:38 crc kubenswrapper[4846]: I0202 13:47:38.142527 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1977e95e-eac6-462d-b7fd-caacae087fee","Type":"ContainerStarted","Data":"ef7013589d21129eb3f31931366d245b7cb70e18a51784ea043b1ad249742e49"} Feb 02 13:47:38 crc kubenswrapper[4846]: I0202 13:47:38.166766 4846 scope.go:117] "RemoveContainer" containerID="b613a7f558aff9f9c293db50d57b1e92483c5da503087919be60368dadf69889" Feb 02 13:47:38 crc kubenswrapper[4846]: I0202 13:47:38.166886 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.166862762 podStartE2EDuration="2.166862762s" podCreationTimestamp="2026-02-02 13:47:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:38.161396619 +0000 UTC m=+5889.389983492" watchObservedRunningTime="2026-02-02 13:47:38.166862762 +0000 UTC m=+5889.395449625" Feb 02 13:47:38 crc kubenswrapper[4846]: I0202 13:47:38.191604 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6mxjq"] Feb 02 13:47:38 crc kubenswrapper[4846]: I0202 13:47:38.202025 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6mxjq"] Feb 02 13:47:38 crc kubenswrapper[4846]: I0202 13:47:38.218532 4846 scope.go:117] "RemoveContainer" containerID="46c1842de9d2086d9301be32aff44bc2987c51c4aed2eb870061fdd7b86bf861" Feb 02 13:47:39 crc kubenswrapper[4846]: I0202 13:47:39.446377 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" path="/var/lib/kubelet/pods/60e91bc7-6a91-4abd-9ea3-e5e462a682cd/volumes" Feb 02 13:47:41 crc kubenswrapper[4846]: I0202 13:47:41.643636 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.420195 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.420874 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.438302 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.438443 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.438527 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.438706 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.439468 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.441648 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.443670 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.447354 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.698541 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bbcfcc989-9vcqg"] Feb 02 13:47:45 crc kubenswrapper[4846]: E0202 13:47:45.698958 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerName="extract-content" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.698974 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerName="extract-content" Feb 02 13:47:45 crc kubenswrapper[4846]: E0202 13:47:45.698998 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerName="extract-utilities" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.699004 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerName="extract-utilities" Feb 02 13:47:45 crc kubenswrapper[4846]: E0202 13:47:45.699015 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerName="registry-server" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.699021 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerName="registry-server" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.699200 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="60e91bc7-6a91-4abd-9ea3-e5e462a682cd" containerName="registry-server" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.700107 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.710481 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bbcfcc989-9vcqg"] Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.775325 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-nb\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.775454 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsqsn\" (UniqueName: \"kubernetes.io/projected/3c239fe3-c133-4207-bc70-846b34c7859d-kube-api-access-rsqsn\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.775484 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-sb\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.775562 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-dns-svc\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.775659 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-config\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.877851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsqsn\" (UniqueName: \"kubernetes.io/projected/3c239fe3-c133-4207-bc70-846b34c7859d-kube-api-access-rsqsn\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.877915 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-sb\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.878028 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-dns-svc\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.878094 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-config\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.878149 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-nb\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.879290 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-sb\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.879302 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-dns-svc\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.879308 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-config\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.879341 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-nb\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:45 crc kubenswrapper[4846]: I0202 13:47:45.896609 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsqsn\" (UniqueName: \"kubernetes.io/projected/3c239fe3-c133-4207-bc70-846b34c7859d-kube-api-access-rsqsn\") pod \"dnsmasq-dns-7bbcfcc989-9vcqg\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:46 crc kubenswrapper[4846]: I0202 13:47:46.028259 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:46 crc kubenswrapper[4846]: I0202 13:47:46.507571 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bbcfcc989-9vcqg"] Feb 02 13:47:46 crc kubenswrapper[4846]: I0202 13:47:46.643665 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 02 13:47:46 crc kubenswrapper[4846]: I0202 13:47:46.676559 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 02 13:47:47 crc kubenswrapper[4846]: I0202 13:47:47.230576 4846 generic.go:334] "Generic (PLEG): container finished" podID="3c239fe3-c133-4207-bc70-846b34c7859d" containerID="da476008128d2691003257fc23608aa5cc53e25a6d2fccac873c7220cbd57440" exitCode=0 Feb 02 13:47:47 crc kubenswrapper[4846]: I0202 13:47:47.230666 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" event={"ID":"3c239fe3-c133-4207-bc70-846b34c7859d","Type":"ContainerDied","Data":"da476008128d2691003257fc23608aa5cc53e25a6d2fccac873c7220cbd57440"} Feb 02 13:47:47 crc kubenswrapper[4846]: I0202 13:47:47.231048 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" event={"ID":"3c239fe3-c133-4207-bc70-846b34c7859d","Type":"ContainerStarted","Data":"31d9a5f42b6af94b78d93773e076df22cc1db37a53a97496027d2118b8edc5f8"} Feb 02 13:47:47 crc kubenswrapper[4846]: I0202 13:47:47.278042 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 02 13:47:47 crc kubenswrapper[4846]: I0202 13:47:47.423819 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:47:47 crc kubenswrapper[4846]: E0202 13:47:47.424121 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:47:48 crc kubenswrapper[4846]: I0202 13:47:48.241764 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" event={"ID":"3c239fe3-c133-4207-bc70-846b34c7859d","Type":"ContainerStarted","Data":"610b151d0fabfd65c7b70a049a079ee2378e097c251eeebc0aaf8a165258dfa9"} Feb 02 13:47:48 crc kubenswrapper[4846]: I0202 13:47:48.241983 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:48 crc kubenswrapper[4846]: I0202 13:47:48.266256 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" podStartSLOduration=3.266238783 podStartE2EDuration="3.266238783s" podCreationTimestamp="2026-02-02 13:47:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:48.258734771 +0000 UTC m=+5899.487321634" watchObservedRunningTime="2026-02-02 13:47:48.266238783 +0000 UTC m=+5899.494825646" Feb 02 13:47:48 crc kubenswrapper[4846]: I0202 13:47:48.573341 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:48 crc kubenswrapper[4846]: I0202 13:47:48.573724 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerName="nova-api-log" containerID="cri-o://dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b" gracePeriod=30 Feb 02 13:47:48 crc kubenswrapper[4846]: I0202 13:47:48.573977 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerName="nova-api-api" containerID="cri-o://25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199" gracePeriod=30 Feb 02 13:47:49 crc kubenswrapper[4846]: I0202 13:47:49.250237 4846 generic.go:334] "Generic (PLEG): container finished" podID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerID="dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b" exitCode=143 Feb 02 13:47:49 crc kubenswrapper[4846]: I0202 13:47:49.250313 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"496f99aa-8e97-489f-8957-7deaba86dc1a","Type":"ContainerDied","Data":"dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b"} Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.203495 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.299064 4846 generic.go:334] "Generic (PLEG): container finished" podID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerID="25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199" exitCode=0 Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.299114 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"496f99aa-8e97-489f-8957-7deaba86dc1a","Type":"ContainerDied","Data":"25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199"} Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.299147 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"496f99aa-8e97-489f-8957-7deaba86dc1a","Type":"ContainerDied","Data":"9d9d802f6fecf6ce509abbb42f3bf13819eb700b3cbe6c88253abf006cde10a7"} Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.299167 4846 scope.go:117] "RemoveContainer" containerID="25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.299347 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.304736 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh5zw\" (UniqueName: \"kubernetes.io/projected/496f99aa-8e97-489f-8957-7deaba86dc1a-kube-api-access-gh5zw\") pod \"496f99aa-8e97-489f-8957-7deaba86dc1a\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.304933 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-combined-ca-bundle\") pod \"496f99aa-8e97-489f-8957-7deaba86dc1a\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.305008 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-config-data\") pod \"496f99aa-8e97-489f-8957-7deaba86dc1a\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.305075 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/496f99aa-8e97-489f-8957-7deaba86dc1a-logs\") pod \"496f99aa-8e97-489f-8957-7deaba86dc1a\" (UID: \"496f99aa-8e97-489f-8957-7deaba86dc1a\") " Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.306295 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/496f99aa-8e97-489f-8957-7deaba86dc1a-logs" (OuterVolumeSpecName: "logs") pod "496f99aa-8e97-489f-8957-7deaba86dc1a" (UID: "496f99aa-8e97-489f-8957-7deaba86dc1a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.315859 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496f99aa-8e97-489f-8957-7deaba86dc1a-kube-api-access-gh5zw" (OuterVolumeSpecName: "kube-api-access-gh5zw") pod "496f99aa-8e97-489f-8957-7deaba86dc1a" (UID: "496f99aa-8e97-489f-8957-7deaba86dc1a"). InnerVolumeSpecName "kube-api-access-gh5zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.354269 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "496f99aa-8e97-489f-8957-7deaba86dc1a" (UID: "496f99aa-8e97-489f-8957-7deaba86dc1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.372852 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-config-data" (OuterVolumeSpecName: "config-data") pod "496f99aa-8e97-489f-8957-7deaba86dc1a" (UID: "496f99aa-8e97-489f-8957-7deaba86dc1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.407243 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh5zw\" (UniqueName: \"kubernetes.io/projected/496f99aa-8e97-489f-8957-7deaba86dc1a-kube-api-access-gh5zw\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.407277 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.407287 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/496f99aa-8e97-489f-8957-7deaba86dc1a-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.407296 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/496f99aa-8e97-489f-8957-7deaba86dc1a-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.434169 4846 scope.go:117] "RemoveContainer" containerID="dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.469123 4846 scope.go:117] "RemoveContainer" containerID="25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199" Feb 02 13:47:52 crc kubenswrapper[4846]: E0202 13:47:52.469591 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199\": container with ID starting with 25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199 not found: ID does not exist" containerID="25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.469661 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199"} err="failed to get container status \"25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199\": rpc error: code = NotFound desc = could not find container \"25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199\": container with ID starting with 25ea5546cccb3bc8e7780c6cfeabd8e46d7626221e62fbd2afea9be0ead4e199 not found: ID does not exist" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.469687 4846 scope.go:117] "RemoveContainer" containerID="dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b" Feb 02 13:47:52 crc kubenswrapper[4846]: E0202 13:47:52.470603 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b\": container with ID starting with dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b not found: ID does not exist" containerID="dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.470654 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b"} err="failed to get container status \"dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b\": rpc error: code = NotFound desc = could not find container \"dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b\": container with ID starting with dea3a330b48d1a46b9e3c126c6411b79a49ae0a207fc5cd66657be5e5336638b not found: ID does not exist" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.631917 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.638897 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.663588 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:52 crc kubenswrapper[4846]: E0202 13:47:52.664002 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerName="nova-api-log" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.664023 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerName="nova-api-log" Feb 02 13:47:52 crc kubenswrapper[4846]: E0202 13:47:52.664039 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerName="nova-api-api" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.664046 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerName="nova-api-api" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.664231 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerName="nova-api-api" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.664271 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" containerName="nova-api-log" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.665292 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.667377 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.667536 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.669355 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.684010 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.813698 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.813779 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-public-tls-certs\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.813809 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.813847 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39fdb60a-d87e-438c-8de9-63e49fdda9b8-logs\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.813872 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-config-data\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.813908 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czqdn\" (UniqueName: \"kubernetes.io/projected/39fdb60a-d87e-438c-8de9-63e49fdda9b8-kube-api-access-czqdn\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.915478 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-config-data\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.915568 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czqdn\" (UniqueName: \"kubernetes.io/projected/39fdb60a-d87e-438c-8de9-63e49fdda9b8-kube-api-access-czqdn\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.915717 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.915801 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-public-tls-certs\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.915839 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.915893 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39fdb60a-d87e-438c-8de9-63e49fdda9b8-logs\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.916565 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39fdb60a-d87e-438c-8de9-63e49fdda9b8-logs\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.920474 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.920482 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-config-data\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.921729 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-public-tls-certs\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.929445 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.933598 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czqdn\" (UniqueName: \"kubernetes.io/projected/39fdb60a-d87e-438c-8de9-63e49fdda9b8-kube-api-access-czqdn\") pod \"nova-api-0\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " pod="openstack/nova-api-0" Feb 02 13:47:52 crc kubenswrapper[4846]: I0202 13:47:52.984330 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 13:47:53 crc kubenswrapper[4846]: I0202 13:47:53.442034 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496f99aa-8e97-489f-8957-7deaba86dc1a" path="/var/lib/kubelet/pods/496f99aa-8e97-489f-8957-7deaba86dc1a/volumes" Feb 02 13:47:53 crc kubenswrapper[4846]: I0202 13:47:53.460764 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 13:47:53 crc kubenswrapper[4846]: W0202 13:47:53.461057 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39fdb60a_d87e_438c_8de9_63e49fdda9b8.slice/crio-f7b53f6664b4d1c22e2151def81b3e2de229c85f54cfba100ed9cdd7b2f467b3 WatchSource:0}: Error finding container f7b53f6664b4d1c22e2151def81b3e2de229c85f54cfba100ed9cdd7b2f467b3: Status 404 returned error can't find the container with id f7b53f6664b4d1c22e2151def81b3e2de229c85f54cfba100ed9cdd7b2f467b3 Feb 02 13:47:54 crc kubenswrapper[4846]: I0202 13:47:54.318587 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"39fdb60a-d87e-438c-8de9-63e49fdda9b8","Type":"ContainerStarted","Data":"42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c"} Feb 02 13:47:54 crc kubenswrapper[4846]: I0202 13:47:54.318986 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"39fdb60a-d87e-438c-8de9-63e49fdda9b8","Type":"ContainerStarted","Data":"a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3"} Feb 02 13:47:54 crc kubenswrapper[4846]: I0202 13:47:54.319002 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"39fdb60a-d87e-438c-8de9-63e49fdda9b8","Type":"ContainerStarted","Data":"f7b53f6664b4d1c22e2151def81b3e2de229c85f54cfba100ed9cdd7b2f467b3"} Feb 02 13:47:54 crc kubenswrapper[4846]: I0202 13:47:54.335971 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.33595066 podStartE2EDuration="2.33595066s" podCreationTimestamp="2026-02-02 13:47:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:47:54.335714404 +0000 UTC m=+5905.564301267" watchObservedRunningTime="2026-02-02 13:47:54.33595066 +0000 UTC m=+5905.564537533" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.029849 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.118526 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8555c84f6c-52q4p"] Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.118804 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" podUID="db6600ab-c487-4e40-ace0-f51e528c46af" containerName="dnsmasq-dns" containerID="cri-o://f9c86bdd27617ceea4f5372425c2905824471788c126325fedd4cf403b48495a" gracePeriod=10 Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.344924 4846 generic.go:334] "Generic (PLEG): container finished" podID="db6600ab-c487-4e40-ace0-f51e528c46af" containerID="f9c86bdd27617ceea4f5372425c2905824471788c126325fedd4cf403b48495a" exitCode=0 Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.344967 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" event={"ID":"db6600ab-c487-4e40-ace0-f51e528c46af","Type":"ContainerDied","Data":"f9c86bdd27617ceea4f5372425c2905824471788c126325fedd4cf403b48495a"} Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.663467 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.815222 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-sb\") pod \"db6600ab-c487-4e40-ace0-f51e528c46af\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.815299 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-config\") pod \"db6600ab-c487-4e40-ace0-f51e528c46af\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.815409 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-nb\") pod \"db6600ab-c487-4e40-ace0-f51e528c46af\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.815533 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8ltz\" (UniqueName: \"kubernetes.io/projected/db6600ab-c487-4e40-ace0-f51e528c46af-kube-api-access-j8ltz\") pod \"db6600ab-c487-4e40-ace0-f51e528c46af\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.815664 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-dns-svc\") pod \"db6600ab-c487-4e40-ace0-f51e528c46af\" (UID: \"db6600ab-c487-4e40-ace0-f51e528c46af\") " Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.825365 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db6600ab-c487-4e40-ace0-f51e528c46af-kube-api-access-j8ltz" (OuterVolumeSpecName: "kube-api-access-j8ltz") pod "db6600ab-c487-4e40-ace0-f51e528c46af" (UID: "db6600ab-c487-4e40-ace0-f51e528c46af"). InnerVolumeSpecName "kube-api-access-j8ltz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.867050 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-config" (OuterVolumeSpecName: "config") pod "db6600ab-c487-4e40-ace0-f51e528c46af" (UID: "db6600ab-c487-4e40-ace0-f51e528c46af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.869317 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "db6600ab-c487-4e40-ace0-f51e528c46af" (UID: "db6600ab-c487-4e40-ace0-f51e528c46af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.873140 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "db6600ab-c487-4e40-ace0-f51e528c46af" (UID: "db6600ab-c487-4e40-ace0-f51e528c46af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.879977 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "db6600ab-c487-4e40-ace0-f51e528c46af" (UID: "db6600ab-c487-4e40-ace0-f51e528c46af"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.918931 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.919170 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.919232 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.919409 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db6600ab-c487-4e40-ace0-f51e528c46af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:56 crc kubenswrapper[4846]: I0202 13:47:56.919473 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8ltz\" (UniqueName: \"kubernetes.io/projected/db6600ab-c487-4e40-ace0-f51e528c46af-kube-api-access-j8ltz\") on node \"crc\" DevicePath \"\"" Feb 02 13:47:57 crc kubenswrapper[4846]: I0202 13:47:57.357057 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" event={"ID":"db6600ab-c487-4e40-ace0-f51e528c46af","Type":"ContainerDied","Data":"8d8cc5e042e621946ff9993859f09a2ce32692e95f79ee89d4081c08b09acc69"} Feb 02 13:47:57 crc kubenswrapper[4846]: I0202 13:47:57.357120 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8555c84f6c-52q4p" Feb 02 13:47:57 crc kubenswrapper[4846]: I0202 13:47:57.357133 4846 scope.go:117] "RemoveContainer" containerID="f9c86bdd27617ceea4f5372425c2905824471788c126325fedd4cf403b48495a" Feb 02 13:47:57 crc kubenswrapper[4846]: I0202 13:47:57.392980 4846 scope.go:117] "RemoveContainer" containerID="0cb523451a149729826966a1d66a091f3787fc72dad8c7e597763e0070bb447c" Feb 02 13:47:57 crc kubenswrapper[4846]: I0202 13:47:57.395959 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8555c84f6c-52q4p"] Feb 02 13:47:57 crc kubenswrapper[4846]: I0202 13:47:57.406454 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8555c84f6c-52q4p"] Feb 02 13:47:57 crc kubenswrapper[4846]: I0202 13:47:57.453092 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db6600ab-c487-4e40-ace0-f51e528c46af" path="/var/lib/kubelet/pods/db6600ab-c487-4e40-ace0-f51e528c46af/volumes" Feb 02 13:47:59 crc kubenswrapper[4846]: I0202 13:47:59.430832 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:47:59 crc kubenswrapper[4846]: E0202 13:47:59.431913 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:48:02 crc kubenswrapper[4846]: I0202 13:48:02.985128 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 13:48:02 crc kubenswrapper[4846]: I0202 13:48:02.985399 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 13:48:03 crc kubenswrapper[4846]: I0202 13:48:03.995868 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.97:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 13:48:03 crc kubenswrapper[4846]: I0202 13:48:03.995929 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.97:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 13:48:10 crc kubenswrapper[4846]: I0202 13:48:10.423566 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:48:10 crc kubenswrapper[4846]: E0202 13:48:10.424524 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:48:12 crc kubenswrapper[4846]: I0202 13:48:12.990816 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 13:48:12 crc kubenswrapper[4846]: I0202 13:48:12.991436 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 13:48:12 crc kubenswrapper[4846]: I0202 13:48:12.996283 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 13:48:12 crc kubenswrapper[4846]: I0202 13:48:12.999695 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 13:48:13 crc kubenswrapper[4846]: I0202 13:48:13.489753 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 13:48:13 crc kubenswrapper[4846]: I0202 13:48:13.500424 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 13:48:24 crc kubenswrapper[4846]: I0202 13:48:24.423718 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:48:24 crc kubenswrapper[4846]: E0202 13:48:24.424556 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.423297 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.725357 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"fcae4c79abc512d4d8a5d4e3219b87abc67dad0f182ce8895f815c1ebd096f68"} Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.938234 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-v8bjp"] Feb 02 13:48:38 crc kubenswrapper[4846]: E0202 13:48:38.938931 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6600ab-c487-4e40-ace0-f51e528c46af" containerName="init" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.939266 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6600ab-c487-4e40-ace0-f51e528c46af" containerName="init" Feb 02 13:48:38 crc kubenswrapper[4846]: E0202 13:48:38.939282 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6600ab-c487-4e40-ace0-f51e528c46af" containerName="dnsmasq-dns" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.939291 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6600ab-c487-4e40-ace0-f51e528c46af" containerName="dnsmasq-dns" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.939536 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="db6600ab-c487-4e40-ace0-f51e528c46af" containerName="dnsmasq-dns" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.940389 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.944797 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.945409 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-fpxgs" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.949508 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.952382 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-v8bjp"] Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.964162 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-mkdn9"] Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.966561 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:38 crc kubenswrapper[4846]: I0202 13:48:38.972978 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mkdn9"] Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.037780 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79522d21-4471-4cdb-8caa-41c17214b4f1-combined-ca-bundle\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.037830 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79522d21-4471-4cdb-8caa-41c17214b4f1-var-log-ovn\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.037875 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-scripts\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.037951 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-var-run\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.037988 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79522d21-4471-4cdb-8caa-41c17214b4f1-scripts\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.038024 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-etc-ovs\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.038085 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-var-log\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.038117 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79522d21-4471-4cdb-8caa-41c17214b4f1-var-run\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.038139 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qncz7\" (UniqueName: \"kubernetes.io/projected/79522d21-4471-4cdb-8caa-41c17214b4f1-kube-api-access-qncz7\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.038163 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79522d21-4471-4cdb-8caa-41c17214b4f1-var-run-ovn\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.039228 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79522d21-4471-4cdb-8caa-41c17214b4f1-ovn-controller-tls-certs\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.039281 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxgzk\" (UniqueName: \"kubernetes.io/projected/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-kube-api-access-gxgzk\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.039390 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-var-lib\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.094350 4846 scope.go:117] "RemoveContainer" containerID="a47babc40ed3ee554051885b0079ab05b8c4af6cd784a66c8b0f4211a8ee9aa5" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.126886 4846 scope.go:117] "RemoveContainer" containerID="a5dc7bfb0f923103f317e2f5f0ef410332dc55538a35f56ab9f3e9af692dc182" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141091 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-etc-ovs\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-var-log\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141182 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79522d21-4471-4cdb-8caa-41c17214b4f1-var-run\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141200 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qncz7\" (UniqueName: \"kubernetes.io/projected/79522d21-4471-4cdb-8caa-41c17214b4f1-kube-api-access-qncz7\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141220 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79522d21-4471-4cdb-8caa-41c17214b4f1-var-run-ovn\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141245 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79522d21-4471-4cdb-8caa-41c17214b4f1-ovn-controller-tls-certs\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141264 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxgzk\" (UniqueName: \"kubernetes.io/projected/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-kube-api-access-gxgzk\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141307 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-var-lib\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141346 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79522d21-4471-4cdb-8caa-41c17214b4f1-combined-ca-bundle\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141365 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79522d21-4471-4cdb-8caa-41c17214b4f1-var-log-ovn\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141388 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-scripts\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141413 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-var-run\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141439 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79522d21-4471-4cdb-8caa-41c17214b4f1-scripts\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.141486 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-etc-ovs\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.142230 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79522d21-4471-4cdb-8caa-41c17214b4f1-var-log-ovn\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.142339 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-var-lib\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.142443 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-var-log\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.142510 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79522d21-4471-4cdb-8caa-41c17214b4f1-var-run\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.142730 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79522d21-4471-4cdb-8caa-41c17214b4f1-var-run-ovn\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.143717 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79522d21-4471-4cdb-8caa-41c17214b4f1-scripts\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.144275 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-var-run\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.145449 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-scripts\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.150226 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79522d21-4471-4cdb-8caa-41c17214b4f1-combined-ca-bundle\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.156421 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79522d21-4471-4cdb-8caa-41c17214b4f1-ovn-controller-tls-certs\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.160155 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxgzk\" (UniqueName: \"kubernetes.io/projected/f7c4abfa-9b37-4c15-accf-fcb0007b4d26-kube-api-access-gxgzk\") pod \"ovn-controller-ovs-mkdn9\" (UID: \"f7c4abfa-9b37-4c15-accf-fcb0007b4d26\") " pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.160901 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qncz7\" (UniqueName: \"kubernetes.io/projected/79522d21-4471-4cdb-8caa-41c17214b4f1-kube-api-access-qncz7\") pod \"ovn-controller-v8bjp\" (UID: \"79522d21-4471-4cdb-8caa-41c17214b4f1\") " pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.264985 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.317922 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:39 crc kubenswrapper[4846]: I0202 13:48:39.783949 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-v8bjp"] Feb 02 13:48:40 crc kubenswrapper[4846]: W0202 13:48:40.217385 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c4abfa_9b37_4c15_accf_fcb0007b4d26.slice/crio-3197f075a0e59beda281d8a37bf1386ecf11f5f489276da50295fdc923c29aec WatchSource:0}: Error finding container 3197f075a0e59beda281d8a37bf1386ecf11f5f489276da50295fdc923c29aec: Status 404 returned error can't find the container with id 3197f075a0e59beda281d8a37bf1386ecf11f5f489276da50295fdc923c29aec Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.219744 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mkdn9"] Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.297021 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-hnvm7"] Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.298388 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.303370 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.307734 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hnvm7"] Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.367246 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7dd42298-2b9c-4382-b25e-9afe253d226f-ovn-rundir\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.367304 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjqcj\" (UniqueName: \"kubernetes.io/projected/7dd42298-2b9c-4382-b25e-9afe253d226f-kube-api-access-bjqcj\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.367490 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dd42298-2b9c-4382-b25e-9afe253d226f-config\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.367552 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd42298-2b9c-4382-b25e-9afe253d226f-combined-ca-bundle\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.367840 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dd42298-2b9c-4382-b25e-9afe253d226f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.367892 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7dd42298-2b9c-4382-b25e-9afe253d226f-ovs-rundir\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.469950 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dd42298-2b9c-4382-b25e-9afe253d226f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.470263 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7dd42298-2b9c-4382-b25e-9afe253d226f-ovs-rundir\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.470341 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7dd42298-2b9c-4382-b25e-9afe253d226f-ovn-rundir\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.470408 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjqcj\" (UniqueName: \"kubernetes.io/projected/7dd42298-2b9c-4382-b25e-9afe253d226f-kube-api-access-bjqcj\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.470485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dd42298-2b9c-4382-b25e-9afe253d226f-config\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.470506 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd42298-2b9c-4382-b25e-9afe253d226f-combined-ca-bundle\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.472471 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7dd42298-2b9c-4382-b25e-9afe253d226f-ovs-rundir\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.472523 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7dd42298-2b9c-4382-b25e-9afe253d226f-ovn-rundir\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.473208 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dd42298-2b9c-4382-b25e-9afe253d226f-config\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.483826 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dd42298-2b9c-4382-b25e-9afe253d226f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.489338 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd42298-2b9c-4382-b25e-9afe253d226f-combined-ca-bundle\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.492089 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjqcj\" (UniqueName: \"kubernetes.io/projected/7dd42298-2b9c-4382-b25e-9afe253d226f-kube-api-access-bjqcj\") pod \"ovn-controller-metrics-hnvm7\" (UID: \"7dd42298-2b9c-4382-b25e-9afe253d226f\") " pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.662828 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hnvm7" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.751814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v8bjp" event={"ID":"79522d21-4471-4cdb-8caa-41c17214b4f1","Type":"ContainerStarted","Data":"794674fcbd362fe8c8b40cfc25c1da4b7e63159083bc56b27425bf7364fd016f"} Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.751862 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v8bjp" event={"ID":"79522d21-4471-4cdb-8caa-41c17214b4f1","Type":"ContainerStarted","Data":"85305f128042f935a452b9f491c72c6998dfd74b0dadc546c071d22152ba81f6"} Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.751898 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-v8bjp" Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.754502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mkdn9" event={"ID":"f7c4abfa-9b37-4c15-accf-fcb0007b4d26","Type":"ContainerStarted","Data":"69e181645dad0d9c436a6c27b63d0bbd6da974cb44d35b91e124f1916f33edd7"} Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.754549 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mkdn9" event={"ID":"f7c4abfa-9b37-4c15-accf-fcb0007b4d26","Type":"ContainerStarted","Data":"3197f075a0e59beda281d8a37bf1386ecf11f5f489276da50295fdc923c29aec"} Feb 02 13:48:40 crc kubenswrapper[4846]: I0202 13:48:40.779005 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-v8bjp" podStartSLOduration=2.778984119 podStartE2EDuration="2.778984119s" podCreationTimestamp="2026-02-02 13:48:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:48:40.776422298 +0000 UTC m=+5952.005009161" watchObservedRunningTime="2026-02-02 13:48:40.778984119 +0000 UTC m=+5952.007570982" Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.122428 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-9v7ph"] Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.123865 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-9v7ph" Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.140959 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-9v7ph"] Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.177786 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hnvm7"] Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.191331 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fk4d\" (UniqueName: \"kubernetes.io/projected/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-kube-api-access-8fk4d\") pod \"octavia-db-create-9v7ph\" (UID: \"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf\") " pod="openstack/octavia-db-create-9v7ph" Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.191495 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-operator-scripts\") pod \"octavia-db-create-9v7ph\" (UID: \"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf\") " pod="openstack/octavia-db-create-9v7ph" Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.293595 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fk4d\" (UniqueName: \"kubernetes.io/projected/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-kube-api-access-8fk4d\") pod \"octavia-db-create-9v7ph\" (UID: \"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf\") " pod="openstack/octavia-db-create-9v7ph" Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.293780 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-operator-scripts\") pod \"octavia-db-create-9v7ph\" (UID: \"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf\") " pod="openstack/octavia-db-create-9v7ph" Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.306062 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-operator-scripts\") pod \"octavia-db-create-9v7ph\" (UID: \"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf\") " pod="openstack/octavia-db-create-9v7ph" Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.311121 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fk4d\" (UniqueName: \"kubernetes.io/projected/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-kube-api-access-8fk4d\") pod \"octavia-db-create-9v7ph\" (UID: \"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf\") " pod="openstack/octavia-db-create-9v7ph" Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.439889 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-9v7ph" Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.779907 4846 generic.go:334] "Generic (PLEG): container finished" podID="f7c4abfa-9b37-4c15-accf-fcb0007b4d26" containerID="69e181645dad0d9c436a6c27b63d0bbd6da974cb44d35b91e124f1916f33edd7" exitCode=0 Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.780855 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mkdn9" event={"ID":"f7c4abfa-9b37-4c15-accf-fcb0007b4d26","Type":"ContainerDied","Data":"69e181645dad0d9c436a6c27b63d0bbd6da974cb44d35b91e124f1916f33edd7"} Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.800668 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hnvm7" event={"ID":"7dd42298-2b9c-4382-b25e-9afe253d226f","Type":"ContainerStarted","Data":"e5a6d823c6f940e7328411ed322d33b6f976e8b657d751b5fa827c00ca70bc56"} Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.800723 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hnvm7" event={"ID":"7dd42298-2b9c-4382-b25e-9afe253d226f","Type":"ContainerStarted","Data":"031ff975768dfa8fda722e6d308b613df3141ae4f75b78849edcac3b9b0bc0bb"} Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.857964 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-hnvm7" podStartSLOduration=1.857947703 podStartE2EDuration="1.857947703s" podCreationTimestamp="2026-02-02 13:48:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:48:41.85577795 +0000 UTC m=+5953.084364813" watchObservedRunningTime="2026-02-02 13:48:41.857947703 +0000 UTC m=+5953.086534566" Feb 02 13:48:41 crc kubenswrapper[4846]: I0202 13:48:41.956299 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-9v7ph"] Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.152397 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-06a1-account-create-update-jj9jt"] Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.156701 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-06a1-account-create-update-jj9jt" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.159840 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.163364 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-06a1-account-create-update-jj9jt"] Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.215566 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6hk9\" (UniqueName: \"kubernetes.io/projected/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-kube-api-access-b6hk9\") pod \"octavia-06a1-account-create-update-jj9jt\" (UID: \"94bb9e87-02ea-4165-bb0a-00acfd0c4c77\") " pod="openstack/octavia-06a1-account-create-update-jj9jt" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.215892 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-operator-scripts\") pod \"octavia-06a1-account-create-update-jj9jt\" (UID: \"94bb9e87-02ea-4165-bb0a-00acfd0c4c77\") " pod="openstack/octavia-06a1-account-create-update-jj9jt" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.317562 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6hk9\" (UniqueName: \"kubernetes.io/projected/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-kube-api-access-b6hk9\") pod \"octavia-06a1-account-create-update-jj9jt\" (UID: \"94bb9e87-02ea-4165-bb0a-00acfd0c4c77\") " pod="openstack/octavia-06a1-account-create-update-jj9jt" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.318198 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-operator-scripts\") pod \"octavia-06a1-account-create-update-jj9jt\" (UID: \"94bb9e87-02ea-4165-bb0a-00acfd0c4c77\") " pod="openstack/octavia-06a1-account-create-update-jj9jt" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.319452 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-operator-scripts\") pod \"octavia-06a1-account-create-update-jj9jt\" (UID: \"94bb9e87-02ea-4165-bb0a-00acfd0c4c77\") " pod="openstack/octavia-06a1-account-create-update-jj9jt" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.341479 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6hk9\" (UniqueName: \"kubernetes.io/projected/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-kube-api-access-b6hk9\") pod \"octavia-06a1-account-create-update-jj9jt\" (UID: \"94bb9e87-02ea-4165-bb0a-00acfd0c4c77\") " pod="openstack/octavia-06a1-account-create-update-jj9jt" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.481809 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-06a1-account-create-update-jj9jt" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.812889 4846 generic.go:334] "Generic (PLEG): container finished" podID="5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf" containerID="a2a9bcd1bf2742b52894d74d287ed03a580068e20aaf6d39d232ca19f9bfee85" exitCode=0 Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.813240 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-9v7ph" event={"ID":"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf","Type":"ContainerDied","Data":"a2a9bcd1bf2742b52894d74d287ed03a580068e20aaf6d39d232ca19f9bfee85"} Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.813272 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-9v7ph" event={"ID":"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf","Type":"ContainerStarted","Data":"06c7e303107a9b393f7b1a1d560e5471726d76d17e91c48d5000bd7af9154535"} Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.817715 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mkdn9" event={"ID":"f7c4abfa-9b37-4c15-accf-fcb0007b4d26","Type":"ContainerStarted","Data":"4fdc032f622b68d6fd989e4edd97d47c2cdc08c41f3cbe1a6a1bb36afabc0fd3"} Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.817757 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.817772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mkdn9" event={"ID":"f7c4abfa-9b37-4c15-accf-fcb0007b4d26","Type":"ContainerStarted","Data":"0bf3ee6253e8c28110f37f5602ffcd2138f3870249a366d57c30f3017affdda0"} Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.817800 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.862029 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-mkdn9" podStartSLOduration=4.862011997 podStartE2EDuration="4.862011997s" podCreationTimestamp="2026-02-02 13:48:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:48:42.855669574 +0000 UTC m=+5954.084256427" watchObservedRunningTime="2026-02-02 13:48:42.862011997 +0000 UTC m=+5954.090598860" Feb 02 13:48:42 crc kubenswrapper[4846]: W0202 13:48:42.943355 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94bb9e87_02ea_4165_bb0a_00acfd0c4c77.slice/crio-e0031575e3e73dda37ab552c7dae2f0667a7aadada427eecbd2a1d6185e55d1d WatchSource:0}: Error finding container e0031575e3e73dda37ab552c7dae2f0667a7aadada427eecbd2a1d6185e55d1d: Status 404 returned error can't find the container with id e0031575e3e73dda37ab552c7dae2f0667a7aadada427eecbd2a1d6185e55d1d Feb 02 13:48:42 crc kubenswrapper[4846]: I0202 13:48:42.944218 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-06a1-account-create-update-jj9jt"] Feb 02 13:48:43 crc kubenswrapper[4846]: I0202 13:48:43.828387 4846 generic.go:334] "Generic (PLEG): container finished" podID="94bb9e87-02ea-4165-bb0a-00acfd0c4c77" containerID="01df9df9171d818929a91e1f45eba8202cd2797e9cbcd282919bd1d1c1bb521c" exitCode=0 Feb 02 13:48:43 crc kubenswrapper[4846]: I0202 13:48:43.828439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-06a1-account-create-update-jj9jt" event={"ID":"94bb9e87-02ea-4165-bb0a-00acfd0c4c77","Type":"ContainerDied","Data":"01df9df9171d818929a91e1f45eba8202cd2797e9cbcd282919bd1d1c1bb521c"} Feb 02 13:48:43 crc kubenswrapper[4846]: I0202 13:48:43.829038 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-06a1-account-create-update-jj9jt" event={"ID":"94bb9e87-02ea-4165-bb0a-00acfd0c4c77","Type":"ContainerStarted","Data":"e0031575e3e73dda37ab552c7dae2f0667a7aadada427eecbd2a1d6185e55d1d"} Feb 02 13:48:44 crc kubenswrapper[4846]: I0202 13:48:44.159176 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-9v7ph" Feb 02 13:48:44 crc kubenswrapper[4846]: I0202 13:48:44.250566 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fk4d\" (UniqueName: \"kubernetes.io/projected/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-kube-api-access-8fk4d\") pod \"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf\" (UID: \"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf\") " Feb 02 13:48:44 crc kubenswrapper[4846]: I0202 13:48:44.250642 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-operator-scripts\") pod \"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf\" (UID: \"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf\") " Feb 02 13:48:44 crc kubenswrapper[4846]: I0202 13:48:44.251362 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf" (UID: "5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:48:44 crc kubenswrapper[4846]: I0202 13:48:44.256863 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-kube-api-access-8fk4d" (OuterVolumeSpecName: "kube-api-access-8fk4d") pod "5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf" (UID: "5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf"). InnerVolumeSpecName "kube-api-access-8fk4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:48:44 crc kubenswrapper[4846]: I0202 13:48:44.352209 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fk4d\" (UniqueName: \"kubernetes.io/projected/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-kube-api-access-8fk4d\") on node \"crc\" DevicePath \"\"" Feb 02 13:48:44 crc kubenswrapper[4846]: I0202 13:48:44.352252 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:48:44 crc kubenswrapper[4846]: I0202 13:48:44.842204 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-9v7ph" event={"ID":"5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf","Type":"ContainerDied","Data":"06c7e303107a9b393f7b1a1d560e5471726d76d17e91c48d5000bd7af9154535"} Feb 02 13:48:44 crc kubenswrapper[4846]: I0202 13:48:44.842240 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-9v7ph" Feb 02 13:48:44 crc kubenswrapper[4846]: I0202 13:48:44.842244 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06c7e303107a9b393f7b1a1d560e5471726d76d17e91c48d5000bd7af9154535" Feb 02 13:48:45 crc kubenswrapper[4846]: I0202 13:48:45.196911 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-06a1-account-create-update-jj9jt" Feb 02 13:48:45 crc kubenswrapper[4846]: I0202 13:48:45.273367 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6hk9\" (UniqueName: \"kubernetes.io/projected/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-kube-api-access-b6hk9\") pod \"94bb9e87-02ea-4165-bb0a-00acfd0c4c77\" (UID: \"94bb9e87-02ea-4165-bb0a-00acfd0c4c77\") " Feb 02 13:48:45 crc kubenswrapper[4846]: I0202 13:48:45.273590 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-operator-scripts\") pod \"94bb9e87-02ea-4165-bb0a-00acfd0c4c77\" (UID: \"94bb9e87-02ea-4165-bb0a-00acfd0c4c77\") " Feb 02 13:48:45 crc kubenswrapper[4846]: I0202 13:48:45.274176 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "94bb9e87-02ea-4165-bb0a-00acfd0c4c77" (UID: "94bb9e87-02ea-4165-bb0a-00acfd0c4c77"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:48:45 crc kubenswrapper[4846]: I0202 13:48:45.277922 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-kube-api-access-b6hk9" (OuterVolumeSpecName: "kube-api-access-b6hk9") pod "94bb9e87-02ea-4165-bb0a-00acfd0c4c77" (UID: "94bb9e87-02ea-4165-bb0a-00acfd0c4c77"). InnerVolumeSpecName "kube-api-access-b6hk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:48:45 crc kubenswrapper[4846]: I0202 13:48:45.375149 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:48:45 crc kubenswrapper[4846]: I0202 13:48:45.375180 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6hk9\" (UniqueName: \"kubernetes.io/projected/94bb9e87-02ea-4165-bb0a-00acfd0c4c77-kube-api-access-b6hk9\") on node \"crc\" DevicePath \"\"" Feb 02 13:48:45 crc kubenswrapper[4846]: I0202 13:48:45.854948 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-06a1-account-create-update-jj9jt" event={"ID":"94bb9e87-02ea-4165-bb0a-00acfd0c4c77","Type":"ContainerDied","Data":"e0031575e3e73dda37ab552c7dae2f0667a7aadada427eecbd2a1d6185e55d1d"} Feb 02 13:48:45 crc kubenswrapper[4846]: I0202 13:48:45.855890 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0031575e3e73dda37ab552c7dae2f0667a7aadada427eecbd2a1d6185e55d1d" Feb 02 13:48:45 crc kubenswrapper[4846]: I0202 13:48:45.854995 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-06a1-account-create-update-jj9jt" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.597300 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-nr6g7"] Feb 02 13:48:47 crc kubenswrapper[4846]: E0202 13:48:47.598054 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf" containerName="mariadb-database-create" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.598070 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf" containerName="mariadb-database-create" Feb 02 13:48:47 crc kubenswrapper[4846]: E0202 13:48:47.598106 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94bb9e87-02ea-4165-bb0a-00acfd0c4c77" containerName="mariadb-account-create-update" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.598115 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="94bb9e87-02ea-4165-bb0a-00acfd0c4c77" containerName="mariadb-account-create-update" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.598360 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="94bb9e87-02ea-4165-bb0a-00acfd0c4c77" containerName="mariadb-account-create-update" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.598386 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf" containerName="mariadb-database-create" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.599213 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-nr6g7" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.605315 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-nr6g7"] Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.616467 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgtzn\" (UniqueName: \"kubernetes.io/projected/c30f1bf7-8638-4199-a977-fbab41f1dad2-kube-api-access-mgtzn\") pod \"octavia-persistence-db-create-nr6g7\" (UID: \"c30f1bf7-8638-4199-a977-fbab41f1dad2\") " pod="openstack/octavia-persistence-db-create-nr6g7" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.616625 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c30f1bf7-8638-4199-a977-fbab41f1dad2-operator-scripts\") pod \"octavia-persistence-db-create-nr6g7\" (UID: \"c30f1bf7-8638-4199-a977-fbab41f1dad2\") " pod="openstack/octavia-persistence-db-create-nr6g7" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.717832 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c30f1bf7-8638-4199-a977-fbab41f1dad2-operator-scripts\") pod \"octavia-persistence-db-create-nr6g7\" (UID: \"c30f1bf7-8638-4199-a977-fbab41f1dad2\") " pod="openstack/octavia-persistence-db-create-nr6g7" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.717959 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgtzn\" (UniqueName: \"kubernetes.io/projected/c30f1bf7-8638-4199-a977-fbab41f1dad2-kube-api-access-mgtzn\") pod \"octavia-persistence-db-create-nr6g7\" (UID: \"c30f1bf7-8638-4199-a977-fbab41f1dad2\") " pod="openstack/octavia-persistence-db-create-nr6g7" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.719182 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c30f1bf7-8638-4199-a977-fbab41f1dad2-operator-scripts\") pod \"octavia-persistence-db-create-nr6g7\" (UID: \"c30f1bf7-8638-4199-a977-fbab41f1dad2\") " pod="openstack/octavia-persistence-db-create-nr6g7" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.745140 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgtzn\" (UniqueName: \"kubernetes.io/projected/c30f1bf7-8638-4199-a977-fbab41f1dad2-kube-api-access-mgtzn\") pod \"octavia-persistence-db-create-nr6g7\" (UID: \"c30f1bf7-8638-4199-a977-fbab41f1dad2\") " pod="openstack/octavia-persistence-db-create-nr6g7" Feb 02 13:48:47 crc kubenswrapper[4846]: I0202 13:48:47.916795 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-nr6g7" Feb 02 13:48:48 crc kubenswrapper[4846]: I0202 13:48:48.387153 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-nr6g7"] Feb 02 13:48:48 crc kubenswrapper[4846]: I0202 13:48:48.861031 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-21cb-account-create-update-p9275"] Feb 02 13:48:48 crc kubenswrapper[4846]: I0202 13:48:48.863109 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-21cb-account-create-update-p9275" Feb 02 13:48:48 crc kubenswrapper[4846]: I0202 13:48:48.868529 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Feb 02 13:48:48 crc kubenswrapper[4846]: I0202 13:48:48.872311 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-21cb-account-create-update-p9275"] Feb 02 13:48:48 crc kubenswrapper[4846]: I0202 13:48:48.900080 4846 generic.go:334] "Generic (PLEG): container finished" podID="c30f1bf7-8638-4199-a977-fbab41f1dad2" containerID="5a64a2f0dbfcc27fcbf3658c317b5037567325030e77d5bb98568fae0d3b49b1" exitCode=0 Feb 02 13:48:48 crc kubenswrapper[4846]: I0202 13:48:48.900145 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-nr6g7" event={"ID":"c30f1bf7-8638-4199-a977-fbab41f1dad2","Type":"ContainerDied","Data":"5a64a2f0dbfcc27fcbf3658c317b5037567325030e77d5bb98568fae0d3b49b1"} Feb 02 13:48:48 crc kubenswrapper[4846]: I0202 13:48:48.900182 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-nr6g7" event={"ID":"c30f1bf7-8638-4199-a977-fbab41f1dad2","Type":"ContainerStarted","Data":"0589d32546da679815673bdeac3cbedf3ef0e7cf3f64c1be94a0b55fd2317141"} Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.049965 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7a3d-account-create-update-h2wdw"] Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.060399 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-wbckr"] Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.063411 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9155b41a-6ea6-4ff8-81d7-18363415052a-operator-scripts\") pod \"octavia-21cb-account-create-update-p9275\" (UID: \"9155b41a-6ea6-4ff8-81d7-18363415052a\") " pod="openstack/octavia-21cb-account-create-update-p9275" Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.063498 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t572\" (UniqueName: \"kubernetes.io/projected/9155b41a-6ea6-4ff8-81d7-18363415052a-kube-api-access-7t572\") pod \"octavia-21cb-account-create-update-p9275\" (UID: \"9155b41a-6ea6-4ff8-81d7-18363415052a\") " pod="openstack/octavia-21cb-account-create-update-p9275" Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.070384 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-wbckr"] Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.082072 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7a3d-account-create-update-h2wdw"] Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.165237 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9155b41a-6ea6-4ff8-81d7-18363415052a-operator-scripts\") pod \"octavia-21cb-account-create-update-p9275\" (UID: \"9155b41a-6ea6-4ff8-81d7-18363415052a\") " pod="openstack/octavia-21cb-account-create-update-p9275" Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.165320 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t572\" (UniqueName: \"kubernetes.io/projected/9155b41a-6ea6-4ff8-81d7-18363415052a-kube-api-access-7t572\") pod \"octavia-21cb-account-create-update-p9275\" (UID: \"9155b41a-6ea6-4ff8-81d7-18363415052a\") " pod="openstack/octavia-21cb-account-create-update-p9275" Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.168150 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9155b41a-6ea6-4ff8-81d7-18363415052a-operator-scripts\") pod \"octavia-21cb-account-create-update-p9275\" (UID: \"9155b41a-6ea6-4ff8-81d7-18363415052a\") " pod="openstack/octavia-21cb-account-create-update-p9275" Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.193263 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t572\" (UniqueName: \"kubernetes.io/projected/9155b41a-6ea6-4ff8-81d7-18363415052a-kube-api-access-7t572\") pod \"octavia-21cb-account-create-update-p9275\" (UID: \"9155b41a-6ea6-4ff8-81d7-18363415052a\") " pod="openstack/octavia-21cb-account-create-update-p9275" Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.211611 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-21cb-account-create-update-p9275" Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.437357 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d318950f-d4db-427f-845d-b4b265d36587" path="/var/lib/kubelet/pods/d318950f-d4db-427f-845d-b4b265d36587/volumes" Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.438181 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e90b7676-c930-4301-8a8c-14d273d1112b" path="/var/lib/kubelet/pods/e90b7676-c930-4301-8a8c-14d273d1112b/volumes" Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.644825 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-21cb-account-create-update-p9275"] Feb 02 13:48:49 crc kubenswrapper[4846]: W0202 13:48:49.657774 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9155b41a_6ea6_4ff8_81d7_18363415052a.slice/crio-87b74cb20a2ec0a0f3e68a4d18da33ce86a898230752582f911dc1101cd96485 WatchSource:0}: Error finding container 87b74cb20a2ec0a0f3e68a4d18da33ce86a898230752582f911dc1101cd96485: Status 404 returned error can't find the container with id 87b74cb20a2ec0a0f3e68a4d18da33ce86a898230752582f911dc1101cd96485 Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.909578 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-21cb-account-create-update-p9275" event={"ID":"9155b41a-6ea6-4ff8-81d7-18363415052a","Type":"ContainerStarted","Data":"b261b51c3bc340880b1917f93aa56e2877153da942e93fe1a361ae5c7dfc2ff5"} Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.909903 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-21cb-account-create-update-p9275" event={"ID":"9155b41a-6ea6-4ff8-81d7-18363415052a","Type":"ContainerStarted","Data":"87b74cb20a2ec0a0f3e68a4d18da33ce86a898230752582f911dc1101cd96485"} Feb 02 13:48:49 crc kubenswrapper[4846]: I0202 13:48:49.932972 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-21cb-account-create-update-p9275" podStartSLOduration=1.932956289 podStartE2EDuration="1.932956289s" podCreationTimestamp="2026-02-02 13:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:48:49.929117046 +0000 UTC m=+5961.157703919" watchObservedRunningTime="2026-02-02 13:48:49.932956289 +0000 UTC m=+5961.161543152" Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.158321 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-nr6g7" Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.198948 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c30f1bf7-8638-4199-a977-fbab41f1dad2-operator-scripts\") pod \"c30f1bf7-8638-4199-a977-fbab41f1dad2\" (UID: \"c30f1bf7-8638-4199-a977-fbab41f1dad2\") " Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.199075 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgtzn\" (UniqueName: \"kubernetes.io/projected/c30f1bf7-8638-4199-a977-fbab41f1dad2-kube-api-access-mgtzn\") pod \"c30f1bf7-8638-4199-a977-fbab41f1dad2\" (UID: \"c30f1bf7-8638-4199-a977-fbab41f1dad2\") " Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.200403 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c30f1bf7-8638-4199-a977-fbab41f1dad2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c30f1bf7-8638-4199-a977-fbab41f1dad2" (UID: "c30f1bf7-8638-4199-a977-fbab41f1dad2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.208073 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c30f1bf7-8638-4199-a977-fbab41f1dad2-kube-api-access-mgtzn" (OuterVolumeSpecName: "kube-api-access-mgtzn") pod "c30f1bf7-8638-4199-a977-fbab41f1dad2" (UID: "c30f1bf7-8638-4199-a977-fbab41f1dad2"). InnerVolumeSpecName "kube-api-access-mgtzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.301481 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c30f1bf7-8638-4199-a977-fbab41f1dad2-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.301522 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgtzn\" (UniqueName: \"kubernetes.io/projected/c30f1bf7-8638-4199-a977-fbab41f1dad2-kube-api-access-mgtzn\") on node \"crc\" DevicePath \"\"" Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.920420 4846 generic.go:334] "Generic (PLEG): container finished" podID="9155b41a-6ea6-4ff8-81d7-18363415052a" containerID="b261b51c3bc340880b1917f93aa56e2877153da942e93fe1a361ae5c7dfc2ff5" exitCode=0 Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.920476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-21cb-account-create-update-p9275" event={"ID":"9155b41a-6ea6-4ff8-81d7-18363415052a","Type":"ContainerDied","Data":"b261b51c3bc340880b1917f93aa56e2877153da942e93fe1a361ae5c7dfc2ff5"} Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.923197 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-nr6g7" event={"ID":"c30f1bf7-8638-4199-a977-fbab41f1dad2","Type":"ContainerDied","Data":"0589d32546da679815673bdeac3cbedf3ef0e7cf3f64c1be94a0b55fd2317141"} Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.923229 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0589d32546da679815673bdeac3cbedf3ef0e7cf3f64c1be94a0b55fd2317141" Feb 02 13:48:50 crc kubenswrapper[4846]: I0202 13:48:50.923280 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-nr6g7" Feb 02 13:48:52 crc kubenswrapper[4846]: I0202 13:48:52.273279 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-21cb-account-create-update-p9275" Feb 02 13:48:52 crc kubenswrapper[4846]: I0202 13:48:52.342595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t572\" (UniqueName: \"kubernetes.io/projected/9155b41a-6ea6-4ff8-81d7-18363415052a-kube-api-access-7t572\") pod \"9155b41a-6ea6-4ff8-81d7-18363415052a\" (UID: \"9155b41a-6ea6-4ff8-81d7-18363415052a\") " Feb 02 13:48:52 crc kubenswrapper[4846]: I0202 13:48:52.342733 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9155b41a-6ea6-4ff8-81d7-18363415052a-operator-scripts\") pod \"9155b41a-6ea6-4ff8-81d7-18363415052a\" (UID: \"9155b41a-6ea6-4ff8-81d7-18363415052a\") " Feb 02 13:48:52 crc kubenswrapper[4846]: I0202 13:48:52.343773 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9155b41a-6ea6-4ff8-81d7-18363415052a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9155b41a-6ea6-4ff8-81d7-18363415052a" (UID: "9155b41a-6ea6-4ff8-81d7-18363415052a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:48:52 crc kubenswrapper[4846]: I0202 13:48:52.347288 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9155b41a-6ea6-4ff8-81d7-18363415052a-kube-api-access-7t572" (OuterVolumeSpecName: "kube-api-access-7t572") pod "9155b41a-6ea6-4ff8-81d7-18363415052a" (UID: "9155b41a-6ea6-4ff8-81d7-18363415052a"). InnerVolumeSpecName "kube-api-access-7t572". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:48:52 crc kubenswrapper[4846]: I0202 13:48:52.445208 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9155b41a-6ea6-4ff8-81d7-18363415052a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:48:52 crc kubenswrapper[4846]: I0202 13:48:52.445246 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t572\" (UniqueName: \"kubernetes.io/projected/9155b41a-6ea6-4ff8-81d7-18363415052a-kube-api-access-7t572\") on node \"crc\" DevicePath \"\"" Feb 02 13:48:52 crc kubenswrapper[4846]: I0202 13:48:52.941814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-21cb-account-create-update-p9275" event={"ID":"9155b41a-6ea6-4ff8-81d7-18363415052a","Type":"ContainerDied","Data":"87b74cb20a2ec0a0f3e68a4d18da33ce86a898230752582f911dc1101cd96485"} Feb 02 13:48:52 crc kubenswrapper[4846]: I0202 13:48:52.941856 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-21cb-account-create-update-p9275" Feb 02 13:48:52 crc kubenswrapper[4846]: I0202 13:48:52.941874 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87b74cb20a2ec0a0f3e68a4d18da33ce86a898230752582f911dc1101cd96485" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.372667 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-7b8b589ff7-l27s9"] Feb 02 13:48:54 crc kubenswrapper[4846]: E0202 13:48:54.373346 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c30f1bf7-8638-4199-a977-fbab41f1dad2" containerName="mariadb-database-create" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.373358 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c30f1bf7-8638-4199-a977-fbab41f1dad2" containerName="mariadb-database-create" Feb 02 13:48:54 crc kubenswrapper[4846]: E0202 13:48:54.373385 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9155b41a-6ea6-4ff8-81d7-18363415052a" containerName="mariadb-account-create-update" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.373391 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9155b41a-6ea6-4ff8-81d7-18363415052a" containerName="mariadb-account-create-update" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.373550 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9155b41a-6ea6-4ff8-81d7-18363415052a" containerName="mariadb-account-create-update" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.373574 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c30f1bf7-8638-4199-a977-fbab41f1dad2" containerName="mariadb-database-create" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.374959 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.383962 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.384056 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.384197 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-g2rjm" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.385217 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-ovndbs" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.405357 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-7b8b589ff7-l27s9"] Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.481326 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-ovndb-tls-certs\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.481384 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-scripts\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.481413 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.481435 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data-merged\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.481553 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-octavia-run\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.481733 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-combined-ca-bundle\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.583159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-octavia-run\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.583274 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-combined-ca-bundle\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.583306 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-ovndb-tls-certs\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.583325 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-scripts\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.583344 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.583366 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data-merged\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.583938 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data-merged\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.584388 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-octavia-run\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.590014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-ovndb-tls-certs\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.590056 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-combined-ca-bundle\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.590190 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-scripts\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.590964 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data\") pod \"octavia-api-7b8b589ff7-l27s9\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:54 crc kubenswrapper[4846]: I0202 13:48:54.701908 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:48:55 crc kubenswrapper[4846]: W0202 13:48:55.201297 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc3d5d0d_95a1_4e28_8144_2ae3729d37d8.slice/crio-2f1056bb1720084a34fb474c3af5b816a7b10f7ba377ea2c803771129932d690 WatchSource:0}: Error finding container 2f1056bb1720084a34fb474c3af5b816a7b10f7ba377ea2c803771129932d690: Status 404 returned error can't find the container with id 2f1056bb1720084a34fb474c3af5b816a7b10f7ba377ea2c803771129932d690 Feb 02 13:48:55 crc kubenswrapper[4846]: I0202 13:48:55.203281 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-7b8b589ff7-l27s9"] Feb 02 13:48:55 crc kubenswrapper[4846]: I0202 13:48:55.976232 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7b8b589ff7-l27s9" event={"ID":"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8","Type":"ContainerStarted","Data":"2f1056bb1720084a34fb474c3af5b816a7b10f7ba377ea2c803771129932d690"} Feb 02 13:48:56 crc kubenswrapper[4846]: I0202 13:48:56.053780 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-h47dj"] Feb 02 13:48:56 crc kubenswrapper[4846]: I0202 13:48:56.065249 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-h47dj"] Feb 02 13:48:57 crc kubenswrapper[4846]: I0202 13:48:57.435576 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dd78cf8-d095-4c69-849d-750d826aba63" path="/var/lib/kubelet/pods/2dd78cf8-d095-4c69-849d-750d826aba63/volumes" Feb 02 13:49:05 crc kubenswrapper[4846]: I0202 13:49:05.075605 4846 generic.go:334] "Generic (PLEG): container finished" podID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerID="2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e" exitCode=0 Feb 02 13:49:05 crc kubenswrapper[4846]: I0202 13:49:05.075815 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7b8b589ff7-l27s9" event={"ID":"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8","Type":"ContainerDied","Data":"2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e"} Feb 02 13:49:06 crc kubenswrapper[4846]: I0202 13:49:06.090633 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7b8b589ff7-l27s9" event={"ID":"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8","Type":"ContainerStarted","Data":"c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119"} Feb 02 13:49:06 crc kubenswrapper[4846]: I0202 13:49:06.091250 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7b8b589ff7-l27s9" event={"ID":"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8","Type":"ContainerStarted","Data":"2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703"} Feb 02 13:49:06 crc kubenswrapper[4846]: I0202 13:49:06.092207 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:49:06 crc kubenswrapper[4846]: I0202 13:49:06.092454 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:49:06 crc kubenswrapper[4846]: I0202 13:49:06.131485 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-7b8b589ff7-l27s9" podStartSLOduration=3.087705697 podStartE2EDuration="12.131459681s" podCreationTimestamp="2026-02-02 13:48:54 +0000 UTC" firstStartedPulling="2026-02-02 13:48:55.205013235 +0000 UTC m=+5966.433600098" lastFinishedPulling="2026-02-02 13:49:04.248767219 +0000 UTC m=+5975.477354082" observedRunningTime="2026-02-02 13:49:06.12480048 +0000 UTC m=+5977.353387343" watchObservedRunningTime="2026-02-02 13:49:06.131459681 +0000 UTC m=+5977.360046554" Feb 02 13:49:10 crc kubenswrapper[4846]: I0202 13:49:10.038825 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2qq96"] Feb 02 13:49:10 crc kubenswrapper[4846]: I0202 13:49:10.052003 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2qq96"] Feb 02 13:49:11 crc kubenswrapper[4846]: I0202 13:49:11.437371 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03f2dde-e1dd-4f93-a8b4-215163728bbd" path="/var/lib/kubelet/pods/b03f2dde-e1dd-4f93-a8b4-215163728bbd/volumes" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.322743 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-v8bjp" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.369111 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.381610 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mkdn9" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.527516 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-v8bjp-config-29kjx"] Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.528950 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.536825 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.541636 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-v8bjp-config-29kjx"] Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.622489 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-scripts\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.622539 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xz5n\" (UniqueName: \"kubernetes.io/projected/6da22a0a-e93c-475e-8bd3-8e5192649685-kube-api-access-2xz5n\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.622666 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run-ovn\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.622701 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-additional-scripts\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.622756 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-log-ovn\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.622853 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.724434 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.724773 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-scripts\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.726771 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xz5n\" (UniqueName: \"kubernetes.io/projected/6da22a0a-e93c-475e-8bd3-8e5192649685-kube-api-access-2xz5n\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.727051 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run-ovn\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.724873 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.727135 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-additional-scripts\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.726728 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-scripts\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.727343 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-log-ovn\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.727412 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-log-ovn\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.727344 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run-ovn\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.727939 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-additional-scripts\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.745384 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xz5n\" (UniqueName: \"kubernetes.io/projected/6da22a0a-e93c-475e-8bd3-8e5192649685-kube-api-access-2xz5n\") pod \"ovn-controller-v8bjp-config-29kjx\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:14 crc kubenswrapper[4846]: I0202 13:49:14.849215 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:15 crc kubenswrapper[4846]: I0202 13:49:15.361906 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-v8bjp-config-29kjx"] Feb 02 13:49:16 crc kubenswrapper[4846]: I0202 13:49:16.189174 4846 generic.go:334] "Generic (PLEG): container finished" podID="6da22a0a-e93c-475e-8bd3-8e5192649685" containerID="6a11abe8177ab0d52fe13b49d6bf7f43d078f1dc4b9c14f25da0f913939da49a" exitCode=0 Feb 02 13:49:16 crc kubenswrapper[4846]: I0202 13:49:16.191672 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v8bjp-config-29kjx" event={"ID":"6da22a0a-e93c-475e-8bd3-8e5192649685","Type":"ContainerDied","Data":"6a11abe8177ab0d52fe13b49d6bf7f43d078f1dc4b9c14f25da0f913939da49a"} Feb 02 13:49:16 crc kubenswrapper[4846]: I0202 13:49:16.191732 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v8bjp-config-29kjx" event={"ID":"6da22a0a-e93c-475e-8bd3-8e5192649685","Type":"ContainerStarted","Data":"a68eaef8b78e5a97ed118171e067036d1b1d7e15b9b0ae7f3527e08188b2e470"} Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.683646 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.787596 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run-ovn\") pod \"6da22a0a-e93c-475e-8bd3-8e5192649685\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.788067 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-additional-scripts\") pod \"6da22a0a-e93c-475e-8bd3-8e5192649685\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.788122 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-scripts\") pod \"6da22a0a-e93c-475e-8bd3-8e5192649685\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.788176 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-log-ovn\") pod \"6da22a0a-e93c-475e-8bd3-8e5192649685\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.788238 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xz5n\" (UniqueName: \"kubernetes.io/projected/6da22a0a-e93c-475e-8bd3-8e5192649685-kube-api-access-2xz5n\") pod \"6da22a0a-e93c-475e-8bd3-8e5192649685\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.787701 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6da22a0a-e93c-475e-8bd3-8e5192649685" (UID: "6da22a0a-e93c-475e-8bd3-8e5192649685"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.788387 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6da22a0a-e93c-475e-8bd3-8e5192649685" (UID: "6da22a0a-e93c-475e-8bd3-8e5192649685"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.788423 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run" (OuterVolumeSpecName: "var-run") pod "6da22a0a-e93c-475e-8bd3-8e5192649685" (UID: "6da22a0a-e93c-475e-8bd3-8e5192649685"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.788393 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run\") pod \"6da22a0a-e93c-475e-8bd3-8e5192649685\" (UID: \"6da22a0a-e93c-475e-8bd3-8e5192649685\") " Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.789004 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6da22a0a-e93c-475e-8bd3-8e5192649685" (UID: "6da22a0a-e93c-475e-8bd3-8e5192649685"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.789336 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.789354 4846 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.789364 4846 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.789374 4846 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6da22a0a-e93c-475e-8bd3-8e5192649685-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.790347 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-scripts" (OuterVolumeSpecName: "scripts") pod "6da22a0a-e93c-475e-8bd3-8e5192649685" (UID: "6da22a0a-e93c-475e-8bd3-8e5192649685"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.799753 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da22a0a-e93c-475e-8bd3-8e5192649685-kube-api-access-2xz5n" (OuterVolumeSpecName: "kube-api-access-2xz5n") pod "6da22a0a-e93c-475e-8bd3-8e5192649685" (UID: "6da22a0a-e93c-475e-8bd3-8e5192649685"). InnerVolumeSpecName "kube-api-access-2xz5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.890889 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6da22a0a-e93c-475e-8bd3-8e5192649685-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:17 crc kubenswrapper[4846]: I0202 13:49:17.891110 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xz5n\" (UniqueName: \"kubernetes.io/projected/6da22a0a-e93c-475e-8bd3-8e5192649685-kube-api-access-2xz5n\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.210044 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v8bjp-config-29kjx" event={"ID":"6da22a0a-e93c-475e-8bd3-8e5192649685","Type":"ContainerDied","Data":"a68eaef8b78e5a97ed118171e067036d1b1d7e15b9b0ae7f3527e08188b2e470"} Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.210088 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a68eaef8b78e5a97ed118171e067036d1b1d7e15b9b0ae7f3527e08188b2e470" Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.210091 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v8bjp-config-29kjx" Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.769666 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-v8bjp-config-29kjx"] Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.780214 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-v8bjp-config-29kjx"] Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.904247 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-v8bjp-config-phmq6"] Feb 02 13:49:18 crc kubenswrapper[4846]: E0202 13:49:18.904673 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da22a0a-e93c-475e-8bd3-8e5192649685" containerName="ovn-config" Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.904689 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da22a0a-e93c-475e-8bd3-8e5192649685" containerName="ovn-config" Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.904909 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da22a0a-e93c-475e-8bd3-8e5192649685" containerName="ovn-config" Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.905510 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.907642 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 02 13:49:18 crc kubenswrapper[4846]: I0202 13:49:18.916673 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-v8bjp-config-phmq6"] Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.010734 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.010779 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-scripts\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.010995 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfmr6\" (UniqueName: \"kubernetes.io/projected/406185e2-2978-42ac-b561-05049828c0c6-kube-api-access-xfmr6\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.011252 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-log-ovn\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.011369 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run-ovn\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.011408 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-additional-scripts\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.113978 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfmr6\" (UniqueName: \"kubernetes.io/projected/406185e2-2978-42ac-b561-05049828c0c6-kube-api-access-xfmr6\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.114207 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-log-ovn\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.114342 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run-ovn\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.114386 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-additional-scripts\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.114553 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.114604 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-scripts\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.114754 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run-ovn\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.114816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-log-ovn\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.114842 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.115325 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-additional-scripts\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.116562 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-scripts\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.136489 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfmr6\" (UniqueName: \"kubernetes.io/projected/406185e2-2978-42ac-b561-05049828c0c6-kube-api-access-xfmr6\") pod \"ovn-controller-v8bjp-config-phmq6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.231314 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.290245 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-rddhp"] Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.292768 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.299093 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.299161 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.299364 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.306718 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-rddhp"] Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.321795 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88517515-ad14-4ea6-93f9-f1bde00e726a-config-data\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.322278 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88517515-ad14-4ea6-93f9-f1bde00e726a-scripts\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.322302 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/88517515-ad14-4ea6-93f9-f1bde00e726a-hm-ports\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.322435 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/88517515-ad14-4ea6-93f9-f1bde00e726a-config-data-merged\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.424315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88517515-ad14-4ea6-93f9-f1bde00e726a-config-data\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.424353 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88517515-ad14-4ea6-93f9-f1bde00e726a-scripts\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.424369 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/88517515-ad14-4ea6-93f9-f1bde00e726a-hm-ports\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.424442 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/88517515-ad14-4ea6-93f9-f1bde00e726a-config-data-merged\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.424939 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/88517515-ad14-4ea6-93f9-f1bde00e726a-config-data-merged\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.427139 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/88517515-ad14-4ea6-93f9-f1bde00e726a-hm-ports\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.434027 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88517515-ad14-4ea6-93f9-f1bde00e726a-config-data\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.438785 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88517515-ad14-4ea6-93f9-f1bde00e726a-scripts\") pod \"octavia-rsyslog-rddhp\" (UID: \"88517515-ad14-4ea6-93f9-f1bde00e726a\") " pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.451601 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da22a0a-e93c-475e-8bd3-8e5192649685" path="/var/lib/kubelet/pods/6da22a0a-e93c-475e-8bd3-8e5192649685/volumes" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.669170 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:19 crc kubenswrapper[4846]: I0202 13:49:19.811597 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-v8bjp-config-phmq6"] Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.167382 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-65dd99cb46-tp6sx"] Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.169562 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.181840 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.185033 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-65dd99cb46-tp6sx"] Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.248071 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v8bjp-config-phmq6" event={"ID":"406185e2-2978-42ac-b561-05049828c0c6","Type":"ContainerStarted","Data":"9f24403eb20e3eebeee4589911d46faa77387cd08614eea20bad2428cbcde459"} Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.248943 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ff73f76-6b20-46a7-9919-ad9ccd05641b-httpd-config\") pod \"octavia-image-upload-65dd99cb46-tp6sx\" (UID: \"5ff73f76-6b20-46a7-9919-ad9ccd05641b\") " pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.249134 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/5ff73f76-6b20-46a7-9919-ad9ccd05641b-amphora-image\") pod \"octavia-image-upload-65dd99cb46-tp6sx\" (UID: \"5ff73f76-6b20-46a7-9919-ad9ccd05641b\") " pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.318011 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-rddhp"] Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.351669 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/5ff73f76-6b20-46a7-9919-ad9ccd05641b-amphora-image\") pod \"octavia-image-upload-65dd99cb46-tp6sx\" (UID: \"5ff73f76-6b20-46a7-9919-ad9ccd05641b\") " pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.351780 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ff73f76-6b20-46a7-9919-ad9ccd05641b-httpd-config\") pod \"octavia-image-upload-65dd99cb46-tp6sx\" (UID: \"5ff73f76-6b20-46a7-9919-ad9ccd05641b\") " pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.352340 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/5ff73f76-6b20-46a7-9919-ad9ccd05641b-amphora-image\") pod \"octavia-image-upload-65dd99cb46-tp6sx\" (UID: \"5ff73f76-6b20-46a7-9919-ad9ccd05641b\") " pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.367594 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ff73f76-6b20-46a7-9919-ad9ccd05641b-httpd-config\") pod \"octavia-image-upload-65dd99cb46-tp6sx\" (UID: \"5ff73f76-6b20-46a7-9919-ad9ccd05641b\") " pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.368657 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-rddhp"] Feb 02 13:49:20 crc kubenswrapper[4846]: I0202 13:49:20.542144 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.027302 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-65dd99cb46-tp6sx"] Feb 02 13:49:21 crc kubenswrapper[4846]: W0202 13:49:21.074815 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ff73f76_6b20_46a7_9919_ad9ccd05641b.slice/crio-b30dd04f2381f9e2228c69b46d28d0e344f689659ad5921e60008ea4ed787311 WatchSource:0}: Error finding container b30dd04f2381f9e2228c69b46d28d0e344f689659ad5921e60008ea4ed787311: Status 404 returned error can't find the container with id b30dd04f2381f9e2228c69b46d28d0e344f689659ad5921e60008ea4ed787311 Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.261903 4846 generic.go:334] "Generic (PLEG): container finished" podID="406185e2-2978-42ac-b561-05049828c0c6" containerID="16aac1532f4b216f5978f276eb1f5873e8a7988440c11f635eb0a7dd25e5988c" exitCode=0 Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.261964 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v8bjp-config-phmq6" event={"ID":"406185e2-2978-42ac-b561-05049828c0c6","Type":"ContainerDied","Data":"16aac1532f4b216f5978f276eb1f5873e8a7988440c11f635eb0a7dd25e5988c"} Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.263011 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" event={"ID":"5ff73f76-6b20-46a7-9919-ad9ccd05641b","Type":"ContainerStarted","Data":"b30dd04f2381f9e2228c69b46d28d0e344f689659ad5921e60008ea4ed787311"} Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.267727 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-rddhp" event={"ID":"88517515-ad14-4ea6-93f9-f1bde00e726a","Type":"ContainerStarted","Data":"ca93334d8ccc0ca1c1eaf088d012c0549a263089ee24af1152a0cba2b41b3566"} Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.442182 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-ml2mq"] Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.444176 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-ml2mq"] Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.444366 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.448125 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.583849 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-config-data\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.583912 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-scripts\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.584172 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/27f64e70-e2a4-4d28-85fa-32421976a607-config-data-merged\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.584236 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-combined-ca-bundle\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.686738 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-config-data\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.686783 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-scripts\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.686883 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/27f64e70-e2a4-4d28-85fa-32421976a607-config-data-merged\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.686926 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-combined-ca-bundle\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.687608 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/27f64e70-e2a4-4d28-85fa-32421976a607-config-data-merged\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.693097 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-config-data\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.694901 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-scripts\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.713328 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-combined-ca-bundle\") pod \"octavia-db-sync-ml2mq\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:21 crc kubenswrapper[4846]: I0202 13:49:21.780741 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.115031 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-ml2mq"] Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.245203 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.307190 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ml2mq" event={"ID":"27f64e70-e2a4-4d28-85fa-32421976a607","Type":"ContainerStarted","Data":"5fb4c9f80081228a83d628f64a9d6aa581f74b29a60630666ac7342d2b94b59a"} Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.310276 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v8bjp-config-phmq6" event={"ID":"406185e2-2978-42ac-b561-05049828c0c6","Type":"ContainerDied","Data":"9f24403eb20e3eebeee4589911d46faa77387cd08614eea20bad2428cbcde459"} Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.310313 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f24403eb20e3eebeee4589911d46faa77387cd08614eea20bad2428cbcde459" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.310375 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v8bjp-config-phmq6" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.332564 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-scripts\") pod \"406185e2-2978-42ac-b561-05049828c0c6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.332898 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-additional-scripts\") pod \"406185e2-2978-42ac-b561-05049828c0c6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.333241 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfmr6\" (UniqueName: \"kubernetes.io/projected/406185e2-2978-42ac-b561-05049828c0c6-kube-api-access-xfmr6\") pod \"406185e2-2978-42ac-b561-05049828c0c6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.333444 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-log-ovn\") pod \"406185e2-2978-42ac-b561-05049828c0c6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.333539 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "406185e2-2978-42ac-b561-05049828c0c6" (UID: "406185e2-2978-42ac-b561-05049828c0c6"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.333549 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "406185e2-2978-42ac-b561-05049828c0c6" (UID: "406185e2-2978-42ac-b561-05049828c0c6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.333559 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run-ovn\") pod \"406185e2-2978-42ac-b561-05049828c0c6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.333686 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run\") pod \"406185e2-2978-42ac-b561-05049828c0c6\" (UID: \"406185e2-2978-42ac-b561-05049828c0c6\") " Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.333705 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "406185e2-2978-42ac-b561-05049828c0c6" (UID: "406185e2-2978-42ac-b561-05049828c0c6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.333817 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-scripts" (OuterVolumeSpecName: "scripts") pod "406185e2-2978-42ac-b561-05049828c0c6" (UID: "406185e2-2978-42ac-b561-05049828c0c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.333852 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run" (OuterVolumeSpecName: "var-run") pod "406185e2-2978-42ac-b561-05049828c0c6" (UID: "406185e2-2978-42ac-b561-05049828c0c6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.334969 4846 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.335037 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-run\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.335050 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.335063 4846 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/406185e2-2978-42ac-b561-05049828c0c6-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.335075 4846 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/406185e2-2978-42ac-b561-05049828c0c6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.350755 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/406185e2-2978-42ac-b561-05049828c0c6-kube-api-access-xfmr6" (OuterVolumeSpecName: "kube-api-access-xfmr6") pod "406185e2-2978-42ac-b561-05049828c0c6" (UID: "406185e2-2978-42ac-b561-05049828c0c6"). InnerVolumeSpecName "kube-api-access-xfmr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:49:23 crc kubenswrapper[4846]: I0202 13:49:23.437421 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfmr6\" (UniqueName: \"kubernetes.io/projected/406185e2-2978-42ac-b561-05049828c0c6-kube-api-access-xfmr6\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:24 crc kubenswrapper[4846]: I0202 13:49:24.348312 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ml2mq" event={"ID":"27f64e70-e2a4-4d28-85fa-32421976a607","Type":"ContainerStarted","Data":"80eb979d842bcaa49dbe868238d121d7d74f6f1ecc1a5f2a2ca20f0ef0009967"} Feb 02 13:49:24 crc kubenswrapper[4846]: I0202 13:49:24.364779 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-v8bjp-config-phmq6"] Feb 02 13:49:24 crc kubenswrapper[4846]: I0202 13:49:24.381420 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-v8bjp-config-phmq6"] Feb 02 13:49:25 crc kubenswrapper[4846]: I0202 13:49:25.361831 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-rddhp" event={"ID":"88517515-ad14-4ea6-93f9-f1bde00e726a","Type":"ContainerStarted","Data":"3a6047fd105a6920271e5fe0580e9eb4729d934e8280e01e0f0c833d02ff3a85"} Feb 02 13:49:25 crc kubenswrapper[4846]: I0202 13:49:25.369804 4846 generic.go:334] "Generic (PLEG): container finished" podID="27f64e70-e2a4-4d28-85fa-32421976a607" containerID="80eb979d842bcaa49dbe868238d121d7d74f6f1ecc1a5f2a2ca20f0ef0009967" exitCode=0 Feb 02 13:49:25 crc kubenswrapper[4846]: I0202 13:49:25.369852 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ml2mq" event={"ID":"27f64e70-e2a4-4d28-85fa-32421976a607","Type":"ContainerDied","Data":"80eb979d842bcaa49dbe868238d121d7d74f6f1ecc1a5f2a2ca20f0ef0009967"} Feb 02 13:49:25 crc kubenswrapper[4846]: I0202 13:49:25.434717 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="406185e2-2978-42ac-b561-05049828c0c6" path="/var/lib/kubelet/pods/406185e2-2978-42ac-b561-05049828c0c6/volumes" Feb 02 13:49:27 crc kubenswrapper[4846]: I0202 13:49:27.399163 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ml2mq" event={"ID":"27f64e70-e2a4-4d28-85fa-32421976a607","Type":"ContainerStarted","Data":"f32c406473cbf1428acf134df8ba011325e01b9f9ce31974b90488b8d9e66468"} Feb 02 13:49:27 crc kubenswrapper[4846]: I0202 13:49:27.411153 4846 generic.go:334] "Generic (PLEG): container finished" podID="88517515-ad14-4ea6-93f9-f1bde00e726a" containerID="3a6047fd105a6920271e5fe0580e9eb4729d934e8280e01e0f0c833d02ff3a85" exitCode=0 Feb 02 13:49:27 crc kubenswrapper[4846]: I0202 13:49:27.411206 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-rddhp" event={"ID":"88517515-ad14-4ea6-93f9-f1bde00e726a","Type":"ContainerDied","Data":"3a6047fd105a6920271e5fe0580e9eb4729d934e8280e01e0f0c833d02ff3a85"} Feb 02 13:49:27 crc kubenswrapper[4846]: I0202 13:49:27.433034 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-ml2mq" podStartSLOduration=6.433010114 podStartE2EDuration="6.433010114s" podCreationTimestamp="2026-02-02 13:49:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:49:27.421145667 +0000 UTC m=+5998.649732540" watchObservedRunningTime="2026-02-02 13:49:27.433010114 +0000 UTC m=+5998.661596977" Feb 02 13:49:29 crc kubenswrapper[4846]: I0202 13:49:29.322399 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:49:29 crc kubenswrapper[4846]: I0202 13:49:29.685195 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:49:30 crc kubenswrapper[4846]: I0202 13:49:30.456377 4846 generic.go:334] "Generic (PLEG): container finished" podID="27f64e70-e2a4-4d28-85fa-32421976a607" containerID="f32c406473cbf1428acf134df8ba011325e01b9f9ce31974b90488b8d9e66468" exitCode=0 Feb 02 13:49:30 crc kubenswrapper[4846]: I0202 13:49:30.456714 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ml2mq" event={"ID":"27f64e70-e2a4-4d28-85fa-32421976a607","Type":"ContainerDied","Data":"f32c406473cbf1428acf134df8ba011325e01b9f9ce31974b90488b8d9e66468"} Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.280978 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.371722 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-scripts\") pod \"27f64e70-e2a4-4d28-85fa-32421976a607\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.371786 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/27f64e70-e2a4-4d28-85fa-32421976a607-config-data-merged\") pod \"27f64e70-e2a4-4d28-85fa-32421976a607\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.371942 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-config-data\") pod \"27f64e70-e2a4-4d28-85fa-32421976a607\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.372704 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-combined-ca-bundle\") pod \"27f64e70-e2a4-4d28-85fa-32421976a607\" (UID: \"27f64e70-e2a4-4d28-85fa-32421976a607\") " Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.398350 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27f64e70-e2a4-4d28-85fa-32421976a607-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "27f64e70-e2a4-4d28-85fa-32421976a607" (UID: "27f64e70-e2a4-4d28-85fa-32421976a607"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.432454 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-scripts" (OuterVolumeSpecName: "scripts") pod "27f64e70-e2a4-4d28-85fa-32421976a607" (UID: "27f64e70-e2a4-4d28-85fa-32421976a607"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.432507 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-config-data" (OuterVolumeSpecName: "config-data") pod "27f64e70-e2a4-4d28-85fa-32421976a607" (UID: "27f64e70-e2a4-4d28-85fa-32421976a607"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.432594 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27f64e70-e2a4-4d28-85fa-32421976a607" (UID: "27f64e70-e2a4-4d28-85fa-32421976a607"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.475586 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.475943 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/27f64e70-e2a4-4d28-85fa-32421976a607-config-data-merged\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.475959 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.475973 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27f64e70-e2a4-4d28-85fa-32421976a607-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.514008 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-ml2mq" event={"ID":"27f64e70-e2a4-4d28-85fa-32421976a607","Type":"ContainerDied","Data":"5fb4c9f80081228a83d628f64a9d6aa581f74b29a60630666ac7342d2b94b59a"} Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.514038 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fb4c9f80081228a83d628f64a9d6aa581f74b29a60630666ac7342d2b94b59a" Feb 02 13:49:34 crc kubenswrapper[4846]: I0202 13:49:34.514089 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-ml2mq" Feb 02 13:49:35 crc kubenswrapper[4846]: I0202 13:49:35.524457 4846 generic.go:334] "Generic (PLEG): container finished" podID="5ff73f76-6b20-46a7-9919-ad9ccd05641b" containerID="b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d" exitCode=0 Feb 02 13:49:35 crc kubenswrapper[4846]: I0202 13:49:35.524644 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" event={"ID":"5ff73f76-6b20-46a7-9919-ad9ccd05641b","Type":"ContainerDied","Data":"b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d"} Feb 02 13:49:35 crc kubenswrapper[4846]: I0202 13:49:35.527166 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-rddhp" event={"ID":"88517515-ad14-4ea6-93f9-f1bde00e726a","Type":"ContainerStarted","Data":"546c5654a782649c9ec601f22bd9f4a6d54f19c02db591b57a7a9fa2b12cabcb"} Feb 02 13:49:35 crc kubenswrapper[4846]: I0202 13:49:35.528957 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:35 crc kubenswrapper[4846]: I0202 13:49:35.592710 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-rddhp" podStartSLOduration=2.290196462 podStartE2EDuration="16.592688478s" podCreationTimestamp="2026-02-02 13:49:19 +0000 UTC" firstStartedPulling="2026-02-02 13:49:20.303367556 +0000 UTC m=+5991.531954419" lastFinishedPulling="2026-02-02 13:49:34.605859572 +0000 UTC m=+6005.834446435" observedRunningTime="2026-02-02 13:49:35.578472493 +0000 UTC m=+6006.807059356" watchObservedRunningTime="2026-02-02 13:49:35.592688478 +0000 UTC m=+6006.821275341" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.329606 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-db4457bbc-xgwkd"] Feb 02 13:49:36 crc kubenswrapper[4846]: E0202 13:49:36.330465 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="406185e2-2978-42ac-b561-05049828c0c6" containerName="ovn-config" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.330487 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="406185e2-2978-42ac-b561-05049828c0c6" containerName="ovn-config" Feb 02 13:49:36 crc kubenswrapper[4846]: E0202 13:49:36.330510 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f64e70-e2a4-4d28-85fa-32421976a607" containerName="octavia-db-sync" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.330529 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f64e70-e2a4-4d28-85fa-32421976a607" containerName="octavia-db-sync" Feb 02 13:49:36 crc kubenswrapper[4846]: E0202 13:49:36.330557 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f64e70-e2a4-4d28-85fa-32421976a607" containerName="init" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.330568 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f64e70-e2a4-4d28-85fa-32421976a607" containerName="init" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.330873 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="406185e2-2978-42ac-b561-05049828c0c6" containerName="ovn-config" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.330909 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="27f64e70-e2a4-4d28-85fa-32421976a607" containerName="octavia-db-sync" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.332480 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.334791 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-internal-svc" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.335167 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-public-svc" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.344950 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-db4457bbc-xgwkd"] Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.408976 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-config-data\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.409080 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-public-tls-certs\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.409140 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/543b5e7c-8a34-45bc-a19c-75f27a036804-config-data-merged\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.409172 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/543b5e7c-8a34-45bc-a19c-75f27a036804-octavia-run\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.409232 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-combined-ca-bundle\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.409300 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-ovndb-tls-certs\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.409324 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-internal-tls-certs\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.409376 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-scripts\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.510943 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-internal-tls-certs\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.511002 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-scripts\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.511069 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-config-data\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.511132 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-public-tls-certs\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.511163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/543b5e7c-8a34-45bc-a19c-75f27a036804-config-data-merged\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.511193 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/543b5e7c-8a34-45bc-a19c-75f27a036804-octavia-run\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.511240 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-combined-ca-bundle\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.511305 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-ovndb-tls-certs\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.511903 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/543b5e7c-8a34-45bc-a19c-75f27a036804-octavia-run\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.511932 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/543b5e7c-8a34-45bc-a19c-75f27a036804-config-data-merged\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.516556 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-public-tls-certs\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.517375 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-internal-tls-certs\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.517666 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-combined-ca-bundle\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.520806 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-config-data\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.520820 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-ovndb-tls-certs\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.529396 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/543b5e7c-8a34-45bc-a19c-75f27a036804-scripts\") pod \"octavia-api-db4457bbc-xgwkd\" (UID: \"543b5e7c-8a34-45bc-a19c-75f27a036804\") " pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.537902 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" event={"ID":"5ff73f76-6b20-46a7-9919-ad9ccd05641b","Type":"ContainerStarted","Data":"74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd"} Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.570175 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" podStartSLOduration=2.79429965 podStartE2EDuration="16.570149716s" podCreationTimestamp="2026-02-02 13:49:20 +0000 UTC" firstStartedPulling="2026-02-02 13:49:21.09289502 +0000 UTC m=+5992.321481883" lastFinishedPulling="2026-02-02 13:49:34.868745086 +0000 UTC m=+6006.097331949" observedRunningTime="2026-02-02 13:49:36.550776865 +0000 UTC m=+6007.779363728" watchObservedRunningTime="2026-02-02 13:49:36.570149716 +0000 UTC m=+6007.798736579" Feb 02 13:49:36 crc kubenswrapper[4846]: I0202 13:49:36.695708 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:37 crc kubenswrapper[4846]: W0202 13:49:37.236751 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod543b5e7c_8a34_45bc_a19c_75f27a036804.slice/crio-5ce77659114cecaa662b07e86936334e5591b74b29c4fc24ac563f835d7cc95a WatchSource:0}: Error finding container 5ce77659114cecaa662b07e86936334e5591b74b29c4fc24ac563f835d7cc95a: Status 404 returned error can't find the container with id 5ce77659114cecaa662b07e86936334e5591b74b29c4fc24ac563f835d7cc95a Feb 02 13:49:37 crc kubenswrapper[4846]: I0202 13:49:37.244590 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-db4457bbc-xgwkd"] Feb 02 13:49:37 crc kubenswrapper[4846]: I0202 13:49:37.550761 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-db4457bbc-xgwkd" event={"ID":"543b5e7c-8a34-45bc-a19c-75f27a036804","Type":"ContainerStarted","Data":"5be91956ca6d888fbe547bacc1df591b92e253446b93716d4f84b02cb665e21b"} Feb 02 13:49:37 crc kubenswrapper[4846]: I0202 13:49:37.551106 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-db4457bbc-xgwkd" event={"ID":"543b5e7c-8a34-45bc-a19c-75f27a036804","Type":"ContainerStarted","Data":"5ce77659114cecaa662b07e86936334e5591b74b29c4fc24ac563f835d7cc95a"} Feb 02 13:49:38 crc kubenswrapper[4846]: I0202 13:49:38.560899 4846 generic.go:334] "Generic (PLEG): container finished" podID="543b5e7c-8a34-45bc-a19c-75f27a036804" containerID="5be91956ca6d888fbe547bacc1df591b92e253446b93716d4f84b02cb665e21b" exitCode=0 Feb 02 13:49:38 crc kubenswrapper[4846]: I0202 13:49:38.560990 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-db4457bbc-xgwkd" event={"ID":"543b5e7c-8a34-45bc-a19c-75f27a036804","Type":"ContainerDied","Data":"5be91956ca6d888fbe547bacc1df591b92e253446b93716d4f84b02cb665e21b"} Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.235791 4846 scope.go:117] "RemoveContainer" containerID="05c09344383f9c8e5d51704d8fa580cf13550a53818d06fb1d07da6df906a258" Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.270883 4846 scope.go:117] "RemoveContainer" containerID="1be1f1eb7f13040ffddb93fdb888ff9b387c46be744cd5ba0613e4acefae0187" Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.306480 4846 scope.go:117] "RemoveContainer" containerID="674e23d0e99a48bd2b58cef4a2d38ab23dbf7ff2689d81a5938d2e82f97afa98" Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.339703 4846 scope.go:117] "RemoveContainer" containerID="9e08b753754ca20b9dd23cd131c22e0726555c25a46b244d99d1a89c470c94b3" Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.384548 4846 scope.go:117] "RemoveContainer" containerID="6d5280498824192bfb8f8dddfa98520726b5ea653d620ec0ff92e2a970352078" Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.454933 4846 scope.go:117] "RemoveContainer" containerID="23cf60ad08070e0a533b416658baa04fbb571b1b7ce7abd32d62d60c6ee62746" Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.581289 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-db4457bbc-xgwkd" event={"ID":"543b5e7c-8a34-45bc-a19c-75f27a036804","Type":"ContainerStarted","Data":"6411ca119add2154643313c21ff635192a6c2b686c24d6046c6bd536a47e5416"} Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.581737 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-db4457bbc-xgwkd" event={"ID":"543b5e7c-8a34-45bc-a19c-75f27a036804","Type":"ContainerStarted","Data":"5dea172c0bedc22dccefb717de8cf4d1f5ee42d2f536841c6231f628971c15cc"} Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.581830 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.581869 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:39 crc kubenswrapper[4846]: I0202 13:49:39.621392 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-db4457bbc-xgwkd" podStartSLOduration=3.621369087 podStartE2EDuration="3.621369087s" podCreationTimestamp="2026-02-02 13:49:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:49:39.609285044 +0000 UTC m=+6010.837871907" watchObservedRunningTime="2026-02-02 13:49:39.621369087 +0000 UTC m=+6010.849955950" Feb 02 13:49:49 crc kubenswrapper[4846]: I0202 13:49:49.704807 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-rddhp" Feb 02 13:49:56 crc kubenswrapper[4846]: I0202 13:49:56.242559 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:56 crc kubenswrapper[4846]: I0202 13:49:56.539892 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-db4457bbc-xgwkd" Feb 02 13:49:56 crc kubenswrapper[4846]: I0202 13:49:56.613954 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-api-7b8b589ff7-l27s9"] Feb 02 13:49:56 crc kubenswrapper[4846]: I0202 13:49:56.614238 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-api-7b8b589ff7-l27s9" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerName="octavia-api" containerID="cri-o://2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703" gracePeriod=30 Feb 02 13:49:56 crc kubenswrapper[4846]: I0202 13:49:56.617263 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-api-7b8b589ff7-l27s9" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerName="octavia-api-provider-agent" containerID="cri-o://c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119" gracePeriod=30 Feb 02 13:49:57 crc kubenswrapper[4846]: I0202 13:49:57.747495 4846 generic.go:334] "Generic (PLEG): container finished" podID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerID="c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119" exitCode=0 Feb 02 13:49:57 crc kubenswrapper[4846]: I0202 13:49:57.747684 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7b8b589ff7-l27s9" event={"ID":"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8","Type":"ContainerDied","Data":"c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119"} Feb 02 13:49:59 crc kubenswrapper[4846]: I0202 13:49:59.667561 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-65dd99cb46-tp6sx"] Feb 02 13:49:59 crc kubenswrapper[4846]: I0202 13:49:59.668094 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" podUID="5ff73f76-6b20-46a7-9919-ad9ccd05641b" containerName="octavia-amphora-httpd" containerID="cri-o://74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd" gracePeriod=30 Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.232861 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.322865 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ff73f76-6b20-46a7-9919-ad9ccd05641b-httpd-config\") pod \"5ff73f76-6b20-46a7-9919-ad9ccd05641b\" (UID: \"5ff73f76-6b20-46a7-9919-ad9ccd05641b\") " Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.323396 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/5ff73f76-6b20-46a7-9919-ad9ccd05641b-amphora-image\") pod \"5ff73f76-6b20-46a7-9919-ad9ccd05641b\" (UID: \"5ff73f76-6b20-46a7-9919-ad9ccd05641b\") " Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.329854 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.355419 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ff73f76-6b20-46a7-9919-ad9ccd05641b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5ff73f76-6b20-46a7-9919-ad9ccd05641b" (UID: "5ff73f76-6b20-46a7-9919-ad9ccd05641b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.396557 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ff73f76-6b20-46a7-9919-ad9ccd05641b-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "5ff73f76-6b20-46a7-9919-ad9ccd05641b" (UID: "5ff73f76-6b20-46a7-9919-ad9ccd05641b"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.425370 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-ovndb-tls-certs\") pod \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.425434 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-combined-ca-bundle\") pod \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.425616 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data\") pod \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.425655 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data-merged\") pod \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.425704 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-octavia-run\") pod \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.425729 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-scripts\") pod \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\" (UID: \"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8\") " Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.426102 4846 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/5ff73f76-6b20-46a7-9919-ad9ccd05641b-amphora-image\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.426122 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ff73f76-6b20-46a7-9919-ad9ccd05641b-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.426964 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-octavia-run" (OuterVolumeSpecName: "octavia-run") pod "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" (UID: "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8"). InnerVolumeSpecName "octavia-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.428937 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-scripts" (OuterVolumeSpecName: "scripts") pod "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" (UID: "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.431599 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data" (OuterVolumeSpecName: "config-data") pod "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" (UID: "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.469726 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" (UID: "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.472697 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" (UID: "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.527771 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.527809 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-config-data-merged\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.527820 4846 reconciler_common.go:293] "Volume detached for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-octavia-run\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.527828 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.527836 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.574933 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" (UID: "dc3d5d0d-95a1-4e28-8144-2ae3729d37d8"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.629811 4846 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.775858 4846 generic.go:334] "Generic (PLEG): container finished" podID="5ff73f76-6b20-46a7-9919-ad9ccd05641b" containerID="74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd" exitCode=0 Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.775922 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.775930 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" event={"ID":"5ff73f76-6b20-46a7-9919-ad9ccd05641b","Type":"ContainerDied","Data":"74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd"} Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.775960 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-65dd99cb46-tp6sx" event={"ID":"5ff73f76-6b20-46a7-9919-ad9ccd05641b","Type":"ContainerDied","Data":"b30dd04f2381f9e2228c69b46d28d0e344f689659ad5921e60008ea4ed787311"} Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.775990 4846 scope.go:117] "RemoveContainer" containerID="74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.787948 4846 generic.go:334] "Generic (PLEG): container finished" podID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerID="2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703" exitCode=0 Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.788075 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7b8b589ff7-l27s9" event={"ID":"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8","Type":"ContainerDied","Data":"2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703"} Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.788307 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7b8b589ff7-l27s9" event={"ID":"dc3d5d0d-95a1-4e28-8144-2ae3729d37d8","Type":"ContainerDied","Data":"2f1056bb1720084a34fb474c3af5b816a7b10f7ba377ea2c803771129932d690"} Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.788952 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-7b8b589ff7-l27s9" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.812839 4846 scope.go:117] "RemoveContainer" containerID="b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.819459 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-65dd99cb46-tp6sx"] Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.828936 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-65dd99cb46-tp6sx"] Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.889933 4846 scope.go:117] "RemoveContainer" containerID="74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd" Feb 02 13:50:00 crc kubenswrapper[4846]: E0202 13:50:00.890297 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd\": container with ID starting with 74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd not found: ID does not exist" containerID="74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.890326 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd"} err="failed to get container status \"74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd\": rpc error: code = NotFound desc = could not find container \"74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd\": container with ID starting with 74d8a741a00b9cb422718f4896d68fba6d9a5b4a5f80bc2f9586906e358a5bdd not found: ID does not exist" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.890345 4846 scope.go:117] "RemoveContainer" containerID="b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d" Feb 02 13:50:00 crc kubenswrapper[4846]: E0202 13:50:00.890562 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d\": container with ID starting with b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d not found: ID does not exist" containerID="b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.890579 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d"} err="failed to get container status \"b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d\": rpc error: code = NotFound desc = could not find container \"b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d\": container with ID starting with b2ad603ab1c600ace73cd6421a4076048249e10e66edddf767366d0beba2a87d not found: ID does not exist" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.890591 4846 scope.go:117] "RemoveContainer" containerID="c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.949712 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-api-7b8b589ff7-l27s9"] Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.953466 4846 scope.go:117] "RemoveContainer" containerID="2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703" Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.958772 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-api-7b8b589ff7-l27s9"] Feb 02 13:50:00 crc kubenswrapper[4846]: I0202 13:50:00.987994 4846 scope.go:117] "RemoveContainer" containerID="2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e" Feb 02 13:50:01 crc kubenswrapper[4846]: I0202 13:50:01.007237 4846 scope.go:117] "RemoveContainer" containerID="c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119" Feb 02 13:50:01 crc kubenswrapper[4846]: E0202 13:50:01.007955 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119\": container with ID starting with c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119 not found: ID does not exist" containerID="c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119" Feb 02 13:50:01 crc kubenswrapper[4846]: I0202 13:50:01.008090 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119"} err="failed to get container status \"c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119\": rpc error: code = NotFound desc = could not find container \"c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119\": container with ID starting with c929685558c980412368216851df9e7872dae354b69c960f0e834f6c4d1a8119 not found: ID does not exist" Feb 02 13:50:01 crc kubenswrapper[4846]: I0202 13:50:01.008208 4846 scope.go:117] "RemoveContainer" containerID="2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703" Feb 02 13:50:01 crc kubenswrapper[4846]: E0202 13:50:01.008776 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703\": container with ID starting with 2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703 not found: ID does not exist" containerID="2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703" Feb 02 13:50:01 crc kubenswrapper[4846]: I0202 13:50:01.008822 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703"} err="failed to get container status \"2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703\": rpc error: code = NotFound desc = could not find container \"2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703\": container with ID starting with 2a0ab0bc23515384f38995fc090c6265224feb989ced9c48122fdefafa0bc703 not found: ID does not exist" Feb 02 13:50:01 crc kubenswrapper[4846]: I0202 13:50:01.008849 4846 scope.go:117] "RemoveContainer" containerID="2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e" Feb 02 13:50:01 crc kubenswrapper[4846]: E0202 13:50:01.009123 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e\": container with ID starting with 2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e not found: ID does not exist" containerID="2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e" Feb 02 13:50:01 crc kubenswrapper[4846]: I0202 13:50:01.009160 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e"} err="failed to get container status \"2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e\": rpc error: code = NotFound desc = could not find container \"2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e\": container with ID starting with 2e1ca8a7b5f8ecd2b596f3f18560691f6924c720ba184db811836682337ccc3e not found: ID does not exist" Feb 02 13:50:01 crc kubenswrapper[4846]: I0202 13:50:01.436908 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ff73f76-6b20-46a7-9919-ad9ccd05641b" path="/var/lib/kubelet/pods/5ff73f76-6b20-46a7-9919-ad9ccd05641b/volumes" Feb 02 13:50:01 crc kubenswrapper[4846]: I0202 13:50:01.438212 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" path="/var/lib/kubelet/pods/dc3d5d0d-95a1-4e28-8144-2ae3729d37d8/volumes" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.863001 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jxfvr"] Feb 02 13:50:07 crc kubenswrapper[4846]: E0202 13:50:07.864032 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff73f76-6b20-46a7-9919-ad9ccd05641b" containerName="octavia-amphora-httpd" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.864048 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff73f76-6b20-46a7-9919-ad9ccd05641b" containerName="octavia-amphora-httpd" Feb 02 13:50:07 crc kubenswrapper[4846]: E0202 13:50:07.864079 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerName="octavia-api" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.864086 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerName="octavia-api" Feb 02 13:50:07 crc kubenswrapper[4846]: E0202 13:50:07.864104 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerName="octavia-api-provider-agent" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.864111 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerName="octavia-api-provider-agent" Feb 02 13:50:07 crc kubenswrapper[4846]: E0202 13:50:07.864128 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerName="init" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.864136 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerName="init" Feb 02 13:50:07 crc kubenswrapper[4846]: E0202 13:50:07.864146 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff73f76-6b20-46a7-9919-ad9ccd05641b" containerName="init" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.864153 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff73f76-6b20-46a7-9919-ad9ccd05641b" containerName="init" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.864367 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ff73f76-6b20-46a7-9919-ad9ccd05641b" containerName="octavia-amphora-httpd" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.864398 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerName="octavia-api-provider-agent" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.864413 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc3d5d0d-95a1-4e28-8144-2ae3729d37d8" containerName="octavia-api" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.866120 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.875825 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxfvr"] Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.984546 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-utilities\") pod \"redhat-marketplace-jxfvr\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.984602 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-548pj\" (UniqueName: \"kubernetes.io/projected/59f33873-fa67-437d-97a1-651ea543ea5a-kube-api-access-548pj\") pod \"redhat-marketplace-jxfvr\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:07 crc kubenswrapper[4846]: I0202 13:50:07.984831 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-catalog-content\") pod \"redhat-marketplace-jxfvr\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.086581 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-utilities\") pod \"redhat-marketplace-jxfvr\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.086648 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-548pj\" (UniqueName: \"kubernetes.io/projected/59f33873-fa67-437d-97a1-651ea543ea5a-kube-api-access-548pj\") pod \"redhat-marketplace-jxfvr\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.086695 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-catalog-content\") pod \"redhat-marketplace-jxfvr\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.087176 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-catalog-content\") pod \"redhat-marketplace-jxfvr\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.087174 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-utilities\") pod \"redhat-marketplace-jxfvr\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.107643 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-548pj\" (UniqueName: \"kubernetes.io/projected/59f33873-fa67-437d-97a1-651ea543ea5a-kube-api-access-548pj\") pod \"redhat-marketplace-jxfvr\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.170788 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-65dd99cb46-gjk5s"] Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.172420 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.176404 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.183582 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-65dd99cb46-gjk5s"] Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.233728 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.294156 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7bae16d8-7405-45de-8a7f-e3714d1fa4eb-httpd-config\") pod \"octavia-image-upload-65dd99cb46-gjk5s\" (UID: \"7bae16d8-7405-45de-8a7f-e3714d1fa4eb\") " pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.294719 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7bae16d8-7405-45de-8a7f-e3714d1fa4eb-amphora-image\") pod \"octavia-image-upload-65dd99cb46-gjk5s\" (UID: \"7bae16d8-7405-45de-8a7f-e3714d1fa4eb\") " pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.397374 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7bae16d8-7405-45de-8a7f-e3714d1fa4eb-amphora-image\") pod \"octavia-image-upload-65dd99cb46-gjk5s\" (UID: \"7bae16d8-7405-45de-8a7f-e3714d1fa4eb\") " pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.397486 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7bae16d8-7405-45de-8a7f-e3714d1fa4eb-httpd-config\") pod \"octavia-image-upload-65dd99cb46-gjk5s\" (UID: \"7bae16d8-7405-45de-8a7f-e3714d1fa4eb\") " pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.399036 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7bae16d8-7405-45de-8a7f-e3714d1fa4eb-amphora-image\") pod \"octavia-image-upload-65dd99cb46-gjk5s\" (UID: \"7bae16d8-7405-45de-8a7f-e3714d1fa4eb\") " pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.412726 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7bae16d8-7405-45de-8a7f-e3714d1fa4eb-httpd-config\") pod \"octavia-image-upload-65dd99cb46-gjk5s\" (UID: \"7bae16d8-7405-45de-8a7f-e3714d1fa4eb\") " pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.494764 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.744496 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxfvr"] Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.872831 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxfvr" event={"ID":"59f33873-fa67-437d-97a1-651ea543ea5a","Type":"ContainerStarted","Data":"230dbe2f429657c0c5d0bf1674510fcfc2a5ebaa7dea669ac7ae39313a18a572"} Feb 02 13:50:08 crc kubenswrapper[4846]: I0202 13:50:08.965207 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-65dd99cb46-gjk5s"] Feb 02 13:50:09 crc kubenswrapper[4846]: I0202 13:50:09.882494 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" event={"ID":"7bae16d8-7405-45de-8a7f-e3714d1fa4eb","Type":"ContainerStarted","Data":"aca617c7e0de7642a1cec1cb1ff329ac7abf588090f81529098384dc973dfca0"} Feb 02 13:50:09 crc kubenswrapper[4846]: I0202 13:50:09.883068 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" event={"ID":"7bae16d8-7405-45de-8a7f-e3714d1fa4eb","Type":"ContainerStarted","Data":"0c19094523b6d7110da6a902239f1f348a904fc3ae9a26a4069eb0da95aa9352"} Feb 02 13:50:09 crc kubenswrapper[4846]: I0202 13:50:09.885904 4846 generic.go:334] "Generic (PLEG): container finished" podID="59f33873-fa67-437d-97a1-651ea543ea5a" containerID="91b34c2dcf3da75d4876649fda48ee7a3d416f5b192915eed72e760ed2a77445" exitCode=0 Feb 02 13:50:09 crc kubenswrapper[4846]: I0202 13:50:09.885936 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxfvr" event={"ID":"59f33873-fa67-437d-97a1-651ea543ea5a","Type":"ContainerDied","Data":"91b34c2dcf3da75d4876649fda48ee7a3d416f5b192915eed72e760ed2a77445"} Feb 02 13:50:10 crc kubenswrapper[4846]: I0202 13:50:10.895554 4846 generic.go:334] "Generic (PLEG): container finished" podID="7bae16d8-7405-45de-8a7f-e3714d1fa4eb" containerID="aca617c7e0de7642a1cec1cb1ff329ac7abf588090f81529098384dc973dfca0" exitCode=0 Feb 02 13:50:10 crc kubenswrapper[4846]: I0202 13:50:10.895614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" event={"ID":"7bae16d8-7405-45de-8a7f-e3714d1fa4eb","Type":"ContainerDied","Data":"aca617c7e0de7642a1cec1cb1ff329ac7abf588090f81529098384dc973dfca0"} Feb 02 13:50:11 crc kubenswrapper[4846]: I0202 13:50:11.910852 4846 generic.go:334] "Generic (PLEG): container finished" podID="59f33873-fa67-437d-97a1-651ea543ea5a" containerID="7ea25db6e970f35ff5c5a2d1c354a5f07715b42c2040b1e06e23ca9bb92235d1" exitCode=0 Feb 02 13:50:11 crc kubenswrapper[4846]: I0202 13:50:11.910973 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxfvr" event={"ID":"59f33873-fa67-437d-97a1-651ea543ea5a","Type":"ContainerDied","Data":"7ea25db6e970f35ff5c5a2d1c354a5f07715b42c2040b1e06e23ca9bb92235d1"} Feb 02 13:50:11 crc kubenswrapper[4846]: I0202 13:50:11.913980 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" event={"ID":"7bae16d8-7405-45de-8a7f-e3714d1fa4eb","Type":"ContainerStarted","Data":"476cdbc67c045581af14a7c8f11b40eb7a0b55e084a143137fcc4bd4e5667263"} Feb 02 13:50:11 crc kubenswrapper[4846]: I0202 13:50:11.964519 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-65dd99cb46-gjk5s" podStartSLOduration=3.493025235 podStartE2EDuration="3.964497305s" podCreationTimestamp="2026-02-02 13:50:08 +0000 UTC" firstStartedPulling="2026-02-02 13:50:09.042491593 +0000 UTC m=+6040.271078456" lastFinishedPulling="2026-02-02 13:50:09.513963663 +0000 UTC m=+6040.742550526" observedRunningTime="2026-02-02 13:50:11.944308295 +0000 UTC m=+6043.172895158" watchObservedRunningTime="2026-02-02 13:50:11.964497305 +0000 UTC m=+6043.193084188" Feb 02 13:50:12 crc kubenswrapper[4846]: I0202 13:50:12.934834 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxfvr" event={"ID":"59f33873-fa67-437d-97a1-651ea543ea5a","Type":"ContainerStarted","Data":"b087f4b9b88aa4255bdf8708423b332cd8dcd3b809497e5de3754abe1aff87f9"} Feb 02 13:50:12 crc kubenswrapper[4846]: I0202 13:50:12.961378 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jxfvr" podStartSLOduration=3.526938962 podStartE2EDuration="5.961356433s" podCreationTimestamp="2026-02-02 13:50:07 +0000 UTC" firstStartedPulling="2026-02-02 13:50:09.887430753 +0000 UTC m=+6041.116017646" lastFinishedPulling="2026-02-02 13:50:12.321848254 +0000 UTC m=+6043.550435117" observedRunningTime="2026-02-02 13:50:12.954518837 +0000 UTC m=+6044.183105720" watchObservedRunningTime="2026-02-02 13:50:12.961356433 +0000 UTC m=+6044.189943296" Feb 02 13:50:18 crc kubenswrapper[4846]: I0202 13:50:18.234883 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:18 crc kubenswrapper[4846]: I0202 13:50:18.235517 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:18 crc kubenswrapper[4846]: I0202 13:50:18.303479 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:19 crc kubenswrapper[4846]: I0202 13:50:19.043604 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:19 crc kubenswrapper[4846]: I0202 13:50:19.107830 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxfvr"] Feb 02 13:50:21 crc kubenswrapper[4846]: I0202 13:50:21.008569 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jxfvr" podUID="59f33873-fa67-437d-97a1-651ea543ea5a" containerName="registry-server" containerID="cri-o://b087f4b9b88aa4255bdf8708423b332cd8dcd3b809497e5de3754abe1aff87f9" gracePeriod=2 Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.018992 4846 generic.go:334] "Generic (PLEG): container finished" podID="59f33873-fa67-437d-97a1-651ea543ea5a" containerID="b087f4b9b88aa4255bdf8708423b332cd8dcd3b809497e5de3754abe1aff87f9" exitCode=0 Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.019060 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxfvr" event={"ID":"59f33873-fa67-437d-97a1-651ea543ea5a","Type":"ContainerDied","Data":"b087f4b9b88aa4255bdf8708423b332cd8dcd3b809497e5de3754abe1aff87f9"} Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.182487 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.370966 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-utilities\") pod \"59f33873-fa67-437d-97a1-651ea543ea5a\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.371095 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-catalog-content\") pod \"59f33873-fa67-437d-97a1-651ea543ea5a\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.371117 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-548pj\" (UniqueName: \"kubernetes.io/projected/59f33873-fa67-437d-97a1-651ea543ea5a-kube-api-access-548pj\") pod \"59f33873-fa67-437d-97a1-651ea543ea5a\" (UID: \"59f33873-fa67-437d-97a1-651ea543ea5a\") " Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.372565 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-utilities" (OuterVolumeSpecName: "utilities") pod "59f33873-fa67-437d-97a1-651ea543ea5a" (UID: "59f33873-fa67-437d-97a1-651ea543ea5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.380463 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59f33873-fa67-437d-97a1-651ea543ea5a-kube-api-access-548pj" (OuterVolumeSpecName: "kube-api-access-548pj") pod "59f33873-fa67-437d-97a1-651ea543ea5a" (UID: "59f33873-fa67-437d-97a1-651ea543ea5a"). InnerVolumeSpecName "kube-api-access-548pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.396548 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59f33873-fa67-437d-97a1-651ea543ea5a" (UID: "59f33873-fa67-437d-97a1-651ea543ea5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.473354 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.473398 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59f33873-fa67-437d-97a1-651ea543ea5a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:22 crc kubenswrapper[4846]: I0202 13:50:22.473414 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-548pj\" (UniqueName: \"kubernetes.io/projected/59f33873-fa67-437d-97a1-651ea543ea5a-kube-api-access-548pj\") on node \"crc\" DevicePath \"\"" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.032832 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxfvr" event={"ID":"59f33873-fa67-437d-97a1-651ea543ea5a","Type":"ContainerDied","Data":"230dbe2f429657c0c5d0bf1674510fcfc2a5ebaa7dea669ac7ae39313a18a572"} Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.032883 4846 scope.go:117] "RemoveContainer" containerID="b087f4b9b88aa4255bdf8708423b332cd8dcd3b809497e5de3754abe1aff87f9" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.033016 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxfvr" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.081335 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxfvr"] Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.092818 4846 scope.go:117] "RemoveContainer" containerID="7ea25db6e970f35ff5c5a2d1c354a5f07715b42c2040b1e06e23ca9bb92235d1" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.094417 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxfvr"] Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.117298 4846 scope.go:117] "RemoveContainer" containerID="91b34c2dcf3da75d4876649fda48ee7a3d416f5b192915eed72e760ed2a77445" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.336376 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-c782n"] Feb 02 13:50:23 crc kubenswrapper[4846]: E0202 13:50:23.346004 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f33873-fa67-437d-97a1-651ea543ea5a" containerName="registry-server" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.346038 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f33873-fa67-437d-97a1-651ea543ea5a" containerName="registry-server" Feb 02 13:50:23 crc kubenswrapper[4846]: E0202 13:50:23.346064 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f33873-fa67-437d-97a1-651ea543ea5a" containerName="extract-utilities" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.346075 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f33873-fa67-437d-97a1-651ea543ea5a" containerName="extract-utilities" Feb 02 13:50:23 crc kubenswrapper[4846]: E0202 13:50:23.346089 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f33873-fa67-437d-97a1-651ea543ea5a" containerName="extract-content" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.346094 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f33873-fa67-437d-97a1-651ea543ea5a" containerName="extract-content" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.346322 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f33873-fa67-437d-97a1-651ea543ea5a" containerName="registry-server" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.347342 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-c782n"] Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.347429 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.350265 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.350502 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.350706 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.435420 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59f33873-fa67-437d-97a1-651ea543ea5a" path="/var/lib/kubelet/pods/59f33873-fa67-437d-97a1-651ea543ea5a/volumes" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.491258 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-amphora-certs\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.491319 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-combined-ca-bundle\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.491360 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-scripts\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.491383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/70a6f659-55db-455b-9264-a9b4cefb6334-hm-ports\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.491467 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-config-data\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.491591 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/70a6f659-55db-455b-9264-a9b4cefb6334-config-data-merged\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.593486 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-amphora-certs\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.593566 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-combined-ca-bundle\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.593610 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-scripts\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.593667 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/70a6f659-55db-455b-9264-a9b4cefb6334-hm-ports\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.593816 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-config-data\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.593842 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/70a6f659-55db-455b-9264-a9b4cefb6334-config-data-merged\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.594851 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/70a6f659-55db-455b-9264-a9b4cefb6334-config-data-merged\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.595913 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/70a6f659-55db-455b-9264-a9b4cefb6334-hm-ports\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.598929 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-amphora-certs\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.599227 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-config-data\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.600770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-combined-ca-bundle\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.602914 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70a6f659-55db-455b-9264-a9b4cefb6334-scripts\") pod \"octavia-healthmanager-c782n\" (UID: \"70a6f659-55db-455b-9264-a9b4cefb6334\") " pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:23 crc kubenswrapper[4846]: I0202 13:50:23.673447 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:24 crc kubenswrapper[4846]: I0202 13:50:24.415887 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-c782n"] Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.052683 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-c782n" event={"ID":"70a6f659-55db-455b-9264-a9b4cefb6334","Type":"ContainerStarted","Data":"daa20b4262d4da5b9fa409c948389b4371074c56197522b56f53d0f6765c5b58"} Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.053168 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-c782n" event={"ID":"70a6f659-55db-455b-9264-a9b4cefb6334","Type":"ContainerStarted","Data":"38419800f9d861862e1bcb79ce318c246d036ccb634872c4ce596469714109ac"} Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.256238 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-cdp5l"] Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.258431 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.260781 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.261094 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.270604 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-cdp5l"] Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.429372 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/70f3f665-75af-4ef6-b89f-4479385845af-hm-ports\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.429421 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-scripts\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.429535 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/70f3f665-75af-4ef6-b89f-4479385845af-config-data-merged\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.429604 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-combined-ca-bundle\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.429650 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-amphora-certs\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.429727 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-config-data\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.531525 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/70f3f665-75af-4ef6-b89f-4479385845af-config-data-merged\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.531668 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-combined-ca-bundle\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.531699 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-amphora-certs\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.531797 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-config-data\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.532054 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/70f3f665-75af-4ef6-b89f-4479385845af-config-data-merged\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.533014 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/70f3f665-75af-4ef6-b89f-4479385845af-hm-ports\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.533070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-scripts\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.533921 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/70f3f665-75af-4ef6-b89f-4479385845af-hm-ports\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.538117 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-combined-ca-bundle\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.538483 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-amphora-certs\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.548038 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-scripts\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.560614 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f3f665-75af-4ef6-b89f-4479385845af-config-data\") pod \"octavia-housekeeping-cdp5l\" (UID: \"70f3f665-75af-4ef6-b89f-4479385845af\") " pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:25 crc kubenswrapper[4846]: I0202 13:50:25.590483 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:26 crc kubenswrapper[4846]: I0202 13:50:26.186593 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-cdp5l"] Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.071234 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-cdp5l" event={"ID":"70f3f665-75af-4ef6-b89f-4479385845af","Type":"ContainerStarted","Data":"230281030c47fd424d186011364add0ec41b40b931548a8e7f873c643c9e9e20"} Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.074072 4846 generic.go:334] "Generic (PLEG): container finished" podID="70a6f659-55db-455b-9264-a9b4cefb6334" containerID="daa20b4262d4da5b9fa409c948389b4371074c56197522b56f53d0f6765c5b58" exitCode=0 Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.074116 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-c782n" event={"ID":"70a6f659-55db-455b-9264-a9b4cefb6334","Type":"ContainerDied","Data":"daa20b4262d4da5b9fa409c948389b4371074c56197522b56f53d0f6765c5b58"} Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.266237 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-7jg2f"] Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.268317 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.271162 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.271451 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.278723 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-7jg2f"] Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.381092 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-scripts\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.381154 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-config-data-merged\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.381243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-combined-ca-bundle\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.381299 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-hm-ports\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.381328 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-amphora-certs\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.381355 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-config-data\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.483537 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-combined-ca-bundle\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.483685 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-hm-ports\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.483735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-amphora-certs\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.483776 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-config-data\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.483816 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-scripts\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.483860 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-config-data-merged\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.484814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-config-data-merged\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.485406 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-hm-ports\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.489505 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-amphora-certs\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.489825 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-combined-ca-bundle\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.489893 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-scripts\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.492215 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea7dbcb7-36a7-4606-ba9f-a85ecb78640a-config-data\") pod \"octavia-worker-7jg2f\" (UID: \"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a\") " pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:27 crc kubenswrapper[4846]: I0202 13:50:27.595516 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:28 crc kubenswrapper[4846]: I0202 13:50:28.087584 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-c782n" event={"ID":"70a6f659-55db-455b-9264-a9b4cefb6334","Type":"ContainerStarted","Data":"b4f046bab6f6eef7729e048862d838916d557a51eb7c8c46b04a2183edc00129"} Feb 02 13:50:28 crc kubenswrapper[4846]: I0202 13:50:28.088095 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:28 crc kubenswrapper[4846]: I0202 13:50:28.151587 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-c782n" podStartSLOduration=5.151568051 podStartE2EDuration="5.151568051s" podCreationTimestamp="2026-02-02 13:50:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:50:28.13706279 +0000 UTC m=+6059.365649663" watchObservedRunningTime="2026-02-02 13:50:28.151568051 +0000 UTC m=+6059.380154914" Feb 02 13:50:28 crc kubenswrapper[4846]: I0202 13:50:28.261017 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-7jg2f"] Feb 02 13:50:28 crc kubenswrapper[4846]: I0202 13:50:28.893130 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-c782n"] Feb 02 13:50:29 crc kubenswrapper[4846]: I0202 13:50:29.098750 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-7jg2f" event={"ID":"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a","Type":"ContainerStarted","Data":"fcd0a14f0ff2f7d7310822d14af9750e85414ddb4f8c0bd63ec6b2da4d552f49"} Feb 02 13:50:29 crc kubenswrapper[4846]: I0202 13:50:29.102602 4846 generic.go:334] "Generic (PLEG): container finished" podID="70f3f665-75af-4ef6-b89f-4479385845af" containerID="aedce66995f5ce1038e746ed0ca4a0e6c014a3388b0cab9790fa94b99b7dcda6" exitCode=0 Feb 02 13:50:29 crc kubenswrapper[4846]: I0202 13:50:29.102703 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-cdp5l" event={"ID":"70f3f665-75af-4ef6-b89f-4479385845af","Type":"ContainerDied","Data":"aedce66995f5ce1038e746ed0ca4a0e6c014a3388b0cab9790fa94b99b7dcda6"} Feb 02 13:50:31 crc kubenswrapper[4846]: I0202 13:50:31.123035 4846 generic.go:334] "Generic (PLEG): container finished" podID="ea7dbcb7-36a7-4606-ba9f-a85ecb78640a" containerID="46d50d086f2e214c0b2f9ce862cc9572ae64e37f8c28b2109c30b286f7b7fb8f" exitCode=0 Feb 02 13:50:31 crc kubenswrapper[4846]: I0202 13:50:31.123094 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-7jg2f" event={"ID":"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a","Type":"ContainerDied","Data":"46d50d086f2e214c0b2f9ce862cc9572ae64e37f8c28b2109c30b286f7b7fb8f"} Feb 02 13:50:31 crc kubenswrapper[4846]: I0202 13:50:31.126238 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-cdp5l" event={"ID":"70f3f665-75af-4ef6-b89f-4479385845af","Type":"ContainerStarted","Data":"71fbb9b57a44103769d2872144d717b348716cd322aa1ff559eab8a464869e29"} Feb 02 13:50:31 crc kubenswrapper[4846]: I0202 13:50:31.126388 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:31 crc kubenswrapper[4846]: I0202 13:50:31.217090 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-cdp5l" podStartSLOduration=4.7000899579999995 podStartE2EDuration="6.217064889s" podCreationTimestamp="2026-02-02 13:50:25 +0000 UTC" firstStartedPulling="2026-02-02 13:50:26.199367702 +0000 UTC m=+6057.427954565" lastFinishedPulling="2026-02-02 13:50:27.716342643 +0000 UTC m=+6058.944929496" observedRunningTime="2026-02-02 13:50:31.188268419 +0000 UTC m=+6062.416855282" watchObservedRunningTime="2026-02-02 13:50:31.217064889 +0000 UTC m=+6062.445651752" Feb 02 13:50:32 crc kubenswrapper[4846]: I0202 13:50:32.143818 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-7jg2f" event={"ID":"ea7dbcb7-36a7-4606-ba9f-a85ecb78640a","Type":"ContainerStarted","Data":"6094ac6bd0e4af9893defa2891f9499edc2c28f9ef45d6845e14ac166235fbca"} Feb 02 13:50:32 crc kubenswrapper[4846]: I0202 13:50:32.172940 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-7jg2f" podStartSLOduration=3.761135075 podStartE2EDuration="5.172923591s" podCreationTimestamp="2026-02-02 13:50:27 +0000 UTC" firstStartedPulling="2026-02-02 13:50:28.270227233 +0000 UTC m=+6059.498814096" lastFinishedPulling="2026-02-02 13:50:29.682015749 +0000 UTC m=+6060.910602612" observedRunningTime="2026-02-02 13:50:32.168383031 +0000 UTC m=+6063.396969904" watchObservedRunningTime="2026-02-02 13:50:32.172923591 +0000 UTC m=+6063.401510454" Feb 02 13:50:33 crc kubenswrapper[4846]: I0202 13:50:33.157772 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-7jg2f" Feb 02 13:50:38 crc kubenswrapper[4846]: I0202 13:50:38.709992 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-c782n" Feb 02 13:50:40 crc kubenswrapper[4846]: I0202 13:50:40.636791 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-cdp5l" Feb 02 13:50:42 crc kubenswrapper[4846]: I0202 13:50:42.625805 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-7jg2f" Feb 02 13:51:00 crc kubenswrapper[4846]: I0202 13:51:00.479097 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:51:00 crc kubenswrapper[4846]: I0202 13:51:00.480022 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.568284 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-48ctv"] Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.574935 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.582413 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48ctv"] Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.656042 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpwdh\" (UniqueName: \"kubernetes.io/projected/26a0bb46-1e22-4ff5-bc22-d54540212e99-kube-api-access-cpwdh\") pod \"redhat-operators-48ctv\" (UID: \"26a0bb46-1e22-4ff5-bc22-d54540212e99\") " pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.656228 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26a0bb46-1e22-4ff5-bc22-d54540212e99-catalog-content\") pod \"redhat-operators-48ctv\" (UID: \"26a0bb46-1e22-4ff5-bc22-d54540212e99\") " pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.656270 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26a0bb46-1e22-4ff5-bc22-d54540212e99-utilities\") pod \"redhat-operators-48ctv\" (UID: \"26a0bb46-1e22-4ff5-bc22-d54540212e99\") " pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.758666 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26a0bb46-1e22-4ff5-bc22-d54540212e99-catalog-content\") pod \"redhat-operators-48ctv\" (UID: \"26a0bb46-1e22-4ff5-bc22-d54540212e99\") " pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.759211 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26a0bb46-1e22-4ff5-bc22-d54540212e99-utilities\") pod \"redhat-operators-48ctv\" (UID: \"26a0bb46-1e22-4ff5-bc22-d54540212e99\") " pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.759151 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26a0bb46-1e22-4ff5-bc22-d54540212e99-catalog-content\") pod \"redhat-operators-48ctv\" (UID: \"26a0bb46-1e22-4ff5-bc22-d54540212e99\") " pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.759288 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpwdh\" (UniqueName: \"kubernetes.io/projected/26a0bb46-1e22-4ff5-bc22-d54540212e99-kube-api-access-cpwdh\") pod \"redhat-operators-48ctv\" (UID: \"26a0bb46-1e22-4ff5-bc22-d54540212e99\") " pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.759696 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26a0bb46-1e22-4ff5-bc22-d54540212e99-utilities\") pod \"redhat-operators-48ctv\" (UID: \"26a0bb46-1e22-4ff5-bc22-d54540212e99\") " pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.780489 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpwdh\" (UniqueName: \"kubernetes.io/projected/26a0bb46-1e22-4ff5-bc22-d54540212e99-kube-api-access-cpwdh\") pod \"redhat-operators-48ctv\" (UID: \"26a0bb46-1e22-4ff5-bc22-d54540212e99\") " pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:01 crc kubenswrapper[4846]: I0202 13:51:01.906505 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:02 crc kubenswrapper[4846]: I0202 13:51:02.401664 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48ctv"] Feb 02 13:51:02 crc kubenswrapper[4846]: I0202 13:51:02.457668 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48ctv" event={"ID":"26a0bb46-1e22-4ff5-bc22-d54540212e99","Type":"ContainerStarted","Data":"7c8d11af8957e4403fd6c527c7358603de98c32b2ef159037691bb270cabbc1f"} Feb 02 13:51:03 crc kubenswrapper[4846]: I0202 13:51:03.467409 4846 generic.go:334] "Generic (PLEG): container finished" podID="26a0bb46-1e22-4ff5-bc22-d54540212e99" containerID="0f82874cbcded6433e34aabf8dab40cfaf5b006f0c88a096ee7ce7d953ba5c3b" exitCode=0 Feb 02 13:51:03 crc kubenswrapper[4846]: I0202 13:51:03.467540 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48ctv" event={"ID":"26a0bb46-1e22-4ff5-bc22-d54540212e99","Type":"ContainerDied","Data":"0f82874cbcded6433e34aabf8dab40cfaf5b006f0c88a096ee7ce7d953ba5c3b"} Feb 02 13:51:16 crc kubenswrapper[4846]: I0202 13:51:16.598882 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48ctv" event={"ID":"26a0bb46-1e22-4ff5-bc22-d54540212e99","Type":"ContainerStarted","Data":"d4857bb06abaf4f3c3af31a82e798114d33c675027486072ef0ac077f36c1fa2"} Feb 02 13:51:18 crc kubenswrapper[4846]: I0202 13:51:18.618171 4846 generic.go:334] "Generic (PLEG): container finished" podID="26a0bb46-1e22-4ff5-bc22-d54540212e99" containerID="d4857bb06abaf4f3c3af31a82e798114d33c675027486072ef0ac077f36c1fa2" exitCode=0 Feb 02 13:51:18 crc kubenswrapper[4846]: I0202 13:51:18.618771 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48ctv" event={"ID":"26a0bb46-1e22-4ff5-bc22-d54540212e99","Type":"ContainerDied","Data":"d4857bb06abaf4f3c3af31a82e798114d33c675027486072ef0ac077f36c1fa2"} Feb 02 13:51:19 crc kubenswrapper[4846]: I0202 13:51:19.662169 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48ctv" event={"ID":"26a0bb46-1e22-4ff5-bc22-d54540212e99","Type":"ContainerStarted","Data":"361bfca9b38f610ed425042ed84ee933cc3aef50caba40eda8af988035fd6acb"} Feb 02 13:51:19 crc kubenswrapper[4846]: I0202 13:51:19.698661 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-48ctv" podStartSLOduration=3.162694161 podStartE2EDuration="18.698639346s" podCreationTimestamp="2026-02-02 13:51:01 +0000 UTC" firstStartedPulling="2026-02-02 13:51:03.470287825 +0000 UTC m=+6094.698874688" lastFinishedPulling="2026-02-02 13:51:19.006233 +0000 UTC m=+6110.234819873" observedRunningTime="2026-02-02 13:51:19.694575587 +0000 UTC m=+6110.923162450" watchObservedRunningTime="2026-02-02 13:51:19.698639346 +0000 UTC m=+6110.927226219" Feb 02 13:51:21 crc kubenswrapper[4846]: I0202 13:51:21.906848 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:21 crc kubenswrapper[4846]: I0202 13:51:21.907129 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:22 crc kubenswrapper[4846]: I0202 13:51:22.956583 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-48ctv" podUID="26a0bb46-1e22-4ff5-bc22-d54540212e99" containerName="registry-server" probeResult="failure" output=< Feb 02 13:51:22 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 13:51:22 crc kubenswrapper[4846]: > Feb 02 13:51:30 crc kubenswrapper[4846]: I0202 13:51:30.479264 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:51:30 crc kubenswrapper[4846]: I0202 13:51:30.479885 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:51:31 crc kubenswrapper[4846]: I0202 13:51:31.955746 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:32 crc kubenswrapper[4846]: I0202 13:51:32.004468 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-48ctv" Feb 02 13:51:32 crc kubenswrapper[4846]: I0202 13:51:32.593666 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48ctv"] Feb 02 13:51:32 crc kubenswrapper[4846]: I0202 13:51:32.770365 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7hsfb"] Feb 02 13:51:32 crc kubenswrapper[4846]: I0202 13:51:32.770592 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7hsfb" podUID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerName="registry-server" containerID="cri-o://3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a" gracePeriod=2 Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.385221 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.423031 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-utilities\") pod \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.423181 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56fw9\" (UniqueName: \"kubernetes.io/projected/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-kube-api-access-56fw9\") pod \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.423233 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-catalog-content\") pod \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\" (UID: \"2fe8aed5-574d-47f9-8822-f7a42dbccb5a\") " Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.432753 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-utilities" (OuterVolumeSpecName: "utilities") pod "2fe8aed5-574d-47f9-8822-f7a42dbccb5a" (UID: "2fe8aed5-574d-47f9-8822-f7a42dbccb5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.440956 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-kube-api-access-56fw9" (OuterVolumeSpecName: "kube-api-access-56fw9") pod "2fe8aed5-574d-47f9-8822-f7a42dbccb5a" (UID: "2fe8aed5-574d-47f9-8822-f7a42dbccb5a"). InnerVolumeSpecName "kube-api-access-56fw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.528483 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.528546 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56fw9\" (UniqueName: \"kubernetes.io/projected/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-kube-api-access-56fw9\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.590862 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fe8aed5-574d-47f9-8822-f7a42dbccb5a" (UID: "2fe8aed5-574d-47f9-8822-f7a42dbccb5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.631176 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fe8aed5-574d-47f9-8822-f7a42dbccb5a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.781433 4846 generic.go:334] "Generic (PLEG): container finished" podID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerID="3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a" exitCode=0 Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.781552 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7hsfb" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.781598 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hsfb" event={"ID":"2fe8aed5-574d-47f9-8822-f7a42dbccb5a","Type":"ContainerDied","Data":"3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a"} Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.781645 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7hsfb" event={"ID":"2fe8aed5-574d-47f9-8822-f7a42dbccb5a","Type":"ContainerDied","Data":"868509b99a7e621fcc156f8b60c01efa5454728c7a23c49018ae95adfa6c982a"} Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.781662 4846 scope.go:117] "RemoveContainer" containerID="3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.817797 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7hsfb"] Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.822045 4846 scope.go:117] "RemoveContainer" containerID="db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.831042 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7hsfb"] Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.846140 4846 scope.go:117] "RemoveContainer" containerID="68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.906318 4846 scope.go:117] "RemoveContainer" containerID="3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a" Feb 02 13:51:33 crc kubenswrapper[4846]: E0202 13:51:33.906823 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a\": container with ID starting with 3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a not found: ID does not exist" containerID="3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.906853 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a"} err="failed to get container status \"3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a\": rpc error: code = NotFound desc = could not find container \"3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a\": container with ID starting with 3275817be0e4041484f8c0d97cf384c61d4d79d01370556d6d5c4ac1b116ba9a not found: ID does not exist" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.906876 4846 scope.go:117] "RemoveContainer" containerID="db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee" Feb 02 13:51:33 crc kubenswrapper[4846]: E0202 13:51:33.907198 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee\": container with ID starting with db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee not found: ID does not exist" containerID="db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.907249 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee"} err="failed to get container status \"db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee\": rpc error: code = NotFound desc = could not find container \"db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee\": container with ID starting with db6f9c721341b88d8513fa6f543457e753bf6cfa8e723ed6282f5c6edb7324ee not found: ID does not exist" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.907274 4846 scope.go:117] "RemoveContainer" containerID="68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc" Feb 02 13:51:33 crc kubenswrapper[4846]: E0202 13:51:33.907605 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc\": container with ID starting with 68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc not found: ID does not exist" containerID="68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc" Feb 02 13:51:33 crc kubenswrapper[4846]: I0202 13:51:33.907759 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc"} err="failed to get container status \"68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc\": rpc error: code = NotFound desc = could not find container \"68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc\": container with ID starting with 68a07506bc3d94f807d9038f6fcbd1ff6e8aef542cf3338c8e528e5744b8d1bc not found: ID does not exist" Feb 02 13:51:34 crc kubenswrapper[4846]: I0202 13:51:34.055691 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-sxfkf"] Feb 02 13:51:34 crc kubenswrapper[4846]: I0202 13:51:34.064767 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-da7f-account-create-update-tbvtt"] Feb 02 13:51:34 crc kubenswrapper[4846]: I0202 13:51:34.074280 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-da7f-account-create-update-tbvtt"] Feb 02 13:51:34 crc kubenswrapper[4846]: I0202 13:51:34.083475 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-sxfkf"] Feb 02 13:51:35 crc kubenswrapper[4846]: I0202 13:51:35.434133 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" path="/var/lib/kubelet/pods/2fe8aed5-574d-47f9-8822-f7a42dbccb5a/volumes" Feb 02 13:51:35 crc kubenswrapper[4846]: I0202 13:51:35.436703 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9de90124-4479-4ae4-a6c2-e12c41df67b4" path="/var/lib/kubelet/pods/9de90124-4479-4ae4-a6c2-e12c41df67b4/volumes" Feb 02 13:51:35 crc kubenswrapper[4846]: I0202 13:51:35.439284 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff66eedf-5da6-4676-9a52-fd6e5d1df411" path="/var/lib/kubelet/pods/ff66eedf-5da6-4676-9a52-fd6e5d1df411/volumes" Feb 02 13:51:39 crc kubenswrapper[4846]: I0202 13:51:39.714091 4846 scope.go:117] "RemoveContainer" containerID="4175c9c531fb6923c23164ac327144d04119c8099e366ea3af132f5ea1af4d14" Feb 02 13:51:39 crc kubenswrapper[4846]: I0202 13:51:39.840224 4846 scope.go:117] "RemoveContainer" containerID="6fff3bdb8343bc093cc9c6be08d6845a546b8a2009d749868f3475a9f27c7ee3" Feb 02 13:51:42 crc kubenswrapper[4846]: I0202 13:51:42.032543 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9r479"] Feb 02 13:51:42 crc kubenswrapper[4846]: I0202 13:51:42.045265 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9r479"] Feb 02 13:51:43 crc kubenswrapper[4846]: I0202 13:51:43.436483 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="957b0ccd-612b-4e78-909e-048644d9fd17" path="/var/lib/kubelet/pods/957b0ccd-612b-4e78-909e-048644d9fd17/volumes" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.669578 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79f6dfcdc7-gb2k8"] Feb 02 13:51:45 crc kubenswrapper[4846]: E0202 13:51:45.671314 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerName="registry-server" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.671342 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerName="registry-server" Feb 02 13:51:45 crc kubenswrapper[4846]: E0202 13:51:45.671375 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerName="extract-content" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.671384 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerName="extract-content" Feb 02 13:51:45 crc kubenswrapper[4846]: E0202 13:51:45.671410 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerName="extract-utilities" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.671421 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerName="extract-utilities" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.671812 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fe8aed5-574d-47f9-8822-f7a42dbccb5a" containerName="registry-server" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.677333 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.681098 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.689165 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.689217 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.689779 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-gl4b4" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.712733 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79f6dfcdc7-gb2k8"] Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.768463 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14cb62be-e7b0-4905-9829-fa63566b377c-logs\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.768666 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-scripts\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.768692 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-config-data\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.768747 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/14cb62be-e7b0-4905-9829-fa63566b377c-horizon-secret-key\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.768795 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s5ld\" (UniqueName: \"kubernetes.io/projected/14cb62be-e7b0-4905-9829-fa63566b377c-kube-api-access-9s5ld\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.787847 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.788148 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6c8c7249-5482-4258-a3a7-069aa4c087dd" containerName="glance-log" containerID="cri-o://418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71" gracePeriod=30 Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.788321 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6c8c7249-5482-4258-a3a7-069aa4c087dd" containerName="glance-httpd" containerID="cri-o://f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013" gracePeriod=30 Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.798304 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-556b694cbc-pxk5w"] Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.800353 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.818743 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-556b694cbc-pxk5w"] Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.835171 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.835454 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" containerName="glance-log" containerID="cri-o://4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20" gracePeriod=30 Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.835644 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" containerName="glance-httpd" containerID="cri-o://41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5" gracePeriod=30 Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.869730 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-horizon-secret-key\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.870043 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/14cb62be-e7b0-4905-9829-fa63566b377c-horizon-secret-key\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.870063 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-config-data\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.870083 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s5ld\" (UniqueName: \"kubernetes.io/projected/14cb62be-e7b0-4905-9829-fa63566b377c-kube-api-access-9s5ld\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.870102 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-logs\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.870417 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14cb62be-e7b0-4905-9829-fa63566b377c-logs\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.870607 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-scripts\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.870661 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-config-data\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.870696 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-scripts\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.870762 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjrv7\" (UniqueName: \"kubernetes.io/projected/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-kube-api-access-fjrv7\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.871226 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14cb62be-e7b0-4905-9829-fa63566b377c-logs\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.871745 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-scripts\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.872825 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-config-data\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.888172 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/14cb62be-e7b0-4905-9829-fa63566b377c-horizon-secret-key\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.888456 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s5ld\" (UniqueName: \"kubernetes.io/projected/14cb62be-e7b0-4905-9829-fa63566b377c-kube-api-access-9s5ld\") pod \"horizon-79f6dfcdc7-gb2k8\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.974430 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-scripts\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.974492 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjrv7\" (UniqueName: \"kubernetes.io/projected/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-kube-api-access-fjrv7\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.974530 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-horizon-secret-key\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.974549 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-config-data\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.974572 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-logs\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.975081 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-logs\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.975607 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-scripts\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.976551 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-config-data\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:45 crc kubenswrapper[4846]: I0202 13:51:45.994094 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-horizon-secret-key\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.007280 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjrv7\" (UniqueName: \"kubernetes.io/projected/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-kube-api-access-fjrv7\") pod \"horizon-556b694cbc-pxk5w\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.015084 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.147554 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.594146 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79f6dfcdc7-gb2k8"] Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.693418 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-556b694cbc-pxk5w"] Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.923740 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" containerID="4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20" exitCode=143 Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.924104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa2668b1-bd01-4fce-93fd-d7fbf9917247","Type":"ContainerDied","Data":"4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20"} Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.925406 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f6dfcdc7-gb2k8" event={"ID":"14cb62be-e7b0-4905-9829-fa63566b377c","Type":"ContainerStarted","Data":"a3ad2915cd67b8340325ad2f1a70e9846a486d9cf027f6d7416859ef6b7a121b"} Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.927982 4846 generic.go:334] "Generic (PLEG): container finished" podID="6c8c7249-5482-4258-a3a7-069aa4c087dd" containerID="418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71" exitCode=143 Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.928032 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c8c7249-5482-4258-a3a7-069aa4c087dd","Type":"ContainerDied","Data":"418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71"} Feb 02 13:51:46 crc kubenswrapper[4846]: I0202 13:51:46.929536 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556b694cbc-pxk5w" event={"ID":"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c","Type":"ContainerStarted","Data":"89e7b78472b01d7c1c132b9a93332668485f128e7cab2d3e233863275003b66a"} Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.544938 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-556b694cbc-pxk5w"] Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.573345 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-766fbf8cc6-nf7ws"] Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.581455 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.585805 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.589051 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-766fbf8cc6-nf7ws"] Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.617816 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8z28\" (UniqueName: \"kubernetes.io/projected/3bada3e7-480c-45df-bc54-452587b69d52-kube-api-access-j8z28\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.617893 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-tls-certs\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.617950 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-config-data\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.617985 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bada3e7-480c-45df-bc54-452587b69d52-logs\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.618013 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-scripts\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.618103 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-secret-key\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.618139 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-combined-ca-bundle\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.660137 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79f6dfcdc7-gb2k8"] Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.700311 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d584cfd6-zscgj"] Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.702302 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720301 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-secret-key\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720367 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-secret-key\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720403 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-combined-ca-bundle\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720421 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-tls-certs\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720448 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-config-data\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720504 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccvc9\" (UniqueName: \"kubernetes.io/projected/510040f2-b874-48d8-bfb1-871b8abf0199-kube-api-access-ccvc9\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720596 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8z28\" (UniqueName: \"kubernetes.io/projected/3bada3e7-480c-45df-bc54-452587b69d52-kube-api-access-j8z28\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720702 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-tls-certs\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/510040f2-b874-48d8-bfb1-871b8abf0199-logs\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720806 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-scripts\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720846 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-config-data\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720903 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bada3e7-480c-45df-bc54-452587b69d52-logs\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720936 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-scripts\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.720977 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-combined-ca-bundle\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.722951 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-config-data\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.723211 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bada3e7-480c-45df-bc54-452587b69d52-logs\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.723586 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-scripts\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.725120 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d584cfd6-zscgj"] Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.731377 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-secret-key\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.733927 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-combined-ca-bundle\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.740272 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-tls-certs\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.750120 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8z28\" (UniqueName: \"kubernetes.io/projected/3bada3e7-480c-45df-bc54-452587b69d52-kube-api-access-j8z28\") pod \"horizon-766fbf8cc6-nf7ws\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.822439 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-combined-ca-bundle\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.822535 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-secret-key\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.822566 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-tls-certs\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.822593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-config-data\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.822647 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccvc9\" (UniqueName: \"kubernetes.io/projected/510040f2-b874-48d8-bfb1-871b8abf0199-kube-api-access-ccvc9\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.822738 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/510040f2-b874-48d8-bfb1-871b8abf0199-logs\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.822789 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-scripts\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.823505 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-scripts\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.823708 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/510040f2-b874-48d8-bfb1-871b8abf0199-logs\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.824395 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-config-data\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.826029 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-combined-ca-bundle\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.826702 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-secret-key\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.826704 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-tls-certs\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.839960 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccvc9\" (UniqueName: \"kubernetes.io/projected/510040f2-b874-48d8-bfb1-871b8abf0199-kube-api-access-ccvc9\") pod \"horizon-d584cfd6-zscgj\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:47 crc kubenswrapper[4846]: I0202 13:51:47.926734 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:48 crc kubenswrapper[4846]: I0202 13:51:48.018587 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:48 crc kubenswrapper[4846]: I0202 13:51:48.421971 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-766fbf8cc6-nf7ws"] Feb 02 13:51:48 crc kubenswrapper[4846]: I0202 13:51:48.676576 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d584cfd6-zscgj"] Feb 02 13:51:48 crc kubenswrapper[4846]: W0202 13:51:48.684839 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod510040f2_b874_48d8_bfb1_871b8abf0199.slice/crio-79842b909a357865e62bad261aad7a07f20c124f6847c50b9ab0bb31f90711d7 WatchSource:0}: Error finding container 79842b909a357865e62bad261aad7a07f20c124f6847c50b9ab0bb31f90711d7: Status 404 returned error can't find the container with id 79842b909a357865e62bad261aad7a07f20c124f6847c50b9ab0bb31f90711d7 Feb 02 13:51:48 crc kubenswrapper[4846]: I0202 13:51:48.952161 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d584cfd6-zscgj" event={"ID":"510040f2-b874-48d8-bfb1-871b8abf0199","Type":"ContainerStarted","Data":"79842b909a357865e62bad261aad7a07f20c124f6847c50b9ab0bb31f90711d7"} Feb 02 13:51:48 crc kubenswrapper[4846]: I0202 13:51:48.954180 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766fbf8cc6-nf7ws" event={"ID":"3bada3e7-480c-45df-bc54-452587b69d52","Type":"ContainerStarted","Data":"6f56143bf78c066b2871c244f7be0b52a4bd2e55df089d7f4fc2d15e085a9ca2"} Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.730394 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.795741 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-config-data\") pod \"6c8c7249-5482-4258-a3a7-069aa4c087dd\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.795805 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-public-tls-certs\") pod \"6c8c7249-5482-4258-a3a7-069aa4c087dd\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.795905 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn9kz\" (UniqueName: \"kubernetes.io/projected/6c8c7249-5482-4258-a3a7-069aa4c087dd-kube-api-access-sn9kz\") pod \"6c8c7249-5482-4258-a3a7-069aa4c087dd\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.795954 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-scripts\") pod \"6c8c7249-5482-4258-a3a7-069aa4c087dd\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.796071 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-combined-ca-bundle\") pod \"6c8c7249-5482-4258-a3a7-069aa4c087dd\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.796093 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-httpd-run\") pod \"6c8c7249-5482-4258-a3a7-069aa4c087dd\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.796150 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-logs\") pod \"6c8c7249-5482-4258-a3a7-069aa4c087dd\" (UID: \"6c8c7249-5482-4258-a3a7-069aa4c087dd\") " Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.797781 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-logs" (OuterVolumeSpecName: "logs") pod "6c8c7249-5482-4258-a3a7-069aa4c087dd" (UID: "6c8c7249-5482-4258-a3a7-069aa4c087dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.802219 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6c8c7249-5482-4258-a3a7-069aa4c087dd" (UID: "6c8c7249-5482-4258-a3a7-069aa4c087dd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.811048 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-scripts" (OuterVolumeSpecName: "scripts") pod "6c8c7249-5482-4258-a3a7-069aa4c087dd" (UID: "6c8c7249-5482-4258-a3a7-069aa4c087dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.811974 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c8c7249-5482-4258-a3a7-069aa4c087dd-kube-api-access-sn9kz" (OuterVolumeSpecName: "kube-api-access-sn9kz") pod "6c8c7249-5482-4258-a3a7-069aa4c087dd" (UID: "6c8c7249-5482-4258-a3a7-069aa4c087dd"). InnerVolumeSpecName "kube-api-access-sn9kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.856187 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c8c7249-5482-4258-a3a7-069aa4c087dd" (UID: "6c8c7249-5482-4258-a3a7-069aa4c087dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.869238 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6c8c7249-5482-4258-a3a7-069aa4c087dd" (UID: "6c8c7249-5482-4258-a3a7-069aa4c087dd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.869576 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-config-data" (OuterVolumeSpecName: "config-data") pod "6c8c7249-5482-4258-a3a7-069aa4c087dd" (UID: "6c8c7249-5482-4258-a3a7-069aa4c087dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.878881 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.899629 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.899679 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.899694 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.899736 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn9kz\" (UniqueName: \"kubernetes.io/projected/6c8c7249-5482-4258-a3a7-069aa4c087dd-kube-api-access-sn9kz\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.899755 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.899767 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c8c7249-5482-4258-a3a7-069aa4c087dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.899803 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6c8c7249-5482-4258-a3a7-069aa4c087dd-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.968728 4846 generic.go:334] "Generic (PLEG): container finished" podID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" containerID="41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5" exitCode=0 Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.968781 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa2668b1-bd01-4fce-93fd-d7fbf9917247","Type":"ContainerDied","Data":"41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5"} Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.968806 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa2668b1-bd01-4fce-93fd-d7fbf9917247","Type":"ContainerDied","Data":"f5b4b95ecadde2a0e7f008aa24d9fc7252cf1b192c151c9987603d74a20fa0cd"} Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.968823 4846 scope.go:117] "RemoveContainer" containerID="41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.968930 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.985497 4846 generic.go:334] "Generic (PLEG): container finished" podID="6c8c7249-5482-4258-a3a7-069aa4c087dd" containerID="f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013" exitCode=0 Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.985548 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c8c7249-5482-4258-a3a7-069aa4c087dd","Type":"ContainerDied","Data":"f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013"} Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.985579 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6c8c7249-5482-4258-a3a7-069aa4c087dd","Type":"ContainerDied","Data":"421880557b5059cba6ac3045d8787aa3bd81b82a0120944659f2a0395a6db752"} Feb 02 13:51:49 crc kubenswrapper[4846]: I0202 13:51:49.985658 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.000532 4846 scope.go:117] "RemoveContainer" containerID="4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.001140 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-combined-ca-bundle\") pod \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.001215 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-scripts\") pod \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.001304 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-internal-tls-certs\") pod \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.001422 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfp5w\" (UniqueName: \"kubernetes.io/projected/aa2668b1-bd01-4fce-93fd-d7fbf9917247-kube-api-access-lfp5w\") pod \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.001562 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-httpd-run\") pod \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.001594 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-config-data\") pod \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.001676 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-logs\") pod \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\" (UID: \"aa2668b1-bd01-4fce-93fd-d7fbf9917247\") " Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.002392 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "aa2668b1-bd01-4fce-93fd-d7fbf9917247" (UID: "aa2668b1-bd01-4fce-93fd-d7fbf9917247"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.003932 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-logs" (OuterVolumeSpecName: "logs") pod "aa2668b1-bd01-4fce-93fd-d7fbf9917247" (UID: "aa2668b1-bd01-4fce-93fd-d7fbf9917247"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.012040 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa2668b1-bd01-4fce-93fd-d7fbf9917247-kube-api-access-lfp5w" (OuterVolumeSpecName: "kube-api-access-lfp5w") pod "aa2668b1-bd01-4fce-93fd-d7fbf9917247" (UID: "aa2668b1-bd01-4fce-93fd-d7fbf9917247"). InnerVolumeSpecName "kube-api-access-lfp5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.044233 4846 scope.go:117] "RemoveContainer" containerID="41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.044411 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-scripts" (OuterVolumeSpecName: "scripts") pod "aa2668b1-bd01-4fce-93fd-d7fbf9917247" (UID: "aa2668b1-bd01-4fce-93fd-d7fbf9917247"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.044470 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:51:50 crc kubenswrapper[4846]: E0202 13:51:50.046488 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5\": container with ID starting with 41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5 not found: ID does not exist" containerID="41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.046535 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5"} err="failed to get container status \"41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5\": rpc error: code = NotFound desc = could not find container \"41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5\": container with ID starting with 41f382a2b1a422b8ae252293693fdbb6d1d6475aabe4e7948a215111edcfa4f5 not found: ID does not exist" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.046569 4846 scope.go:117] "RemoveContainer" containerID="4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20" Feb 02 13:51:50 crc kubenswrapper[4846]: E0202 13:51:50.047319 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20\": container with ID starting with 4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20 not found: ID does not exist" containerID="4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.047359 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20"} err="failed to get container status \"4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20\": rpc error: code = NotFound desc = could not find container \"4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20\": container with ID starting with 4f6819f76f24e4fe4a6cd964765bbcbb95e5f3d6b0a171d9d334217938b17a20 not found: ID does not exist" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.047381 4846 scope.go:117] "RemoveContainer" containerID="f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.054921 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa2668b1-bd01-4fce-93fd-d7fbf9917247" (UID: "aa2668b1-bd01-4fce-93fd-d7fbf9917247"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.065413 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.081232 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:51:50 crc kubenswrapper[4846]: E0202 13:51:50.082416 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" containerName="glance-httpd" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.082442 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" containerName="glance-httpd" Feb 02 13:51:50 crc kubenswrapper[4846]: E0202 13:51:50.082455 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c8c7249-5482-4258-a3a7-069aa4c087dd" containerName="glance-httpd" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.082774 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-config-data" (OuterVolumeSpecName: "config-data") pod "aa2668b1-bd01-4fce-93fd-d7fbf9917247" (UID: "aa2668b1-bd01-4fce-93fd-d7fbf9917247"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.084227 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c8c7249-5482-4258-a3a7-069aa4c087dd" containerName="glance-httpd" Feb 02 13:51:50 crc kubenswrapper[4846]: E0202 13:51:50.084401 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" containerName="glance-log" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.084418 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" containerName="glance-log" Feb 02 13:51:50 crc kubenswrapper[4846]: E0202 13:51:50.084438 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c8c7249-5482-4258-a3a7-069aa4c087dd" containerName="glance-log" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.084447 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c8c7249-5482-4258-a3a7-069aa4c087dd" containerName="glance-log" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.084817 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" containerName="glance-log" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.084857 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" containerName="glance-httpd" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.084870 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c8c7249-5482-4258-a3a7-069aa4c087dd" containerName="glance-httpd" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.084897 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c8c7249-5482-4258-a3a7-069aa4c087dd" containerName="glance-log" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.088651 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "aa2668b1-bd01-4fce-93fd-d7fbf9917247" (UID: "aa2668b1-bd01-4fce-93fd-d7fbf9917247"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.091792 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.094993 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.097247 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.100004 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.110636 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.110712 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfp5w\" (UniqueName: \"kubernetes.io/projected/aa2668b1-bd01-4fce-93fd-d7fbf9917247-kube-api-access-lfp5w\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.110728 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.110740 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.110750 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa2668b1-bd01-4fce-93fd-d7fbf9917247-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.110763 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.110774 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa2668b1-bd01-4fce-93fd-d7fbf9917247-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.214306 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-767r4\" (UniqueName: \"kubernetes.io/projected/98a97c37-c689-484f-9f21-c3bd2fc20d32-kube-api-access-767r4\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.214373 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-config-data\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.214596 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98a97c37-c689-484f-9f21-c3bd2fc20d32-logs\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.214776 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-scripts\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.214939 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98a97c37-c689-484f-9f21-c3bd2fc20d32-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.214971 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.215091 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.240016 4846 scope.go:117] "RemoveContainer" containerID="418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.315797 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.317229 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-767r4\" (UniqueName: \"kubernetes.io/projected/98a97c37-c689-484f-9f21-c3bd2fc20d32-kube-api-access-767r4\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.317603 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-config-data\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.318303 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98a97c37-c689-484f-9f21-c3bd2fc20d32-logs\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.318442 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-scripts\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.318562 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98a97c37-c689-484f-9f21-c3bd2fc20d32-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.318684 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.318836 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.320349 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98a97c37-c689-484f-9f21-c3bd2fc20d32-logs\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.325748 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/98a97c37-c689-484f-9f21-c3bd2fc20d32-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.332482 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.340863 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.344317 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-config-data\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.345612 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.347296 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.347774 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-767r4\" (UniqueName: \"kubernetes.io/projected/98a97c37-c689-484f-9f21-c3bd2fc20d32-kube-api-access-767r4\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.349151 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-scripts\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.349297 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.349447 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a97c37-c689-484f-9f21-c3bd2fc20d32-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"98a97c37-c689-484f-9f21-c3bd2fc20d32\") " pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.357071 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.361610 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.523019 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07db105d-693c-46f6-b19c-e2f470acb949-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.523455 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-scripts\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.523533 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07db105d-693c-46f6-b19c-e2f470acb949-logs\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.523558 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.523589 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.523668 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gblhn\" (UniqueName: \"kubernetes.io/projected/07db105d-693c-46f6-b19c-e2f470acb949-kube-api-access-gblhn\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.523772 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-config-data\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.525883 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.625931 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07db105d-693c-46f6-b19c-e2f470acb949-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.626030 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-scripts\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.626075 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07db105d-693c-46f6-b19c-e2f470acb949-logs\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.626098 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.626123 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.626153 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gblhn\" (UniqueName: \"kubernetes.io/projected/07db105d-693c-46f6-b19c-e2f470acb949-kube-api-access-gblhn\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.626261 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-config-data\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.627072 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/07db105d-693c-46f6-b19c-e2f470acb949-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.627206 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07db105d-693c-46f6-b19c-e2f470acb949-logs\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.632347 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-config-data\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.635516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-scripts\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.644183 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.645826 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gblhn\" (UniqueName: \"kubernetes.io/projected/07db105d-693c-46f6-b19c-e2f470acb949-kube-api-access-gblhn\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.650866 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07db105d-693c-46f6-b19c-e2f470acb949-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"07db105d-693c-46f6-b19c-e2f470acb949\") " pod="openstack/glance-default-internal-api-0" Feb 02 13:51:50 crc kubenswrapper[4846]: I0202 13:51:50.742206 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 02 13:51:51 crc kubenswrapper[4846]: I0202 13:51:51.436466 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c8c7249-5482-4258-a3a7-069aa4c087dd" path="/var/lib/kubelet/pods/6c8c7249-5482-4258-a3a7-069aa4c087dd/volumes" Feb 02 13:51:51 crc kubenswrapper[4846]: I0202 13:51:51.438723 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa2668b1-bd01-4fce-93fd-d7fbf9917247" path="/var/lib/kubelet/pods/aa2668b1-bd01-4fce-93fd-d7fbf9917247/volumes" Feb 02 13:51:55 crc kubenswrapper[4846]: I0202 13:51:55.222792 4846 scope.go:117] "RemoveContainer" containerID="f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013" Feb 02 13:51:55 crc kubenswrapper[4846]: E0202 13:51:55.223757 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013\": container with ID starting with f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013 not found: ID does not exist" containerID="f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013" Feb 02 13:51:55 crc kubenswrapper[4846]: I0202 13:51:55.223794 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013"} err="failed to get container status \"f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013\": rpc error: code = NotFound desc = could not find container \"f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013\": container with ID starting with f148fc311e92921d232e85c4534e74a3136c9cff251bce2ecf57b64e38558013 not found: ID does not exist" Feb 02 13:51:55 crc kubenswrapper[4846]: I0202 13:51:55.223820 4846 scope.go:117] "RemoveContainer" containerID="418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71" Feb 02 13:51:55 crc kubenswrapper[4846]: E0202 13:51:55.224205 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71\": container with ID starting with 418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71 not found: ID does not exist" containerID="418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71" Feb 02 13:51:55 crc kubenswrapper[4846]: I0202 13:51:55.224228 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71"} err="failed to get container status \"418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71\": rpc error: code = NotFound desc = could not find container \"418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71\": container with ID starting with 418b6ba8ae789f22485529fe400910371c15427f0330e90370bc92a133196c71 not found: ID does not exist" Feb 02 13:51:55 crc kubenswrapper[4846]: I0202 13:51:55.879401 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.065475 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d584cfd6-zscgj" event={"ID":"510040f2-b874-48d8-bfb1-871b8abf0199","Type":"ContainerStarted","Data":"1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658"} Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.065517 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d584cfd6-zscgj" event={"ID":"510040f2-b874-48d8-bfb1-871b8abf0199","Type":"ContainerStarted","Data":"bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572"} Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.069099 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f6dfcdc7-gb2k8" event={"ID":"14cb62be-e7b0-4905-9829-fa63566b377c","Type":"ContainerStarted","Data":"5bcf66edfbdda5e8efcfe55190c003ef831323e86fb8db879bdedd241b1a18d2"} Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.069144 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f6dfcdc7-gb2k8" event={"ID":"14cb62be-e7b0-4905-9829-fa63566b377c","Type":"ContainerStarted","Data":"222c8cb57929a2e28a680f136da4e6c6884d7ea3f8c78184d48dac2787309c98"} Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.069205 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79f6dfcdc7-gb2k8" podUID="14cb62be-e7b0-4905-9829-fa63566b377c" containerName="horizon-log" containerID="cri-o://222c8cb57929a2e28a680f136da4e6c6884d7ea3f8c78184d48dac2787309c98" gracePeriod=30 Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.069255 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79f6dfcdc7-gb2k8" podUID="14cb62be-e7b0-4905-9829-fa63566b377c" containerName="horizon" containerID="cri-o://5bcf66edfbdda5e8efcfe55190c003ef831323e86fb8db879bdedd241b1a18d2" gracePeriod=30 Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.073418 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766fbf8cc6-nf7ws" event={"ID":"3bada3e7-480c-45df-bc54-452587b69d52","Type":"ContainerStarted","Data":"5f9d52f18a37a63a4bc28d032601d1879bbf315c55d1fe5db1b5650199b27dce"} Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.073467 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766fbf8cc6-nf7ws" event={"ID":"3bada3e7-480c-45df-bc54-452587b69d52","Type":"ContainerStarted","Data":"ddfad6f8cdba946ad016e61bd42dc8962bec94675a6cdbbc07cc91eb2c9ebc9d"} Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.076643 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556b694cbc-pxk5w" event={"ID":"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c","Type":"ContainerStarted","Data":"a10ff26431ed9a74985cd9c143602a06b8cbfa8d5ca21e16a34421382b51859b"} Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.076686 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556b694cbc-pxk5w" event={"ID":"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c","Type":"ContainerStarted","Data":"a9fc91af8bdff38c9bfa3429a5b5e049c7bcc48cfb5f7309b874c5f112bf7cde"} Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.076772 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-556b694cbc-pxk5w" podUID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" containerName="horizon" containerID="cri-o://a10ff26431ed9a74985cd9c143602a06b8cbfa8d5ca21e16a34421382b51859b" gracePeriod=30 Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.076770 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-556b694cbc-pxk5w" podUID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" containerName="horizon-log" containerID="cri-o://a9fc91af8bdff38c9bfa3429a5b5e049c7bcc48cfb5f7309b874c5f112bf7cde" gracePeriod=30 Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.082719 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"07db105d-693c-46f6-b19c-e2f470acb949","Type":"ContainerStarted","Data":"279b7dae47e316829f43e9a3700a62bb92c18c9bd2baba8e9fe3630ae1b6ec61"} Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.087468 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-d584cfd6-zscgj" podStartSLOduration=2.433481445 podStartE2EDuration="9.087451458s" podCreationTimestamp="2026-02-02 13:51:47 +0000 UTC" firstStartedPulling="2026-02-02 13:51:48.687904979 +0000 UTC m=+6139.916491842" lastFinishedPulling="2026-02-02 13:51:55.341874982 +0000 UTC m=+6146.570461855" observedRunningTime="2026-02-02 13:51:56.084321432 +0000 UTC m=+6147.312908295" watchObservedRunningTime="2026-02-02 13:51:56.087451458 +0000 UTC m=+6147.316038321" Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.115350 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-766fbf8cc6-nf7ws" podStartSLOduration=2.217176742 podStartE2EDuration="9.115328145s" podCreationTimestamp="2026-02-02 13:51:47 +0000 UTC" firstStartedPulling="2026-02-02 13:51:48.451425846 +0000 UTC m=+6139.680012709" lastFinishedPulling="2026-02-02 13:51:55.349577249 +0000 UTC m=+6146.578164112" observedRunningTime="2026-02-02 13:51:56.105464586 +0000 UTC m=+6147.334051449" watchObservedRunningTime="2026-02-02 13:51:56.115328145 +0000 UTC m=+6147.343915008" Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.134792 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-556b694cbc-pxk5w" podStartSLOduration=2.390031218 podStartE2EDuration="11.134765127s" podCreationTimestamp="2026-02-02 13:51:45 +0000 UTC" firstStartedPulling="2026-02-02 13:51:46.696082696 +0000 UTC m=+6137.924669559" lastFinishedPulling="2026-02-02 13:51:55.440816605 +0000 UTC m=+6146.669403468" observedRunningTime="2026-02-02 13:51:56.124205751 +0000 UTC m=+6147.352792624" watchObservedRunningTime="2026-02-02 13:51:56.134765127 +0000 UTC m=+6147.363351990" Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.149307 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.155731 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79f6dfcdc7-gb2k8" podStartSLOduration=2.409895762 podStartE2EDuration="11.155706776s" podCreationTimestamp="2026-02-02 13:51:45 +0000 UTC" firstStartedPulling="2026-02-02 13:51:46.593669089 +0000 UTC m=+6137.822255952" lastFinishedPulling="2026-02-02 13:51:55.339480093 +0000 UTC m=+6146.568066966" observedRunningTime="2026-02-02 13:51:56.145114919 +0000 UTC m=+6147.373701792" watchObservedRunningTime="2026-02-02 13:51:56.155706776 +0000 UTC m=+6147.384293649" Feb 02 13:51:56 crc kubenswrapper[4846]: I0202 13:51:56.846851 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 02 13:51:56 crc kubenswrapper[4846]: W0202 13:51:56.867829 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98a97c37_c689_484f_9f21_c3bd2fc20d32.slice/crio-f9b847bd6fac93e4882a59584b43aa232ee9b2e43e05381deda9858e2e7be56f WatchSource:0}: Error finding container f9b847bd6fac93e4882a59584b43aa232ee9b2e43e05381deda9858e2e7be56f: Status 404 returned error can't find the container with id f9b847bd6fac93e4882a59584b43aa232ee9b2e43e05381deda9858e2e7be56f Feb 02 13:51:57 crc kubenswrapper[4846]: I0202 13:51:57.096518 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"98a97c37-c689-484f-9f21-c3bd2fc20d32","Type":"ContainerStarted","Data":"f9b847bd6fac93e4882a59584b43aa232ee9b2e43e05381deda9858e2e7be56f"} Feb 02 13:51:57 crc kubenswrapper[4846]: I0202 13:51:57.098616 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"07db105d-693c-46f6-b19c-e2f470acb949","Type":"ContainerStarted","Data":"4e2c705c80f10ff5ce63b924f4f09225bab7e0cdfb8a1efedb9d32a462b7f586"} Feb 02 13:51:57 crc kubenswrapper[4846]: I0202 13:51:57.098683 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"07db105d-693c-46f6-b19c-e2f470acb949","Type":"ContainerStarted","Data":"d6390a061df369abf2fddbd4f2e53d0d618186a3330d540bd633d8879064f6ea"} Feb 02 13:51:57 crc kubenswrapper[4846]: I0202 13:51:57.131010 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.130986081 podStartE2EDuration="7.130986081s" podCreationTimestamp="2026-02-02 13:51:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:51:57.123179621 +0000 UTC m=+6148.351766504" watchObservedRunningTime="2026-02-02 13:51:57.130986081 +0000 UTC m=+6148.359572954" Feb 02 13:51:57 crc kubenswrapper[4846]: I0202 13:51:57.926876 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:57 crc kubenswrapper[4846]: I0202 13:51:57.927230 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:51:58 crc kubenswrapper[4846]: I0202 13:51:58.019221 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:58 crc kubenswrapper[4846]: I0202 13:51:58.019340 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:51:58 crc kubenswrapper[4846]: I0202 13:51:58.116756 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"98a97c37-c689-484f-9f21-c3bd2fc20d32","Type":"ContainerStarted","Data":"815eade641462ef59377ec4d614bf22d753229adc2d7bc9147dbd7e3f7900326"} Feb 02 13:51:58 crc kubenswrapper[4846]: I0202 13:51:58.117192 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"98a97c37-c689-484f-9f21-c3bd2fc20d32","Type":"ContainerStarted","Data":"2213e81a9074c9b42102e8f4bedd305302ee252eae33bfc2e70535895d828e3d"} Feb 02 13:51:59 crc kubenswrapper[4846]: I0202 13:51:59.152087 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=9.152056853 podStartE2EDuration="9.152056853s" podCreationTimestamp="2026-02-02 13:51:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:51:59.146492738 +0000 UTC m=+6150.375079631" watchObservedRunningTime="2026-02-02 13:51:59.152056853 +0000 UTC m=+6150.380643756" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.479020 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.479352 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.479396 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.480106 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fcae4c79abc512d4d8a5d4e3219b87abc67dad0f182ce8895f815c1ebd096f68"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.480156 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://fcae4c79abc512d4d8a5d4e3219b87abc67dad0f182ce8895f815c1ebd096f68" gracePeriod=600 Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.526852 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.526921 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.568392 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.574372 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.743396 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.744073 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.777509 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 02 13:52:00 crc kubenswrapper[4846]: I0202 13:52:00.788583 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 02 13:52:01 crc kubenswrapper[4846]: I0202 13:52:01.149221 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="fcae4c79abc512d4d8a5d4e3219b87abc67dad0f182ce8895f815c1ebd096f68" exitCode=0 Feb 02 13:52:01 crc kubenswrapper[4846]: I0202 13:52:01.149312 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"fcae4c79abc512d4d8a5d4e3219b87abc67dad0f182ce8895f815c1ebd096f68"} Feb 02 13:52:01 crc kubenswrapper[4846]: I0202 13:52:01.149393 4846 scope.go:117] "RemoveContainer" containerID="e9d7c4a3498c89a2353f29bef9ee80f98161106390f93552ff5adaba6e391bd4" Feb 02 13:52:01 crc kubenswrapper[4846]: I0202 13:52:01.149645 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 02 13:52:01 crc kubenswrapper[4846]: I0202 13:52:01.149671 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 02 13:52:01 crc kubenswrapper[4846]: I0202 13:52:01.149683 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 02 13:52:01 crc kubenswrapper[4846]: I0202 13:52:01.149691 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 02 13:52:02 crc kubenswrapper[4846]: I0202 13:52:02.167407 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab"} Feb 02 13:52:03 crc kubenswrapper[4846]: I0202 13:52:03.559150 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 02 13:52:04 crc kubenswrapper[4846]: I0202 13:52:04.162314 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 02 13:52:04 crc kubenswrapper[4846]: I0202 13:52:04.328253 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 02 13:52:06 crc kubenswrapper[4846]: I0202 13:52:06.015388 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:52:06 crc kubenswrapper[4846]: I0202 13:52:06.408391 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 02 13:52:07 crc kubenswrapper[4846]: I0202 13:52:07.929905 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-766fbf8cc6-nf7ws" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.120:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.120:8443: connect: connection refused" Feb 02 13:52:08 crc kubenswrapper[4846]: I0202 13:52:08.021453 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-d584cfd6-zscgj" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.121:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.121:8443: connect: connection refused" Feb 02 13:52:08 crc kubenswrapper[4846]: I0202 13:52:08.040116 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-s4ngv"] Feb 02 13:52:08 crc kubenswrapper[4846]: I0202 13:52:08.052944 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-edfe-account-create-update-ks4pd"] Feb 02 13:52:08 crc kubenswrapper[4846]: I0202 13:52:08.062788 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-s4ngv"] Feb 02 13:52:08 crc kubenswrapper[4846]: I0202 13:52:08.077469 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-edfe-account-create-update-ks4pd"] Feb 02 13:52:09 crc kubenswrapper[4846]: I0202 13:52:09.438958 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2726ef7-8ba0-4415-86e0-cf918301446a" path="/var/lib/kubelet/pods/d2726ef7-8ba0-4415-86e0-cf918301446a/volumes" Feb 02 13:52:09 crc kubenswrapper[4846]: I0202 13:52:09.440264 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8175f9f-b434-44ff-b770-252b30d2cf73" path="/var/lib/kubelet/pods/d8175f9f-b434-44ff-b770-252b30d2cf73/volumes" Feb 02 13:52:18 crc kubenswrapper[4846]: I0202 13:52:18.056900 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4glkm"] Feb 02 13:52:18 crc kubenswrapper[4846]: I0202 13:52:18.067250 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4glkm"] Feb 02 13:52:19 crc kubenswrapper[4846]: I0202 13:52:19.467195 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77c19163-c5ff-45f2-8410-87c0eee441a0" path="/var/lib/kubelet/pods/77c19163-c5ff-45f2-8410-87c0eee441a0/volumes" Feb 02 13:52:19 crc kubenswrapper[4846]: I0202 13:52:19.868170 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:52:19 crc kubenswrapper[4846]: I0202 13:52:19.869576 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:52:21 crc kubenswrapper[4846]: I0202 13:52:21.745251 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:52:21 crc kubenswrapper[4846]: I0202 13:52:21.888549 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:52:21 crc kubenswrapper[4846]: I0202 13:52:21.977267 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-766fbf8cc6-nf7ws"] Feb 02 13:52:22 crc kubenswrapper[4846]: I0202 13:52:22.385439 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-766fbf8cc6-nf7ws" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon-log" containerID="cri-o://ddfad6f8cdba946ad016e61bd42dc8962bec94675a6cdbbc07cc91eb2c9ebc9d" gracePeriod=30 Feb 02 13:52:22 crc kubenswrapper[4846]: I0202 13:52:22.385599 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-766fbf8cc6-nf7ws" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon" containerID="cri-o://5f9d52f18a37a63a4bc28d032601d1879bbf315c55d1fe5db1b5650199b27dce" gracePeriod=30 Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.449841 4846 generic.go:334] "Generic (PLEG): container finished" podID="14cb62be-e7b0-4905-9829-fa63566b377c" containerID="5bcf66edfbdda5e8efcfe55190c003ef831323e86fb8db879bdedd241b1a18d2" exitCode=137 Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.450377 4846 generic.go:334] "Generic (PLEG): container finished" podID="14cb62be-e7b0-4905-9829-fa63566b377c" containerID="222c8cb57929a2e28a680f136da4e6c6884d7ea3f8c78184d48dac2787309c98" exitCode=137 Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.450416 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f6dfcdc7-gb2k8" event={"ID":"14cb62be-e7b0-4905-9829-fa63566b377c","Type":"ContainerDied","Data":"5bcf66edfbdda5e8efcfe55190c003ef831323e86fb8db879bdedd241b1a18d2"} Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.450439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f6dfcdc7-gb2k8" event={"ID":"14cb62be-e7b0-4905-9829-fa63566b377c","Type":"ContainerDied","Data":"222c8cb57929a2e28a680f136da4e6c6884d7ea3f8c78184d48dac2787309c98"} Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.451894 4846 generic.go:334] "Generic (PLEG): container finished" podID="3bada3e7-480c-45df-bc54-452587b69d52" containerID="5f9d52f18a37a63a4bc28d032601d1879bbf315c55d1fe5db1b5650199b27dce" exitCode=0 Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.451923 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766fbf8cc6-nf7ws" event={"ID":"3bada3e7-480c-45df-bc54-452587b69d52","Type":"ContainerDied","Data":"5f9d52f18a37a63a4bc28d032601d1879bbf315c55d1fe5db1b5650199b27dce"} Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.453416 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" containerID="a10ff26431ed9a74985cd9c143602a06b8cbfa8d5ca21e16a34421382b51859b" exitCode=137 Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.453437 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" containerID="a9fc91af8bdff38c9bfa3429a5b5e049c7bcc48cfb5f7309b874c5f112bf7cde" exitCode=137 Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.453455 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556b694cbc-pxk5w" event={"ID":"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c","Type":"ContainerDied","Data":"a10ff26431ed9a74985cd9c143602a06b8cbfa8d5ca21e16a34421382b51859b"} Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.453475 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556b694cbc-pxk5w" event={"ID":"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c","Type":"ContainerDied","Data":"a9fc91af8bdff38c9bfa3429a5b5e049c7bcc48cfb5f7309b874c5f112bf7cde"} Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.532449 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.538121 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.641981 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjrv7\" (UniqueName: \"kubernetes.io/projected/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-kube-api-access-fjrv7\") pod \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.642094 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-scripts\") pod \"14cb62be-e7b0-4905-9829-fa63566b377c\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.642159 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-horizon-secret-key\") pod \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.642185 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-logs\") pod \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.642202 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-config-data\") pod \"14cb62be-e7b0-4905-9829-fa63566b377c\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.642275 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-scripts\") pod \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.642293 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s5ld\" (UniqueName: \"kubernetes.io/projected/14cb62be-e7b0-4905-9829-fa63566b377c-kube-api-access-9s5ld\") pod \"14cb62be-e7b0-4905-9829-fa63566b377c\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.642311 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/14cb62be-e7b0-4905-9829-fa63566b377c-horizon-secret-key\") pod \"14cb62be-e7b0-4905-9829-fa63566b377c\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.642345 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14cb62be-e7b0-4905-9829-fa63566b377c-logs\") pod \"14cb62be-e7b0-4905-9829-fa63566b377c\" (UID: \"14cb62be-e7b0-4905-9829-fa63566b377c\") " Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.642378 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-config-data\") pod \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\" (UID: \"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c\") " Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.642847 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-logs" (OuterVolumeSpecName: "logs") pod "f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" (UID: "f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.643214 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14cb62be-e7b0-4905-9829-fa63566b377c-logs" (OuterVolumeSpecName: "logs") pod "14cb62be-e7b0-4905-9829-fa63566b377c" (UID: "14cb62be-e7b0-4905-9829-fa63566b377c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.650920 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-kube-api-access-fjrv7" (OuterVolumeSpecName: "kube-api-access-fjrv7") pod "f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" (UID: "f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c"). InnerVolumeSpecName "kube-api-access-fjrv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.651731 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14cb62be-e7b0-4905-9829-fa63566b377c-kube-api-access-9s5ld" (OuterVolumeSpecName: "kube-api-access-9s5ld") pod "14cb62be-e7b0-4905-9829-fa63566b377c" (UID: "14cb62be-e7b0-4905-9829-fa63566b377c"). InnerVolumeSpecName "kube-api-access-9s5ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.653430 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" (UID: "f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.653569 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14cb62be-e7b0-4905-9829-fa63566b377c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "14cb62be-e7b0-4905-9829-fa63566b377c" (UID: "14cb62be-e7b0-4905-9829-fa63566b377c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.669100 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-config-data" (OuterVolumeSpecName: "config-data") pod "f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" (UID: "f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.670553 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-scripts" (OuterVolumeSpecName: "scripts") pod "f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" (UID: "f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.677575 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-config-data" (OuterVolumeSpecName: "config-data") pod "14cb62be-e7b0-4905-9829-fa63566b377c" (UID: "14cb62be-e7b0-4905-9829-fa63566b377c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.679121 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-scripts" (OuterVolumeSpecName: "scripts") pod "14cb62be-e7b0-4905-9829-fa63566b377c" (UID: "14cb62be-e7b0-4905-9829-fa63566b377c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.745073 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.745113 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.745129 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.745147 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s5ld\" (UniqueName: \"kubernetes.io/projected/14cb62be-e7b0-4905-9829-fa63566b377c-kube-api-access-9s5ld\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.745165 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.745179 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/14cb62be-e7b0-4905-9829-fa63566b377c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.745194 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14cb62be-e7b0-4905-9829-fa63566b377c-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.745209 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.745223 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjrv7\" (UniqueName: \"kubernetes.io/projected/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c-kube-api-access-fjrv7\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:26 crc kubenswrapper[4846]: I0202 13:52:26.745267 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14cb62be-e7b0-4905-9829-fa63566b377c-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.472508 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-556b694cbc-pxk5w" Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.474010 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-556b694cbc-pxk5w" event={"ID":"f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c","Type":"ContainerDied","Data":"89e7b78472b01d7c1c132b9a93332668485f128e7cab2d3e233863275003b66a"} Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.474219 4846 scope.go:117] "RemoveContainer" containerID="a10ff26431ed9a74985cd9c143602a06b8cbfa8d5ca21e16a34421382b51859b" Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.481112 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79f6dfcdc7-gb2k8" event={"ID":"14cb62be-e7b0-4905-9829-fa63566b377c","Type":"ContainerDied","Data":"a3ad2915cd67b8340325ad2f1a70e9846a486d9cf027f6d7416859ef6b7a121b"} Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.481225 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79f6dfcdc7-gb2k8" Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.510630 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-556b694cbc-pxk5w"] Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.527009 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-556b694cbc-pxk5w"] Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.536997 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79f6dfcdc7-gb2k8"] Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.545653 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79f6dfcdc7-gb2k8"] Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.709598 4846 scope.go:117] "RemoveContainer" containerID="a9fc91af8bdff38c9bfa3429a5b5e049c7bcc48cfb5f7309b874c5f112bf7cde" Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.738084 4846 scope.go:117] "RemoveContainer" containerID="5bcf66edfbdda5e8efcfe55190c003ef831323e86fb8db879bdedd241b1a18d2" Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.927953 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-766fbf8cc6-nf7ws" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.120:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.120:8443: connect: connection refused" Feb 02 13:52:27 crc kubenswrapper[4846]: I0202 13:52:27.977251 4846 scope.go:117] "RemoveContainer" containerID="222c8cb57929a2e28a680f136da4e6c6884d7ea3f8c78184d48dac2787309c98" Feb 02 13:52:29 crc kubenswrapper[4846]: I0202 13:52:29.440351 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14cb62be-e7b0-4905-9829-fa63566b377c" path="/var/lib/kubelet/pods/14cb62be-e7b0-4905-9829-fa63566b377c/volumes" Feb 02 13:52:29 crc kubenswrapper[4846]: I0202 13:52:29.441592 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" path="/var/lib/kubelet/pods/f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c/volumes" Feb 02 13:52:37 crc kubenswrapper[4846]: I0202 13:52:37.927896 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-766fbf8cc6-nf7ws" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.120:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.120:8443: connect: connection refused" Feb 02 13:52:39 crc kubenswrapper[4846]: I0202 13:52:39.940718 4846 scope.go:117] "RemoveContainer" containerID="6156327aecadf498e51f3ef6d8243ebb8c487db008b029f9fb2954bb8cacaaa2" Feb 02 13:52:39 crc kubenswrapper[4846]: I0202 13:52:39.975177 4846 scope.go:117] "RemoveContainer" containerID="f338a53a8f99bcc13f977c2a11df73ae5ad23888cbca856da2180ea35bcd2fe7" Feb 02 13:52:40 crc kubenswrapper[4846]: I0202 13:52:40.043827 4846 scope.go:117] "RemoveContainer" containerID="9ae519a9b309c537001888466260ad26927cfdf3c7b19f69c4f9ea422bff5265" Feb 02 13:52:40 crc kubenswrapper[4846]: I0202 13:52:40.092003 4846 scope.go:117] "RemoveContainer" containerID="66d119cc3db1736125c3cf810a2c411382e995cc2ec2648c44eb775d460acd9e" Feb 02 13:52:47 crc kubenswrapper[4846]: I0202 13:52:47.927723 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-766fbf8cc6-nf7ws" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.120:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.120:8443: connect: connection refused" Feb 02 13:52:47 crc kubenswrapper[4846]: I0202 13:52:47.928524 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.724950 4846 generic.go:334] "Generic (PLEG): container finished" podID="3bada3e7-480c-45df-bc54-452587b69d52" containerID="ddfad6f8cdba946ad016e61bd42dc8962bec94675a6cdbbc07cc91eb2c9ebc9d" exitCode=137 Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.725029 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766fbf8cc6-nf7ws" event={"ID":"3bada3e7-480c-45df-bc54-452587b69d52","Type":"ContainerDied","Data":"ddfad6f8cdba946ad016e61bd42dc8962bec94675a6cdbbc07cc91eb2c9ebc9d"} Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.725389 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-766fbf8cc6-nf7ws" event={"ID":"3bada3e7-480c-45df-bc54-452587b69d52","Type":"ContainerDied","Data":"6f56143bf78c066b2871c244f7be0b52a4bd2e55df089d7f4fc2d15e085a9ca2"} Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.725404 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f56143bf78c066b2871c244f7be0b52a4bd2e55df089d7f4fc2d15e085a9ca2" Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.804075 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.935097 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bada3e7-480c-45df-bc54-452587b69d52-logs\") pod \"3bada3e7-480c-45df-bc54-452587b69d52\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.935162 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-tls-certs\") pod \"3bada3e7-480c-45df-bc54-452587b69d52\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.935233 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-config-data\") pod \"3bada3e7-480c-45df-bc54-452587b69d52\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.936154 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8z28\" (UniqueName: \"kubernetes.io/projected/3bada3e7-480c-45df-bc54-452587b69d52-kube-api-access-j8z28\") pod \"3bada3e7-480c-45df-bc54-452587b69d52\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.935644 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bada3e7-480c-45df-bc54-452587b69d52-logs" (OuterVolumeSpecName: "logs") pod "3bada3e7-480c-45df-bc54-452587b69d52" (UID: "3bada3e7-480c-45df-bc54-452587b69d52"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.936339 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-combined-ca-bundle\") pod \"3bada3e7-480c-45df-bc54-452587b69d52\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.936416 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-scripts\") pod \"3bada3e7-480c-45df-bc54-452587b69d52\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.936475 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-secret-key\") pod \"3bada3e7-480c-45df-bc54-452587b69d52\" (UID: \"3bada3e7-480c-45df-bc54-452587b69d52\") " Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.937219 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bada3e7-480c-45df-bc54-452587b69d52-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.940929 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3bada3e7-480c-45df-bc54-452587b69d52" (UID: "3bada3e7-480c-45df-bc54-452587b69d52"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.941166 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bada3e7-480c-45df-bc54-452587b69d52-kube-api-access-j8z28" (OuterVolumeSpecName: "kube-api-access-j8z28") pod "3bada3e7-480c-45df-bc54-452587b69d52" (UID: "3bada3e7-480c-45df-bc54-452587b69d52"). InnerVolumeSpecName "kube-api-access-j8z28". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.961994 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-config-data" (OuterVolumeSpecName: "config-data") pod "3bada3e7-480c-45df-bc54-452587b69d52" (UID: "3bada3e7-480c-45df-bc54-452587b69d52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.973796 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bada3e7-480c-45df-bc54-452587b69d52" (UID: "3bada3e7-480c-45df-bc54-452587b69d52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:52:52 crc kubenswrapper[4846]: I0202 13:52:52.980603 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-scripts" (OuterVolumeSpecName: "scripts") pod "3bada3e7-480c-45df-bc54-452587b69d52" (UID: "3bada3e7-480c-45df-bc54-452587b69d52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:52:53 crc kubenswrapper[4846]: I0202 13:52:53.008512 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "3bada3e7-480c-45df-bc54-452587b69d52" (UID: "3bada3e7-480c-45df-bc54-452587b69d52"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:52:53 crc kubenswrapper[4846]: I0202 13:52:53.038821 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:53 crc kubenswrapper[4846]: I0202 13:52:53.038856 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:53 crc kubenswrapper[4846]: I0202 13:52:53.038867 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8z28\" (UniqueName: \"kubernetes.io/projected/3bada3e7-480c-45df-bc54-452587b69d52-kube-api-access-j8z28\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:53 crc kubenswrapper[4846]: I0202 13:52:53.038877 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:53 crc kubenswrapper[4846]: I0202 13:52:53.038887 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3bada3e7-480c-45df-bc54-452587b69d52-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:53 crc kubenswrapper[4846]: I0202 13:52:53.038897 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3bada3e7-480c-45df-bc54-452587b69d52-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 02 13:52:53 crc kubenswrapper[4846]: I0202 13:52:53.734977 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-766fbf8cc6-nf7ws" Feb 02 13:52:53 crc kubenswrapper[4846]: I0202 13:52:53.775768 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-766fbf8cc6-nf7ws"] Feb 02 13:52:53 crc kubenswrapper[4846]: I0202 13:52:53.785976 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-766fbf8cc6-nf7ws"] Feb 02 13:52:55 crc kubenswrapper[4846]: I0202 13:52:55.442790 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bada3e7-480c-45df-bc54-452587b69d52" path="/var/lib/kubelet/pods/3bada3e7-480c-45df-bc54-452587b69d52/volumes" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.058946 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5c49958db-lgw6w"] Feb 02 13:53:03 crc kubenswrapper[4846]: E0202 13:53:03.060033 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon-log" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060053 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon-log" Feb 02 13:53:03 crc kubenswrapper[4846]: E0202 13:53:03.060076 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" containerName="horizon-log" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060084 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" containerName="horizon-log" Feb 02 13:53:03 crc kubenswrapper[4846]: E0202 13:53:03.060103 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" containerName="horizon" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060110 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" containerName="horizon" Feb 02 13:53:03 crc kubenswrapper[4846]: E0202 13:53:03.060139 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060146 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon" Feb 02 13:53:03 crc kubenswrapper[4846]: E0202 13:53:03.060166 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14cb62be-e7b0-4905-9829-fa63566b377c" containerName="horizon-log" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060173 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="14cb62be-e7b0-4905-9829-fa63566b377c" containerName="horizon-log" Feb 02 13:53:03 crc kubenswrapper[4846]: E0202 13:53:03.060186 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14cb62be-e7b0-4905-9829-fa63566b377c" containerName="horizon" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060194 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="14cb62be-e7b0-4905-9829-fa63566b377c" containerName="horizon" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060394 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon-log" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060411 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bada3e7-480c-45df-bc54-452587b69d52" containerName="horizon" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060420 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="14cb62be-e7b0-4905-9829-fa63566b377c" containerName="horizon-log" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060441 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="14cb62be-e7b0-4905-9829-fa63566b377c" containerName="horizon" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060455 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" containerName="horizon-log" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.060468 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4fd18b4-4e0f-43a9-b70a-7fdb5ba6b49c" containerName="horizon" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.061520 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.088820 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c49958db-lgw6w"] Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.149163 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf96j\" (UniqueName: \"kubernetes.io/projected/46333d82-5d6b-4bae-b1ad-7a30e6a88490-kube-api-access-lf96j\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.149215 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/46333d82-5d6b-4bae-b1ad-7a30e6a88490-horizon-tls-certs\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.149286 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46333d82-5d6b-4bae-b1ad-7a30e6a88490-scripts\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.149316 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46333d82-5d6b-4bae-b1ad-7a30e6a88490-config-data\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.149444 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46333d82-5d6b-4bae-b1ad-7a30e6a88490-logs\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.149639 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46333d82-5d6b-4bae-b1ad-7a30e6a88490-horizon-secret-key\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.149701 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46333d82-5d6b-4bae-b1ad-7a30e6a88490-combined-ca-bundle\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.258491 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46333d82-5d6b-4bae-b1ad-7a30e6a88490-config-data\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.258563 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46333d82-5d6b-4bae-b1ad-7a30e6a88490-logs\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.258695 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46333d82-5d6b-4bae-b1ad-7a30e6a88490-horizon-secret-key\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.258739 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46333d82-5d6b-4bae-b1ad-7a30e6a88490-combined-ca-bundle\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.258871 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf96j\" (UniqueName: \"kubernetes.io/projected/46333d82-5d6b-4bae-b1ad-7a30e6a88490-kube-api-access-lf96j\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.258903 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/46333d82-5d6b-4bae-b1ad-7a30e6a88490-horizon-tls-certs\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.259084 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46333d82-5d6b-4bae-b1ad-7a30e6a88490-scripts\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.259583 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46333d82-5d6b-4bae-b1ad-7a30e6a88490-scripts\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.259839 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46333d82-5d6b-4bae-b1ad-7a30e6a88490-logs\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.260250 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/46333d82-5d6b-4bae-b1ad-7a30e6a88490-config-data\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.264888 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46333d82-5d6b-4bae-b1ad-7a30e6a88490-combined-ca-bundle\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.265047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/46333d82-5d6b-4bae-b1ad-7a30e6a88490-horizon-tls-certs\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.267967 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/46333d82-5d6b-4bae-b1ad-7a30e6a88490-horizon-secret-key\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.291906 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf96j\" (UniqueName: \"kubernetes.io/projected/46333d82-5d6b-4bae-b1ad-7a30e6a88490-kube-api-access-lf96j\") pod \"horizon-5c49958db-lgw6w\" (UID: \"46333d82-5d6b-4bae-b1ad-7a30e6a88490\") " pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.387839 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:03 crc kubenswrapper[4846]: I0202 13:53:03.884576 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c49958db-lgw6w"] Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.792053 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-tms4z"] Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.794154 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-tms4z" Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.804722 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-tms4z"] Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.843555 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c49958db-lgw6w" event={"ID":"46333d82-5d6b-4bae-b1ad-7a30e6a88490","Type":"ContainerStarted","Data":"808382b60730b5d827054ebde85cf187bde674676fc1f567ec2e39d603fd76a5"} Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.843853 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c49958db-lgw6w" event={"ID":"46333d82-5d6b-4bae-b1ad-7a30e6a88490","Type":"ContainerStarted","Data":"d4bf1ed25ecff5bcbf216bd154db7015231b6e049b6439df6dca0a9017407f4a"} Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.843943 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c49958db-lgw6w" event={"ID":"46333d82-5d6b-4bae-b1ad-7a30e6a88490","Type":"ContainerStarted","Data":"f69788d56229e37e8dd90e3d7ddd419853c9ce0b1d6b387af2e73a618646db7b"} Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.886544 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-fb24-account-create-update-6hv7c"] Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.888343 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fb24-account-create-update-6hv7c" Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.891865 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.894023 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5c49958db-lgw6w" podStartSLOduration=1.893977285 podStartE2EDuration="1.893977285s" podCreationTimestamp="2026-02-02 13:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:53:04.883131271 +0000 UTC m=+6216.111718144" watchObservedRunningTime="2026-02-02 13:53:04.893977285 +0000 UTC m=+6216.122564148" Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.897496 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86ba6367-a1ef-4568-96e2-cddbcab98c68-operator-scripts\") pod \"heat-db-create-tms4z\" (UID: \"86ba6367-a1ef-4568-96e2-cddbcab98c68\") " pod="openstack/heat-db-create-tms4z" Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.898226 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbpds\" (UniqueName: \"kubernetes.io/projected/86ba6367-a1ef-4568-96e2-cddbcab98c68-kube-api-access-mbpds\") pod \"heat-db-create-tms4z\" (UID: \"86ba6367-a1ef-4568-96e2-cddbcab98c68\") " pod="openstack/heat-db-create-tms4z" Feb 02 13:53:04 crc kubenswrapper[4846]: I0202 13:53:04.910810 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-fb24-account-create-update-6hv7c"] Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.000841 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbpds\" (UniqueName: \"kubernetes.io/projected/86ba6367-a1ef-4568-96e2-cddbcab98c68-kube-api-access-mbpds\") pod \"heat-db-create-tms4z\" (UID: \"86ba6367-a1ef-4568-96e2-cddbcab98c68\") " pod="openstack/heat-db-create-tms4z" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.000911 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ea28d1f-bc99-4cd3-9f37-922870704b52-operator-scripts\") pod \"heat-fb24-account-create-update-6hv7c\" (UID: \"5ea28d1f-bc99-4cd3-9f37-922870704b52\") " pod="openstack/heat-fb24-account-create-update-6hv7c" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.001231 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkn8h\" (UniqueName: \"kubernetes.io/projected/5ea28d1f-bc99-4cd3-9f37-922870704b52-kube-api-access-rkn8h\") pod \"heat-fb24-account-create-update-6hv7c\" (UID: \"5ea28d1f-bc99-4cd3-9f37-922870704b52\") " pod="openstack/heat-fb24-account-create-update-6hv7c" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.001276 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86ba6367-a1ef-4568-96e2-cddbcab98c68-operator-scripts\") pod \"heat-db-create-tms4z\" (UID: \"86ba6367-a1ef-4568-96e2-cddbcab98c68\") " pod="openstack/heat-db-create-tms4z" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.002106 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86ba6367-a1ef-4568-96e2-cddbcab98c68-operator-scripts\") pod \"heat-db-create-tms4z\" (UID: \"86ba6367-a1ef-4568-96e2-cddbcab98c68\") " pod="openstack/heat-db-create-tms4z" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.020215 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbpds\" (UniqueName: \"kubernetes.io/projected/86ba6367-a1ef-4568-96e2-cddbcab98c68-kube-api-access-mbpds\") pod \"heat-db-create-tms4z\" (UID: \"86ba6367-a1ef-4568-96e2-cddbcab98c68\") " pod="openstack/heat-db-create-tms4z" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.102958 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkn8h\" (UniqueName: \"kubernetes.io/projected/5ea28d1f-bc99-4cd3-9f37-922870704b52-kube-api-access-rkn8h\") pod \"heat-fb24-account-create-update-6hv7c\" (UID: \"5ea28d1f-bc99-4cd3-9f37-922870704b52\") " pod="openstack/heat-fb24-account-create-update-6hv7c" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.103303 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ea28d1f-bc99-4cd3-9f37-922870704b52-operator-scripts\") pod \"heat-fb24-account-create-update-6hv7c\" (UID: \"5ea28d1f-bc99-4cd3-9f37-922870704b52\") " pod="openstack/heat-fb24-account-create-update-6hv7c" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.104019 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ea28d1f-bc99-4cd3-9f37-922870704b52-operator-scripts\") pod \"heat-fb24-account-create-update-6hv7c\" (UID: \"5ea28d1f-bc99-4cd3-9f37-922870704b52\") " pod="openstack/heat-fb24-account-create-update-6hv7c" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.119892 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkn8h\" (UniqueName: \"kubernetes.io/projected/5ea28d1f-bc99-4cd3-9f37-922870704b52-kube-api-access-rkn8h\") pod \"heat-fb24-account-create-update-6hv7c\" (UID: \"5ea28d1f-bc99-4cd3-9f37-922870704b52\") " pod="openstack/heat-fb24-account-create-update-6hv7c" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.122980 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-tms4z" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.207663 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fb24-account-create-update-6hv7c" Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.622804 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-tms4z"] Feb 02 13:53:05 crc kubenswrapper[4846]: W0202 13:53:05.623689 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86ba6367_a1ef_4568_96e2_cddbcab98c68.slice/crio-404d2f5dca870549deb7843c9c8a2a2327a519ffd049c206d0a13a64ec4bc62e WatchSource:0}: Error finding container 404d2f5dca870549deb7843c9c8a2a2327a519ffd049c206d0a13a64ec4bc62e: Status 404 returned error can't find the container with id 404d2f5dca870549deb7843c9c8a2a2327a519ffd049c206d0a13a64ec4bc62e Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.752566 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-fb24-account-create-update-6hv7c"] Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.861683 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fb24-account-create-update-6hv7c" event={"ID":"5ea28d1f-bc99-4cd3-9f37-922870704b52","Type":"ContainerStarted","Data":"3201b7c5cfea8d454d46d624fa07a9c0f4146abe12d66ee8566a8cb875351d80"} Feb 02 13:53:05 crc kubenswrapper[4846]: I0202 13:53:05.863466 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-tms4z" event={"ID":"86ba6367-a1ef-4568-96e2-cddbcab98c68","Type":"ContainerStarted","Data":"404d2f5dca870549deb7843c9c8a2a2327a519ffd049c206d0a13a64ec4bc62e"} Feb 02 13:53:06 crc kubenswrapper[4846]: I0202 13:53:06.871838 4846 generic.go:334] "Generic (PLEG): container finished" podID="5ea28d1f-bc99-4cd3-9f37-922870704b52" containerID="7650dafd98d9fd70ffe724a2db03c28b9c0d9052eb70c0e3042f1828171adc1a" exitCode=0 Feb 02 13:53:06 crc kubenswrapper[4846]: I0202 13:53:06.871918 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fb24-account-create-update-6hv7c" event={"ID":"5ea28d1f-bc99-4cd3-9f37-922870704b52","Type":"ContainerDied","Data":"7650dafd98d9fd70ffe724a2db03c28b9c0d9052eb70c0e3042f1828171adc1a"} Feb 02 13:53:06 crc kubenswrapper[4846]: I0202 13:53:06.873951 4846 generic.go:334] "Generic (PLEG): container finished" podID="86ba6367-a1ef-4568-96e2-cddbcab98c68" containerID="a0a6c3dd28e50d762ec4e4bde884d757f3d597a1560d821b6ff6cb795d82dc0d" exitCode=0 Feb 02 13:53:06 crc kubenswrapper[4846]: I0202 13:53:06.873984 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-tms4z" event={"ID":"86ba6367-a1ef-4568-96e2-cddbcab98c68","Type":"ContainerDied","Data":"a0a6c3dd28e50d762ec4e4bde884d757f3d597a1560d821b6ff6cb795d82dc0d"} Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.341804 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fb24-account-create-update-6hv7c" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.348748 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-tms4z" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.491757 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ea28d1f-bc99-4cd3-9f37-922870704b52-operator-scripts\") pod \"5ea28d1f-bc99-4cd3-9f37-922870704b52\" (UID: \"5ea28d1f-bc99-4cd3-9f37-922870704b52\") " Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.491839 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86ba6367-a1ef-4568-96e2-cddbcab98c68-operator-scripts\") pod \"86ba6367-a1ef-4568-96e2-cddbcab98c68\" (UID: \"86ba6367-a1ef-4568-96e2-cddbcab98c68\") " Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.491949 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkn8h\" (UniqueName: \"kubernetes.io/projected/5ea28d1f-bc99-4cd3-9f37-922870704b52-kube-api-access-rkn8h\") pod \"5ea28d1f-bc99-4cd3-9f37-922870704b52\" (UID: \"5ea28d1f-bc99-4cd3-9f37-922870704b52\") " Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.492039 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbpds\" (UniqueName: \"kubernetes.io/projected/86ba6367-a1ef-4568-96e2-cddbcab98c68-kube-api-access-mbpds\") pod \"86ba6367-a1ef-4568-96e2-cddbcab98c68\" (UID: \"86ba6367-a1ef-4568-96e2-cddbcab98c68\") " Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.492573 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ea28d1f-bc99-4cd3-9f37-922870704b52-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ea28d1f-bc99-4cd3-9f37-922870704b52" (UID: "5ea28d1f-bc99-4cd3-9f37-922870704b52"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.493226 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86ba6367-a1ef-4568-96e2-cddbcab98c68-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "86ba6367-a1ef-4568-96e2-cddbcab98c68" (UID: "86ba6367-a1ef-4568-96e2-cddbcab98c68"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.497821 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ea28d1f-bc99-4cd3-9f37-922870704b52-kube-api-access-rkn8h" (OuterVolumeSpecName: "kube-api-access-rkn8h") pod "5ea28d1f-bc99-4cd3-9f37-922870704b52" (UID: "5ea28d1f-bc99-4cd3-9f37-922870704b52"). InnerVolumeSpecName "kube-api-access-rkn8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.501541 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86ba6367-a1ef-4568-96e2-cddbcab98c68-kube-api-access-mbpds" (OuterVolumeSpecName: "kube-api-access-mbpds") pod "86ba6367-a1ef-4568-96e2-cddbcab98c68" (UID: "86ba6367-a1ef-4568-96e2-cddbcab98c68"). InnerVolumeSpecName "kube-api-access-mbpds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.594626 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkn8h\" (UniqueName: \"kubernetes.io/projected/5ea28d1f-bc99-4cd3-9f37-922870704b52-kube-api-access-rkn8h\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.594678 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbpds\" (UniqueName: \"kubernetes.io/projected/86ba6367-a1ef-4568-96e2-cddbcab98c68-kube-api-access-mbpds\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.594690 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ea28d1f-bc99-4cd3-9f37-922870704b52-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.594700 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86ba6367-a1ef-4568-96e2-cddbcab98c68-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.892756 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-tms4z" event={"ID":"86ba6367-a1ef-4568-96e2-cddbcab98c68","Type":"ContainerDied","Data":"404d2f5dca870549deb7843c9c8a2a2327a519ffd049c206d0a13a64ec4bc62e"} Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.892994 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="404d2f5dca870549deb7843c9c8a2a2327a519ffd049c206d0a13a64ec4bc62e" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.892932 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-tms4z" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.894359 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-fb24-account-create-update-6hv7c" event={"ID":"5ea28d1f-bc99-4cd3-9f37-922870704b52","Type":"ContainerDied","Data":"3201b7c5cfea8d454d46d624fa07a9c0f4146abe12d66ee8566a8cb875351d80"} Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.894391 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3201b7c5cfea8d454d46d624fa07a9c0f4146abe12d66ee8566a8cb875351d80" Feb 02 13:53:08 crc kubenswrapper[4846]: I0202 13:53:08.894427 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-fb24-account-create-update-6hv7c" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.218669 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-g6986"] Feb 02 13:53:10 crc kubenswrapper[4846]: E0202 13:53:10.220591 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea28d1f-bc99-4cd3-9f37-922870704b52" containerName="mariadb-account-create-update" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.220624 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea28d1f-bc99-4cd3-9f37-922870704b52" containerName="mariadb-account-create-update" Feb 02 13:53:10 crc kubenswrapper[4846]: E0202 13:53:10.220692 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ba6367-a1ef-4568-96e2-cddbcab98c68" containerName="mariadb-database-create" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.220702 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ba6367-a1ef-4568-96e2-cddbcab98c68" containerName="mariadb-database-create" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.220963 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ea28d1f-bc99-4cd3-9f37-922870704b52" containerName="mariadb-account-create-update" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.220984 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="86ba6367-a1ef-4568-96e2-cddbcab98c68" containerName="mariadb-database-create" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.221949 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.231096 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.231164 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-hkvrg" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.245893 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-g6986"] Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.334250 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-combined-ca-bundle\") pod \"heat-db-sync-g6986\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.334869 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgb76\" (UniqueName: \"kubernetes.io/projected/337292e8-0928-4f18-86d6-e433f406af6b-kube-api-access-fgb76\") pod \"heat-db-sync-g6986\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.335100 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-config-data\") pod \"heat-db-sync-g6986\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.438019 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-combined-ca-bundle\") pod \"heat-db-sync-g6986\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.438163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgb76\" (UniqueName: \"kubernetes.io/projected/337292e8-0928-4f18-86d6-e433f406af6b-kube-api-access-fgb76\") pod \"heat-db-sync-g6986\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.438204 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-config-data\") pod \"heat-db-sync-g6986\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.455779 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-combined-ca-bundle\") pod \"heat-db-sync-g6986\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.456063 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-config-data\") pod \"heat-db-sync-g6986\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.460360 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgb76\" (UniqueName: \"kubernetes.io/projected/337292e8-0928-4f18-86d6-e433f406af6b-kube-api-access-fgb76\") pod \"heat-db-sync-g6986\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.550842 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-g6986" Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.958917 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-g6986"] Feb 02 13:53:10 crc kubenswrapper[4846]: I0202 13:53:10.960554 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 13:53:11 crc kubenswrapper[4846]: I0202 13:53:11.929496 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-g6986" event={"ID":"337292e8-0928-4f18-86d6-e433f406af6b","Type":"ContainerStarted","Data":"7ae24a7275644041817cf0f2d88a1320cb54cf03e140709fa8a6e2a3e546fe03"} Feb 02 13:53:13 crc kubenswrapper[4846]: I0202 13:53:13.387979 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:13 crc kubenswrapper[4846]: I0202 13:53:13.388553 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:18 crc kubenswrapper[4846]: I0202 13:53:18.998942 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-g6986" event={"ID":"337292e8-0928-4f18-86d6-e433f406af6b","Type":"ContainerStarted","Data":"bdc95a26941ce1d40c00fc2320b558f64a8062946be22fdd0e03417633000c42"} Feb 02 13:53:19 crc kubenswrapper[4846]: I0202 13:53:19.027429 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-g6986" podStartSLOduration=1.919897001 podStartE2EDuration="9.027404999s" podCreationTimestamp="2026-02-02 13:53:10 +0000 UTC" firstStartedPulling="2026-02-02 13:53:10.960323688 +0000 UTC m=+6222.188910561" lastFinishedPulling="2026-02-02 13:53:18.067831696 +0000 UTC m=+6229.296418559" observedRunningTime="2026-02-02 13:53:19.015835929 +0000 UTC m=+6230.244422802" watchObservedRunningTime="2026-02-02 13:53:19.027404999 +0000 UTC m=+6230.255991882" Feb 02 13:53:21 crc kubenswrapper[4846]: I0202 13:53:21.035201 4846 generic.go:334] "Generic (PLEG): container finished" podID="337292e8-0928-4f18-86d6-e433f406af6b" containerID="bdc95a26941ce1d40c00fc2320b558f64a8062946be22fdd0e03417633000c42" exitCode=0 Feb 02 13:53:21 crc kubenswrapper[4846]: I0202 13:53:21.035550 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-g6986" event={"ID":"337292e8-0928-4f18-86d6-e433f406af6b","Type":"ContainerDied","Data":"bdc95a26941ce1d40c00fc2320b558f64a8062946be22fdd0e03417633000c42"} Feb 02 13:53:22 crc kubenswrapper[4846]: I0202 13:53:22.444303 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-g6986" Feb 02 13:53:22 crc kubenswrapper[4846]: I0202 13:53:22.520670 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgb76\" (UniqueName: \"kubernetes.io/projected/337292e8-0928-4f18-86d6-e433f406af6b-kube-api-access-fgb76\") pod \"337292e8-0928-4f18-86d6-e433f406af6b\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " Feb 02 13:53:22 crc kubenswrapper[4846]: I0202 13:53:22.520854 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-config-data\") pod \"337292e8-0928-4f18-86d6-e433f406af6b\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " Feb 02 13:53:22 crc kubenswrapper[4846]: I0202 13:53:22.520889 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-combined-ca-bundle\") pod \"337292e8-0928-4f18-86d6-e433f406af6b\" (UID: \"337292e8-0928-4f18-86d6-e433f406af6b\") " Feb 02 13:53:22 crc kubenswrapper[4846]: I0202 13:53:22.533926 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/337292e8-0928-4f18-86d6-e433f406af6b-kube-api-access-fgb76" (OuterVolumeSpecName: "kube-api-access-fgb76") pod "337292e8-0928-4f18-86d6-e433f406af6b" (UID: "337292e8-0928-4f18-86d6-e433f406af6b"). InnerVolumeSpecName "kube-api-access-fgb76". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:53:22 crc kubenswrapper[4846]: I0202 13:53:22.547525 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "337292e8-0928-4f18-86d6-e433f406af6b" (UID: "337292e8-0928-4f18-86d6-e433f406af6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:22 crc kubenswrapper[4846]: I0202 13:53:22.623235 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgb76\" (UniqueName: \"kubernetes.io/projected/337292e8-0928-4f18-86d6-e433f406af6b-kube-api-access-fgb76\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:22 crc kubenswrapper[4846]: I0202 13:53:22.623284 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:22 crc kubenswrapper[4846]: I0202 13:53:22.635071 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-config-data" (OuterVolumeSpecName: "config-data") pod "337292e8-0928-4f18-86d6-e433f406af6b" (UID: "337292e8-0928-4f18-86d6-e433f406af6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:22 crc kubenswrapper[4846]: I0202 13:53:22.725508 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337292e8-0928-4f18-86d6-e433f406af6b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:23 crc kubenswrapper[4846]: I0202 13:53:23.052951 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-g6986" event={"ID":"337292e8-0928-4f18-86d6-e433f406af6b","Type":"ContainerDied","Data":"7ae24a7275644041817cf0f2d88a1320cb54cf03e140709fa8a6e2a3e546fe03"} Feb 02 13:53:23 crc kubenswrapper[4846]: I0202 13:53:23.053527 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ae24a7275644041817cf0f2d88a1320cb54cf03e140709fa8a6e2a3e546fe03" Feb 02 13:53:23 crc kubenswrapper[4846]: I0202 13:53:23.053011 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-g6986" Feb 02 13:53:23 crc kubenswrapper[4846]: I0202 13:53:23.390191 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c49958db-lgw6w" podUID="46333d82-5d6b-4bae-b1ad-7a30e6a88490" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.124:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.124:8443: connect: connection refused" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.163197 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-677998f9bd-zfdvm"] Feb 02 13:53:24 crc kubenswrapper[4846]: E0202 13:53:24.163982 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337292e8-0928-4f18-86d6-e433f406af6b" containerName="heat-db-sync" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.163998 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="337292e8-0928-4f18-86d6-e433f406af6b" containerName="heat-db-sync" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.164254 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="337292e8-0928-4f18-86d6-e433f406af6b" containerName="heat-db-sync" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.165043 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.167095 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-hkvrg" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.167497 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.168106 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.177897 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-677998f9bd-zfdvm"] Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.257445 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6vl6\" (UniqueName: \"kubernetes.io/projected/70b265c4-72ec-485b-a147-97e1a9ac0a1b-kube-api-access-w6vl6\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.257502 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data-custom\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.257527 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-combined-ca-bundle\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.257749 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.394945 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6vl6\" (UniqueName: \"kubernetes.io/projected/70b265c4-72ec-485b-a147-97e1a9ac0a1b-kube-api-access-w6vl6\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.395028 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data-custom\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.395058 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-combined-ca-bundle\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.395121 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.401197 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-56c8998cb4-v2r4d"] Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.410352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data-custom\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.410496 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-combined-ca-bundle\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.416875 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.436292 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.448500 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.465414 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6vl6\" (UniqueName: \"kubernetes.io/projected/70b265c4-72ec-485b-a147-97e1a9ac0a1b-kube-api-access-w6vl6\") pod \"heat-engine-677998f9bd-zfdvm\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.493347 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.497822 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpcrm\" (UniqueName: \"kubernetes.io/projected/55343bd7-9622-44c6-81b2-9f140da38a90-kube-api-access-cpcrm\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.498233 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-combined-ca-bundle\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.498352 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.498433 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data-custom\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.520459 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-56c8998cb4-v2r4d"] Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.541952 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-769b6db55f-vttw8"] Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.544140 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.546837 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.565739 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-769b6db55f-vttw8"] Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.601078 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.601156 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data-custom\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.601242 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data-custom\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.601280 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-combined-ca-bundle\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.601350 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.601460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpcrm\" (UniqueName: \"kubernetes.io/projected/55343bd7-9622-44c6-81b2-9f140da38a90-kube-api-access-cpcrm\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.601713 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf4kj\" (UniqueName: \"kubernetes.io/projected/1e35b053-4619-428e-82d2-46f4a12e475c-kube-api-access-qf4kj\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.601806 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-combined-ca-bundle\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.615205 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.615790 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data-custom\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.615924 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-combined-ca-bundle\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.623692 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpcrm\" (UniqueName: \"kubernetes.io/projected/55343bd7-9622-44c6-81b2-9f140da38a90-kube-api-access-cpcrm\") pod \"heat-cfnapi-56c8998cb4-v2r4d\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.703463 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf4kj\" (UniqueName: \"kubernetes.io/projected/1e35b053-4619-428e-82d2-46f4a12e475c-kube-api-access-qf4kj\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.704548 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data-custom\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.704669 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-combined-ca-bundle\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.704729 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.711796 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data-custom\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.724274 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.730364 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf4kj\" (UniqueName: \"kubernetes.io/projected/1e35b053-4619-428e-82d2-46f4a12e475c-kube-api-access-qf4kj\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.733191 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-combined-ca-bundle\") pod \"heat-api-769b6db55f-vttw8\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.790045 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:24 crc kubenswrapper[4846]: I0202 13:53:24.814063 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:25 crc kubenswrapper[4846]: I0202 13:53:25.127204 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-677998f9bd-zfdvm"] Feb 02 13:53:25 crc kubenswrapper[4846]: W0202 13:53:25.134267 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70b265c4_72ec_485b_a147_97e1a9ac0a1b.slice/crio-702733b699c3f65bd3b79776ce18f65048346f0835223c453039e58d9b6eb78d WatchSource:0}: Error finding container 702733b699c3f65bd3b79776ce18f65048346f0835223c453039e58d9b6eb78d: Status 404 returned error can't find the container with id 702733b699c3f65bd3b79776ce18f65048346f0835223c453039e58d9b6eb78d Feb 02 13:53:25 crc kubenswrapper[4846]: I0202 13:53:25.329161 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-56c8998cb4-v2r4d"] Feb 02 13:53:25 crc kubenswrapper[4846]: W0202 13:53:25.407394 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e35b053_4619_428e_82d2_46f4a12e475c.slice/crio-67c1f1bc89130ad735e40e7b9ca8600439f8eca70b19e32bfc9216922ef76766 WatchSource:0}: Error finding container 67c1f1bc89130ad735e40e7b9ca8600439f8eca70b19e32bfc9216922ef76766: Status 404 returned error can't find the container with id 67c1f1bc89130ad735e40e7b9ca8600439f8eca70b19e32bfc9216922ef76766 Feb 02 13:53:25 crc kubenswrapper[4846]: I0202 13:53:25.414104 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-769b6db55f-vttw8"] Feb 02 13:53:26 crc kubenswrapper[4846]: I0202 13:53:26.091207 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" event={"ID":"55343bd7-9622-44c6-81b2-9f140da38a90","Type":"ContainerStarted","Data":"0c5cb5857a6b32938e58c700e3b8dccbf930f277818807abf700b1d24a214deb"} Feb 02 13:53:26 crc kubenswrapper[4846]: I0202 13:53:26.094480 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-677998f9bd-zfdvm" event={"ID":"70b265c4-72ec-485b-a147-97e1a9ac0a1b","Type":"ContainerStarted","Data":"47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053"} Feb 02 13:53:26 crc kubenswrapper[4846]: I0202 13:53:26.094520 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-677998f9bd-zfdvm" event={"ID":"70b265c4-72ec-485b-a147-97e1a9ac0a1b","Type":"ContainerStarted","Data":"702733b699c3f65bd3b79776ce18f65048346f0835223c453039e58d9b6eb78d"} Feb 02 13:53:26 crc kubenswrapper[4846]: I0202 13:53:26.095881 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:26 crc kubenswrapper[4846]: I0202 13:53:26.097022 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-769b6db55f-vttw8" event={"ID":"1e35b053-4619-428e-82d2-46f4a12e475c","Type":"ContainerStarted","Data":"67c1f1bc89130ad735e40e7b9ca8600439f8eca70b19e32bfc9216922ef76766"} Feb 02 13:53:26 crc kubenswrapper[4846]: I0202 13:53:26.122699 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-677998f9bd-zfdvm" podStartSLOduration=2.12265932 podStartE2EDuration="2.12265932s" podCreationTimestamp="2026-02-02 13:53:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:53:26.110770731 +0000 UTC m=+6237.339357604" watchObservedRunningTime="2026-02-02 13:53:26.12265932 +0000 UTC m=+6237.351246183" Feb 02 13:53:28 crc kubenswrapper[4846]: I0202 13:53:28.145414 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" event={"ID":"55343bd7-9622-44c6-81b2-9f140da38a90","Type":"ContainerStarted","Data":"2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd"} Feb 02 13:53:28 crc kubenswrapper[4846]: I0202 13:53:28.146070 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:28 crc kubenswrapper[4846]: I0202 13:53:28.149109 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-769b6db55f-vttw8" event={"ID":"1e35b053-4619-428e-82d2-46f4a12e475c","Type":"ContainerStarted","Data":"cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c"} Feb 02 13:53:28 crc kubenswrapper[4846]: I0202 13:53:28.149872 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:28 crc kubenswrapper[4846]: I0202 13:53:28.171047 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" podStartSLOduration=2.198960725 podStartE2EDuration="4.171026386s" podCreationTimestamp="2026-02-02 13:53:24 +0000 UTC" firstStartedPulling="2026-02-02 13:53:25.327333086 +0000 UTC m=+6236.555919949" lastFinishedPulling="2026-02-02 13:53:27.299398747 +0000 UTC m=+6238.527985610" observedRunningTime="2026-02-02 13:53:28.168194727 +0000 UTC m=+6239.396781610" watchObservedRunningTime="2026-02-02 13:53:28.171026386 +0000 UTC m=+6239.399613249" Feb 02 13:53:28 crc kubenswrapper[4846]: I0202 13:53:28.204662 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-769b6db55f-vttw8" podStartSLOduration=2.314629653 podStartE2EDuration="4.204640082s" podCreationTimestamp="2026-02-02 13:53:24 +0000 UTC" firstStartedPulling="2026-02-02 13:53:25.409886231 +0000 UTC m=+6236.638473094" lastFinishedPulling="2026-02-02 13:53:27.29989666 +0000 UTC m=+6238.528483523" observedRunningTime="2026-02-02 13:53:28.193796338 +0000 UTC m=+6239.422383191" watchObservedRunningTime="2026-02-02 13:53:28.204640082 +0000 UTC m=+6239.433226945" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.619108 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-555b878bf4-mzv78"] Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.620742 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.634390 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-555b878bf4-mzv78"] Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.650162 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-869cdc4dcc-vmzv2"] Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.652068 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.687719 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-869cdc4dcc-vmzv2"] Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.708588 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6fd74dfb66-674x6"] Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.720994 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.729964 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6fd74dfb66-674x6"] Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.775661 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b20035b-bca2-4bfd-86c8-e53deda9da15-config-data-custom\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.776010 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj9wf\" (UniqueName: \"kubernetes.io/projected/39ecd283-215f-442b-9b13-39ddf4bb3be9-kube-api-access-pj9wf\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.776059 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b20035b-bca2-4bfd-86c8-e53deda9da15-config-data\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.776092 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-combined-ca-bundle\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.776265 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data-custom\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.776425 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.776548 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b20035b-bca2-4bfd-86c8-e53deda9da15-combined-ca-bundle\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.776765 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxqnk\" (UniqueName: \"kubernetes.io/projected/4b20035b-bca2-4bfd-86c8-e53deda9da15-kube-api-access-sxqnk\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878279 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-combined-ca-bundle\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878342 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data-custom\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878381 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878411 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878432 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8ztg\" (UniqueName: \"kubernetes.io/projected/434d073d-68ad-4a28-97f0-8aab974ff3d9-kube-api-access-g8ztg\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878455 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b20035b-bca2-4bfd-86c8-e53deda9da15-combined-ca-bundle\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878470 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data-custom\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878537 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxqnk\" (UniqueName: \"kubernetes.io/projected/4b20035b-bca2-4bfd-86c8-e53deda9da15-kube-api-access-sxqnk\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878585 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-combined-ca-bundle\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878615 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b20035b-bca2-4bfd-86c8-e53deda9da15-config-data-custom\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878677 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj9wf\" (UniqueName: \"kubernetes.io/projected/39ecd283-215f-442b-9b13-39ddf4bb3be9-kube-api-access-pj9wf\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.878723 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b20035b-bca2-4bfd-86c8-e53deda9da15-config-data\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.887368 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b20035b-bca2-4bfd-86c8-e53deda9da15-config-data\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.887389 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b20035b-bca2-4bfd-86c8-e53deda9da15-combined-ca-bundle\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.890515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b20035b-bca2-4bfd-86c8-e53deda9da15-config-data-custom\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.890670 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data-custom\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.891094 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.892443 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-combined-ca-bundle\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.909640 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxqnk\" (UniqueName: \"kubernetes.io/projected/4b20035b-bca2-4bfd-86c8-e53deda9da15-kube-api-access-sxqnk\") pod \"heat-engine-555b878bf4-mzv78\" (UID: \"4b20035b-bca2-4bfd-86c8-e53deda9da15\") " pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.927592 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj9wf\" (UniqueName: \"kubernetes.io/projected/39ecd283-215f-442b-9b13-39ddf4bb3be9-kube-api-access-pj9wf\") pod \"heat-api-869cdc4dcc-vmzv2\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.944514 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.980545 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.980595 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8ztg\" (UniqueName: \"kubernetes.io/projected/434d073d-68ad-4a28-97f0-8aab974ff3d9-kube-api-access-g8ztg\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.980631 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data-custom\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.980767 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-combined-ca-bundle\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.985677 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data-custom\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.985967 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:31 crc kubenswrapper[4846]: I0202 13:53:31.986019 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-combined-ca-bundle\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:32 crc kubenswrapper[4846]: I0202 13:53:32.001807 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8ztg\" (UniqueName: \"kubernetes.io/projected/434d073d-68ad-4a28-97f0-8aab974ff3d9-kube-api-access-g8ztg\") pod \"heat-cfnapi-6fd74dfb66-674x6\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:32 crc kubenswrapper[4846]: I0202 13:53:32.010031 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:32 crc kubenswrapper[4846]: I0202 13:53:32.042421 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:32 crc kubenswrapper[4846]: I0202 13:53:32.500486 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-555b878bf4-mzv78"] Feb 02 13:53:32 crc kubenswrapper[4846]: W0202 13:53:32.503019 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b20035b_bca2_4bfd_86c8_e53deda9da15.slice/crio-b67e19b772749999cb5a3552d80aace44bb51c201283ac7ae9e9422e2e20f9d8 WatchSource:0}: Error finding container b67e19b772749999cb5a3552d80aace44bb51c201283ac7ae9e9422e2e20f9d8: Status 404 returned error can't find the container with id b67e19b772749999cb5a3552d80aace44bb51c201283ac7ae9e9422e2e20f9d8 Feb 02 13:53:32 crc kubenswrapper[4846]: I0202 13:53:32.614060 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-869cdc4dcc-vmzv2"] Feb 02 13:53:32 crc kubenswrapper[4846]: W0202 13:53:32.617235 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39ecd283_215f_442b_9b13_39ddf4bb3be9.slice/crio-992203a1f5996e0fe7192e32e778050b733917c075fa1e15f625820933d0a109 WatchSource:0}: Error finding container 992203a1f5996e0fe7192e32e778050b733917c075fa1e15f625820933d0a109: Status 404 returned error can't find the container with id 992203a1f5996e0fe7192e32e778050b733917c075fa1e15f625820933d0a109 Feb 02 13:53:32 crc kubenswrapper[4846]: I0202 13:53:32.700127 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6fd74dfb66-674x6"] Feb 02 13:53:32 crc kubenswrapper[4846]: W0202 13:53:32.712318 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod434d073d_68ad_4a28_97f0_8aab974ff3d9.slice/crio-aa750e5e829902ff418562536cff664b0aa3926955809bd8ae03ff326c5f2a6f WatchSource:0}: Error finding container aa750e5e829902ff418562536cff664b0aa3926955809bd8ae03ff326c5f2a6f: Status 404 returned error can't find the container with id aa750e5e829902ff418562536cff664b0aa3926955809bd8ae03ff326c5f2a6f Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.046236 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-769b6db55f-vttw8"] Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.046452 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-769b6db55f-vttw8" podUID="1e35b053-4619-428e-82d2-46f4a12e475c" containerName="heat-api" containerID="cri-o://cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c" gracePeriod=60 Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.055325 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-769b6db55f-vttw8" podUID="1e35b053-4619-428e-82d2-46f4a12e475c" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.130:8004/healthcheck\": EOF" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.072361 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-56c8998cb4-v2r4d"] Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.072792 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" podUID="55343bd7-9622-44c6-81b2-9f140da38a90" containerName="heat-cfnapi" containerID="cri-o://2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd" gracePeriod=60 Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.096481 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" podUID="55343bd7-9622-44c6-81b2-9f140da38a90" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.129:8000/healthcheck\": EOF" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.117497 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-684d5dd44f-9r7lc"] Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.119181 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.130020 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.131241 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.147661 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-684d5dd44f-9r7lc"] Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.163692 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-54dbdfd74b-wqbr7"] Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.165263 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.169593 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.169910 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.186818 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-54dbdfd74b-wqbr7"] Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.224605 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-combined-ca-bundle\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.224760 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-internal-tls-certs\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.224788 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-public-tls-certs\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.224847 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-config-data-custom\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.224889 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-config-data\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.224923 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdqwv\" (UniqueName: \"kubernetes.io/projected/fddebaab-459c-4aba-b375-859a15797a03-kube-api-access-xdqwv\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.229144 4846 generic.go:334] "Generic (PLEG): container finished" podID="39ecd283-215f-442b-9b13-39ddf4bb3be9" containerID="9f499cb4ef72a2cedc1ade2babdf3218bd8998b52d1c09b0e90d5ed6d08f0652" exitCode=1 Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.229255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-869cdc4dcc-vmzv2" event={"ID":"39ecd283-215f-442b-9b13-39ddf4bb3be9","Type":"ContainerDied","Data":"9f499cb4ef72a2cedc1ade2babdf3218bd8998b52d1c09b0e90d5ed6d08f0652"} Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.229290 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-869cdc4dcc-vmzv2" event={"ID":"39ecd283-215f-442b-9b13-39ddf4bb3be9","Type":"ContainerStarted","Data":"992203a1f5996e0fe7192e32e778050b733917c075fa1e15f625820933d0a109"} Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.230016 4846 scope.go:117] "RemoveContainer" containerID="9f499cb4ef72a2cedc1ade2babdf3218bd8998b52d1c09b0e90d5ed6d08f0652" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.250016 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-555b878bf4-mzv78" event={"ID":"4b20035b-bca2-4bfd-86c8-e53deda9da15","Type":"ContainerStarted","Data":"e310c9f6e54b0dfecc3cb6e2a7154bbdddfb86385966525158ef1d9a42e23d34"} Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.250078 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-555b878bf4-mzv78" event={"ID":"4b20035b-bca2-4bfd-86c8-e53deda9da15","Type":"ContainerStarted","Data":"b67e19b772749999cb5a3552d80aace44bb51c201283ac7ae9e9422e2e20f9d8"} Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.251747 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.258968 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" event={"ID":"434d073d-68ad-4a28-97f0-8aab974ff3d9","Type":"ContainerStarted","Data":"aa9ecd9b19a89e65448aa9db00722f5e1b1ba968c979299c76cda72f3fed0902"} Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.259034 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" event={"ID":"434d073d-68ad-4a28-97f0-8aab974ff3d9","Type":"ContainerStarted","Data":"aa750e5e829902ff418562536cff664b0aa3926955809bd8ae03ff326c5f2a6f"} Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.259360 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.353311 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-public-tls-certs\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.353420 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-combined-ca-bundle\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.353498 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-config-data-custom\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.353558 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-internal-tls-certs\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.353576 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-public-tls-certs\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.353594 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-combined-ca-bundle\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.409768 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-config-data-custom\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.409915 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-config-data\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.409981 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdqwv\" (UniqueName: \"kubernetes.io/projected/fddebaab-459c-4aba-b375-859a15797a03-kube-api-access-xdqwv\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.410036 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-internal-tls-certs\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.410079 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh69b\" (UniqueName: \"kubernetes.io/projected/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-kube-api-access-zh69b\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.410122 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-config-data\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.410104 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-555b878bf4-mzv78" podStartSLOduration=2.410088208 podStartE2EDuration="2.410088208s" podCreationTimestamp="2026-02-02 13:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:53:33.342401424 +0000 UTC m=+6244.570988287" watchObservedRunningTime="2026-02-02 13:53:33.410088208 +0000 UTC m=+6244.638675061" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.411258 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-combined-ca-bundle\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.426310 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-public-tls-certs\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.426900 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-internal-tls-certs\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.435433 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-config-data-custom\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.436858 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddebaab-459c-4aba-b375-859a15797a03-config-data\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.487464 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdqwv\" (UniqueName: \"kubernetes.io/projected/fddebaab-459c-4aba-b375-859a15797a03-kube-api-access-xdqwv\") pod \"heat-api-684d5dd44f-9r7lc\" (UID: \"fddebaab-459c-4aba-b375-859a15797a03\") " pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.512186 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.512836 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-config-data-custom\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.512902 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-combined-ca-bundle\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.512971 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-internal-tls-certs\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.512994 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh69b\" (UniqueName: \"kubernetes.io/projected/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-kube-api-access-zh69b\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.513017 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-config-data\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.513058 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-public-tls-certs\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.528185 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" podStartSLOduration=2.528167845 podStartE2EDuration="2.528167845s" podCreationTimestamp="2026-02-02 13:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:53:33.411426831 +0000 UTC m=+6244.640013694" watchObservedRunningTime="2026-02-02 13:53:33.528167845 +0000 UTC m=+6244.756754708" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.531458 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-internal-tls-certs\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.542177 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-config-data-custom\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.542413 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-public-tls-certs\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.547322 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh69b\" (UniqueName: \"kubernetes.io/projected/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-kube-api-access-zh69b\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.547556 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-combined-ca-bundle\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.550049 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc91827-42a5-4fc6-9720-b4dfd5e246d4-config-data\") pod \"heat-cfnapi-54dbdfd74b-wqbr7\" (UID: \"bfc91827-42a5-4fc6-9720-b4dfd5e246d4\") " pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:33 crc kubenswrapper[4846]: I0202 13:53:33.596107 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:34 crc kubenswrapper[4846]: I0202 13:53:34.290061 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-54dbdfd74b-wqbr7"] Feb 02 13:53:34 crc kubenswrapper[4846]: I0202 13:53:34.306668 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-684d5dd44f-9r7lc"] Feb 02 13:53:34 crc kubenswrapper[4846]: I0202 13:53:34.310660 4846 generic.go:334] "Generic (PLEG): container finished" podID="434d073d-68ad-4a28-97f0-8aab974ff3d9" containerID="aa9ecd9b19a89e65448aa9db00722f5e1b1ba968c979299c76cda72f3fed0902" exitCode=1 Feb 02 13:53:34 crc kubenswrapper[4846]: I0202 13:53:34.310764 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" event={"ID":"434d073d-68ad-4a28-97f0-8aab974ff3d9","Type":"ContainerDied","Data":"aa9ecd9b19a89e65448aa9db00722f5e1b1ba968c979299c76cda72f3fed0902"} Feb 02 13:53:34 crc kubenswrapper[4846]: I0202 13:53:34.313480 4846 scope.go:117] "RemoveContainer" containerID="aa9ecd9b19a89e65448aa9db00722f5e1b1ba968c979299c76cda72f3fed0902" Feb 02 13:53:34 crc kubenswrapper[4846]: I0202 13:53:34.384749 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-869cdc4dcc-vmzv2" event={"ID":"39ecd283-215f-442b-9b13-39ddf4bb3be9","Type":"ContainerStarted","Data":"228f04de122620d2f00bf59c0b7b5a3301c007c6d1ea1c9178123954e8e72687"} Feb 02 13:53:34 crc kubenswrapper[4846]: I0202 13:53:34.384816 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:34 crc kubenswrapper[4846]: I0202 13:53:34.438757 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-869cdc4dcc-vmzv2" podStartSLOduration=3.438739708 podStartE2EDuration="3.438739708s" podCreationTimestamp="2026-02-02 13:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:53:34.407850308 +0000 UTC m=+6245.636437181" watchObservedRunningTime="2026-02-02 13:53:34.438739708 +0000 UTC m=+6245.667326571" Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.395271 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-684d5dd44f-9r7lc" event={"ID":"fddebaab-459c-4aba-b375-859a15797a03","Type":"ContainerStarted","Data":"14bd50e6a802d66f6d2f79e44a8bf8ff21316e917c9e415b86f9e1cfb0cc6468"} Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.395693 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-684d5dd44f-9r7lc" event={"ID":"fddebaab-459c-4aba-b375-859a15797a03","Type":"ContainerStarted","Data":"cc379ccf8f4fe01b3bba98bfc0b3828e7894e42af8325f22f0ac475a3095d95f"} Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.395967 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.396857 4846 generic.go:334] "Generic (PLEG): container finished" podID="39ecd283-215f-442b-9b13-39ddf4bb3be9" containerID="228f04de122620d2f00bf59c0b7b5a3301c007c6d1ea1c9178123954e8e72687" exitCode=1 Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.396925 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-869cdc4dcc-vmzv2" event={"ID":"39ecd283-215f-442b-9b13-39ddf4bb3be9","Type":"ContainerDied","Data":"228f04de122620d2f00bf59c0b7b5a3301c007c6d1ea1c9178123954e8e72687"} Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.396954 4846 scope.go:117] "RemoveContainer" containerID="9f499cb4ef72a2cedc1ade2babdf3218bd8998b52d1c09b0e90d5ed6d08f0652" Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.397291 4846 scope.go:117] "RemoveContainer" containerID="228f04de122620d2f00bf59c0b7b5a3301c007c6d1ea1c9178123954e8e72687" Feb 02 13:53:35 crc kubenswrapper[4846]: E0202 13:53:35.397488 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-869cdc4dcc-vmzv2_openstack(39ecd283-215f-442b-9b13-39ddf4bb3be9)\"" pod="openstack/heat-api-869cdc4dcc-vmzv2" podUID="39ecd283-215f-442b-9b13-39ddf4bb3be9" Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.400273 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" event={"ID":"bfc91827-42a5-4fc6-9720-b4dfd5e246d4","Type":"ContainerStarted","Data":"a2050dc2a056f3a99c05a3968f25999bbc0104fd3e51764565282ce181951d0f"} Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.400304 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" event={"ID":"bfc91827-42a5-4fc6-9720-b4dfd5e246d4","Type":"ContainerStarted","Data":"3cf7f64e09573aaf5b95c2fb24e91b06c0b25f025eda4e275ba35b4164766b32"} Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.400871 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.403368 4846 generic.go:334] "Generic (PLEG): container finished" podID="434d073d-68ad-4a28-97f0-8aab974ff3d9" containerID="9ec7f639eb3b8839a57292b30ab53f0fafa83ef8ad644376bfbaa36b26940917" exitCode=1 Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.403396 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" event={"ID":"434d073d-68ad-4a28-97f0-8aab974ff3d9","Type":"ContainerDied","Data":"9ec7f639eb3b8839a57292b30ab53f0fafa83ef8ad644376bfbaa36b26940917"} Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.403821 4846 scope.go:117] "RemoveContainer" containerID="9ec7f639eb3b8839a57292b30ab53f0fafa83ef8ad644376bfbaa36b26940917" Feb 02 13:53:35 crc kubenswrapper[4846]: E0202 13:53:35.404047 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6fd74dfb66-674x6_openstack(434d073d-68ad-4a28-97f0-8aab974ff3d9)\"" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" podUID="434d073d-68ad-4a28-97f0-8aab974ff3d9" Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.446774 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-684d5dd44f-9r7lc" podStartSLOduration=2.446755918 podStartE2EDuration="2.446755918s" podCreationTimestamp="2026-02-02 13:53:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:53:35.437606837 +0000 UTC m=+6246.666193710" watchObservedRunningTime="2026-02-02 13:53:35.446755918 +0000 UTC m=+6246.675342781" Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.478575 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" podStartSLOduration=2.478549611 podStartE2EDuration="2.478549611s" podCreationTimestamp="2026-02-02 13:53:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:53:35.470079335 +0000 UTC m=+6246.698666198" watchObservedRunningTime="2026-02-02 13:53:35.478549611 +0000 UTC m=+6246.707136474" Feb 02 13:53:35 crc kubenswrapper[4846]: I0202 13:53:35.479302 4846 scope.go:117] "RemoveContainer" containerID="aa9ecd9b19a89e65448aa9db00722f5e1b1ba968c979299c76cda72f3fed0902" Feb 02 13:53:36 crc kubenswrapper[4846]: I0202 13:53:36.416261 4846 scope.go:117] "RemoveContainer" containerID="9ec7f639eb3b8839a57292b30ab53f0fafa83ef8ad644376bfbaa36b26940917" Feb 02 13:53:36 crc kubenswrapper[4846]: E0202 13:53:36.416804 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6fd74dfb66-674x6_openstack(434d073d-68ad-4a28-97f0-8aab974ff3d9)\"" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" podUID="434d073d-68ad-4a28-97f0-8aab974ff3d9" Feb 02 13:53:36 crc kubenswrapper[4846]: I0202 13:53:36.418942 4846 scope.go:117] "RemoveContainer" containerID="228f04de122620d2f00bf59c0b7b5a3301c007c6d1ea1c9178123954e8e72687" Feb 02 13:53:36 crc kubenswrapper[4846]: E0202 13:53:36.419129 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-869cdc4dcc-vmzv2_openstack(39ecd283-215f-442b-9b13-39ddf4bb3be9)\"" pod="openstack/heat-api-869cdc4dcc-vmzv2" podUID="39ecd283-215f-442b-9b13-39ddf4bb3be9" Feb 02 13:53:36 crc kubenswrapper[4846]: I0202 13:53:36.582181 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:37 crc kubenswrapper[4846]: I0202 13:53:37.010834 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:37 crc kubenswrapper[4846]: I0202 13:53:37.043684 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:37 crc kubenswrapper[4846]: I0202 13:53:37.043735 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:37 crc kubenswrapper[4846]: I0202 13:53:37.427075 4846 scope.go:117] "RemoveContainer" containerID="228f04de122620d2f00bf59c0b7b5a3301c007c6d1ea1c9178123954e8e72687" Feb 02 13:53:37 crc kubenswrapper[4846]: I0202 13:53:37.427147 4846 scope.go:117] "RemoveContainer" containerID="9ec7f639eb3b8839a57292b30ab53f0fafa83ef8ad644376bfbaa36b26940917" Feb 02 13:53:37 crc kubenswrapper[4846]: E0202 13:53:37.427371 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-869cdc4dcc-vmzv2_openstack(39ecd283-215f-442b-9b13-39ddf4bb3be9)\"" pod="openstack/heat-api-869cdc4dcc-vmzv2" podUID="39ecd283-215f-442b-9b13-39ddf4bb3be9" Feb 02 13:53:37 crc kubenswrapper[4846]: E0202 13:53:37.427375 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6fd74dfb66-674x6_openstack(434d073d-68ad-4a28-97f0-8aab974ff3d9)\"" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" podUID="434d073d-68ad-4a28-97f0-8aab974ff3d9" Feb 02 13:53:38 crc kubenswrapper[4846]: I0202 13:53:38.444727 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5c49958db-lgw6w" Feb 02 13:53:38 crc kubenswrapper[4846]: I0202 13:53:38.489453 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" podUID="55343bd7-9622-44c6-81b2-9f140da38a90" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.129:8000/healthcheck\": read tcp 10.217.0.2:33350->10.217.1.129:8000: read: connection reset by peer" Feb 02 13:53:38 crc kubenswrapper[4846]: I0202 13:53:38.506836 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d584cfd6-zscgj"] Feb 02 13:53:38 crc kubenswrapper[4846]: I0202 13:53:38.507063 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d584cfd6-zscgj" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon-log" containerID="cri-o://bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572" gracePeriod=30 Feb 02 13:53:38 crc kubenswrapper[4846]: I0202 13:53:38.507153 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d584cfd6-zscgj" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon" containerID="cri-o://1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658" gracePeriod=30 Feb 02 13:53:38 crc kubenswrapper[4846]: I0202 13:53:38.996964 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.062750 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data\") pod \"55343bd7-9622-44c6-81b2-9f140da38a90\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.063149 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpcrm\" (UniqueName: \"kubernetes.io/projected/55343bd7-9622-44c6-81b2-9f140da38a90-kube-api-access-cpcrm\") pod \"55343bd7-9622-44c6-81b2-9f140da38a90\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.063338 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data-custom\") pod \"55343bd7-9622-44c6-81b2-9f140da38a90\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.063445 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-combined-ca-bundle\") pod \"55343bd7-9622-44c6-81b2-9f140da38a90\" (UID: \"55343bd7-9622-44c6-81b2-9f140da38a90\") " Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.069513 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "55343bd7-9622-44c6-81b2-9f140da38a90" (UID: "55343bd7-9622-44c6-81b2-9f140da38a90"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.069883 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55343bd7-9622-44c6-81b2-9f140da38a90-kube-api-access-cpcrm" (OuterVolumeSpecName: "kube-api-access-cpcrm") pod "55343bd7-9622-44c6-81b2-9f140da38a90" (UID: "55343bd7-9622-44c6-81b2-9f140da38a90"). InnerVolumeSpecName "kube-api-access-cpcrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.112563 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55343bd7-9622-44c6-81b2-9f140da38a90" (UID: "55343bd7-9622-44c6-81b2-9f140da38a90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.128681 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data" (OuterVolumeSpecName: "config-data") pod "55343bd7-9622-44c6-81b2-9f140da38a90" (UID: "55343bd7-9622-44c6-81b2-9f140da38a90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.164907 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.164947 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpcrm\" (UniqueName: \"kubernetes.io/projected/55343bd7-9622-44c6-81b2-9f140da38a90-kube-api-access-cpcrm\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.164959 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.164973 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55343bd7-9622-44c6-81b2-9f140da38a90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.445965 4846 generic.go:334] "Generic (PLEG): container finished" podID="55343bd7-9622-44c6-81b2-9f140da38a90" containerID="2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd" exitCode=0 Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.446025 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" event={"ID":"55343bd7-9622-44c6-81b2-9f140da38a90","Type":"ContainerDied","Data":"2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd"} Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.446057 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" event={"ID":"55343bd7-9622-44c6-81b2-9f140da38a90","Type":"ContainerDied","Data":"0c5cb5857a6b32938e58c700e3b8dccbf930f277818807abf700b1d24a214deb"} Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.446076 4846 scope.go:117] "RemoveContainer" containerID="2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.446200 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-56c8998cb4-v2r4d" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.481578 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-769b6db55f-vttw8" podUID="1e35b053-4619-428e-82d2-46f4a12e475c" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.130:8004/healthcheck\": read tcp 10.217.0.2:45902->10.217.1.130:8004: read: connection reset by peer" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.488228 4846 scope.go:117] "RemoveContainer" containerID="2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd" Feb 02 13:53:39 crc kubenswrapper[4846]: E0202 13:53:39.490124 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd\": container with ID starting with 2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd not found: ID does not exist" containerID="2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.490168 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd"} err="failed to get container status \"2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd\": rpc error: code = NotFound desc = could not find container \"2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd\": container with ID starting with 2c0114d2afc18dcdd7bbc057c2dcf48064095a48a45583fae4071e4d484475dd not found: ID does not exist" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.497831 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-56c8998cb4-v2r4d"] Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.509744 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-56c8998cb4-v2r4d"] Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.953713 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.996653 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data-custom\") pod \"1e35b053-4619-428e-82d2-46f4a12e475c\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.996898 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf4kj\" (UniqueName: \"kubernetes.io/projected/1e35b053-4619-428e-82d2-46f4a12e475c-kube-api-access-qf4kj\") pod \"1e35b053-4619-428e-82d2-46f4a12e475c\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.996929 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data\") pod \"1e35b053-4619-428e-82d2-46f4a12e475c\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " Feb 02 13:53:39 crc kubenswrapper[4846]: I0202 13:53:39.997024 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-combined-ca-bundle\") pod \"1e35b053-4619-428e-82d2-46f4a12e475c\" (UID: \"1e35b053-4619-428e-82d2-46f4a12e475c\") " Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.013642 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e35b053-4619-428e-82d2-46f4a12e475c-kube-api-access-qf4kj" (OuterVolumeSpecName: "kube-api-access-qf4kj") pod "1e35b053-4619-428e-82d2-46f4a12e475c" (UID: "1e35b053-4619-428e-82d2-46f4a12e475c"). InnerVolumeSpecName "kube-api-access-qf4kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.015446 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1e35b053-4619-428e-82d2-46f4a12e475c" (UID: "1e35b053-4619-428e-82d2-46f4a12e475c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.046863 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e35b053-4619-428e-82d2-46f4a12e475c" (UID: "1e35b053-4619-428e-82d2-46f4a12e475c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.062743 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5fd7-account-create-update-2k6hx"] Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.072211 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data" (OuterVolumeSpecName: "config-data") pod "1e35b053-4619-428e-82d2-46f4a12e475c" (UID: "1e35b053-4619-428e-82d2-46f4a12e475c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.078954 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5fd7-account-create-update-2k6hx"] Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.091991 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-cgm6d"] Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.099547 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf4kj\" (UniqueName: \"kubernetes.io/projected/1e35b053-4619-428e-82d2-46f4a12e475c-kube-api-access-qf4kj\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.099584 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.099596 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.099609 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e35b053-4619-428e-82d2-46f4a12e475c-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.104057 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-cgm6d"] Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.459172 4846 generic.go:334] "Generic (PLEG): container finished" podID="1e35b053-4619-428e-82d2-46f4a12e475c" containerID="cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c" exitCode=0 Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.459215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-769b6db55f-vttw8" event={"ID":"1e35b053-4619-428e-82d2-46f4a12e475c","Type":"ContainerDied","Data":"cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c"} Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.459237 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-769b6db55f-vttw8" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.459258 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-769b6db55f-vttw8" event={"ID":"1e35b053-4619-428e-82d2-46f4a12e475c","Type":"ContainerDied","Data":"67c1f1bc89130ad735e40e7b9ca8600439f8eca70b19e32bfc9216922ef76766"} Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.459284 4846 scope.go:117] "RemoveContainer" containerID="cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.489395 4846 scope.go:117] "RemoveContainer" containerID="cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c" Feb 02 13:53:40 crc kubenswrapper[4846]: E0202 13:53:40.490159 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c\": container with ID starting with cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c not found: ID does not exist" containerID="cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.490201 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c"} err="failed to get container status \"cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c\": rpc error: code = NotFound desc = could not find container \"cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c\": container with ID starting with cdb121fc5582c58ce1106634ade7daacd944300c957ee340310bd82a72d19a8c not found: ID does not exist" Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.497681 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-769b6db55f-vttw8"] Feb 02 13:53:40 crc kubenswrapper[4846]: I0202 13:53:40.514762 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-769b6db55f-vttw8"] Feb 02 13:53:42 crc kubenswrapper[4846]: I0202 13:53:42.107417 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e35b053-4619-428e-82d2-46f4a12e475c" path="/var/lib/kubelet/pods/1e35b053-4619-428e-82d2-46f4a12e475c/volumes" Feb 02 13:53:42 crc kubenswrapper[4846]: I0202 13:53:42.108008 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55343bd7-9622-44c6-81b2-9f140da38a90" path="/var/lib/kubelet/pods/55343bd7-9622-44c6-81b2-9f140da38a90/volumes" Feb 02 13:53:42 crc kubenswrapper[4846]: I0202 13:53:42.109894 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81aec00d-a484-44aa-a06d-57bc3a25c481" path="/var/lib/kubelet/pods/81aec00d-a484-44aa-a06d-57bc3a25c481/volumes" Feb 02 13:53:42 crc kubenswrapper[4846]: I0202 13:53:42.111356 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eb3025e-a013-471a-bf21-d28e12c189f1" path="/var/lib/kubelet/pods/9eb3025e-a013-471a-bf21-d28e12c189f1/volumes" Feb 02 13:53:42 crc kubenswrapper[4846]: I0202 13:53:42.167874 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-555b878bf4-mzv78" Feb 02 13:53:42 crc kubenswrapper[4846]: I0202 13:53:42.212479 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-677998f9bd-zfdvm"] Feb 02 13:53:42 crc kubenswrapper[4846]: I0202 13:53:42.212932 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-677998f9bd-zfdvm" podUID="70b265c4-72ec-485b-a147-97e1a9ac0a1b" containerName="heat-engine" containerID="cri-o://47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" gracePeriod=60 Feb 02 13:53:42 crc kubenswrapper[4846]: E0202 13:53:42.218669 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 02 13:53:42 crc kubenswrapper[4846]: E0202 13:53:42.251731 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 02 13:53:42 crc kubenswrapper[4846]: E0202 13:53:42.262075 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 02 13:53:42 crc kubenswrapper[4846]: E0202 13:53:42.262189 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-677998f9bd-zfdvm" podUID="70b265c4-72ec-485b-a147-97e1a9ac0a1b" containerName="heat-engine" Feb 02 13:53:43 crc kubenswrapper[4846]: I0202 13:53:43.114636 4846 generic.go:334] "Generic (PLEG): container finished" podID="510040f2-b874-48d8-bfb1-871b8abf0199" containerID="1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658" exitCode=0 Feb 02 13:53:43 crc kubenswrapper[4846]: I0202 13:53:43.114893 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d584cfd6-zscgj" event={"ID":"510040f2-b874-48d8-bfb1-871b8abf0199","Type":"ContainerDied","Data":"1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658"} Feb 02 13:53:44 crc kubenswrapper[4846]: E0202 13:53:44.501709 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 02 13:53:44 crc kubenswrapper[4846]: E0202 13:53:44.507123 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 02 13:53:44 crc kubenswrapper[4846]: E0202 13:53:44.523866 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 02 13:53:44 crc kubenswrapper[4846]: E0202 13:53:44.523949 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-677998f9bd-zfdvm" podUID="70b265c4-72ec-485b-a147-97e1a9ac0a1b" containerName="heat-engine" Feb 02 13:53:45 crc kubenswrapper[4846]: I0202 13:53:45.005152 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-54dbdfd74b-wqbr7" Feb 02 13:53:45 crc kubenswrapper[4846]: I0202 13:53:45.025531 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-684d5dd44f-9r7lc" Feb 02 13:53:45 crc kubenswrapper[4846]: I0202 13:53:45.122016 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6fd74dfb66-674x6"] Feb 02 13:53:45 crc kubenswrapper[4846]: I0202 13:53:45.165608 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-869cdc4dcc-vmzv2"] Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.665833 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.671515 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.786164 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-combined-ca-bundle\") pod \"39ecd283-215f-442b-9b13-39ddf4bb3be9\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.786276 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data\") pod \"39ecd283-215f-442b-9b13-39ddf4bb3be9\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.786341 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data-custom\") pod \"39ecd283-215f-442b-9b13-39ddf4bb3be9\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.786484 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-combined-ca-bundle\") pod \"434d073d-68ad-4a28-97f0-8aab974ff3d9\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.786524 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj9wf\" (UniqueName: \"kubernetes.io/projected/39ecd283-215f-442b-9b13-39ddf4bb3be9-kube-api-access-pj9wf\") pod \"39ecd283-215f-442b-9b13-39ddf4bb3be9\" (UID: \"39ecd283-215f-442b-9b13-39ddf4bb3be9\") " Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.786577 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data\") pod \"434d073d-68ad-4a28-97f0-8aab974ff3d9\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.786597 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data-custom\") pod \"434d073d-68ad-4a28-97f0-8aab974ff3d9\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.786661 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8ztg\" (UniqueName: \"kubernetes.io/projected/434d073d-68ad-4a28-97f0-8aab974ff3d9-kube-api-access-g8ztg\") pod \"434d073d-68ad-4a28-97f0-8aab974ff3d9\" (UID: \"434d073d-68ad-4a28-97f0-8aab974ff3d9\") " Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.794427 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434d073d-68ad-4a28-97f0-8aab974ff3d9-kube-api-access-g8ztg" (OuterVolumeSpecName: "kube-api-access-g8ztg") pod "434d073d-68ad-4a28-97f0-8aab974ff3d9" (UID: "434d073d-68ad-4a28-97f0-8aab974ff3d9"). InnerVolumeSpecName "kube-api-access-g8ztg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.803929 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "39ecd283-215f-442b-9b13-39ddf4bb3be9" (UID: "39ecd283-215f-442b-9b13-39ddf4bb3be9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.804022 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "434d073d-68ad-4a28-97f0-8aab974ff3d9" (UID: "434d073d-68ad-4a28-97f0-8aab974ff3d9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.811332 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39ecd283-215f-442b-9b13-39ddf4bb3be9-kube-api-access-pj9wf" (OuterVolumeSpecName: "kube-api-access-pj9wf") pod "39ecd283-215f-442b-9b13-39ddf4bb3be9" (UID: "39ecd283-215f-442b-9b13-39ddf4bb3be9"). InnerVolumeSpecName "kube-api-access-pj9wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.830638 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39ecd283-215f-442b-9b13-39ddf4bb3be9" (UID: "39ecd283-215f-442b-9b13-39ddf4bb3be9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.834900 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "434d073d-68ad-4a28-97f0-8aab974ff3d9" (UID: "434d073d-68ad-4a28-97f0-8aab974ff3d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.861776 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data" (OuterVolumeSpecName: "config-data") pod "39ecd283-215f-442b-9b13-39ddf4bb3be9" (UID: "39ecd283-215f-442b-9b13-39ddf4bb3be9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.879363 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data" (OuterVolumeSpecName: "config-data") pod "434d073d-68ad-4a28-97f0-8aab974ff3d9" (UID: "434d073d-68ad-4a28-97f0-8aab974ff3d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.889606 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.889665 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj9wf\" (UniqueName: \"kubernetes.io/projected/39ecd283-215f-442b-9b13-39ddf4bb3be9-kube-api-access-pj9wf\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.889679 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.889691 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/434d073d-68ad-4a28-97f0-8aab974ff3d9-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.889702 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8ztg\" (UniqueName: \"kubernetes.io/projected/434d073d-68ad-4a28-97f0-8aab974ff3d9-kube-api-access-g8ztg\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.889713 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.889724 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:45.889737 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39ecd283-215f-442b-9b13-39ddf4bb3be9-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:46.160027 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" event={"ID":"434d073d-68ad-4a28-97f0-8aab974ff3d9","Type":"ContainerDied","Data":"aa750e5e829902ff418562536cff664b0aa3926955809bd8ae03ff326c5f2a6f"} Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:46.160146 4846 scope.go:117] "RemoveContainer" containerID="9ec7f639eb3b8839a57292b30ab53f0fafa83ef8ad644376bfbaa36b26940917" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:46.160303 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6fd74dfb66-674x6" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:46.164124 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-869cdc4dcc-vmzv2" event={"ID":"39ecd283-215f-442b-9b13-39ddf4bb3be9","Type":"ContainerDied","Data":"992203a1f5996e0fe7192e32e778050b733917c075fa1e15f625820933d0a109"} Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:46.164210 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-869cdc4dcc-vmzv2" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:46.213686 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6fd74dfb66-674x6"] Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:46.225435 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-6fd74dfb66-674x6"] Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:46.225642 4846 scope.go:117] "RemoveContainer" containerID="228f04de122620d2f00bf59c0b7b5a3301c007c6d1ea1c9178123954e8e72687" Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:46.247455 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-869cdc4dcc-vmzv2"] Feb 02 13:53:46 crc kubenswrapper[4846]: I0202 13:53:46.261866 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-869cdc4dcc-vmzv2"] Feb 02 13:53:47 crc kubenswrapper[4846]: I0202 13:53:47.438153 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39ecd283-215f-442b-9b13-39ddf4bb3be9" path="/var/lib/kubelet/pods/39ecd283-215f-442b-9b13-39ddf4bb3be9/volumes" Feb 02 13:53:47 crc kubenswrapper[4846]: I0202 13:53:47.439251 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="434d073d-68ad-4a28-97f0-8aab974ff3d9" path="/var/lib/kubelet/pods/434d073d-68ad-4a28-97f0-8aab974ff3d9/volumes" Feb 02 13:53:48 crc kubenswrapper[4846]: I0202 13:53:48.019698 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d584cfd6-zscgj" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.121:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.121:8443: connect: connection refused" Feb 02 13:53:48 crc kubenswrapper[4846]: I0202 13:53:48.051374 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-wxpwk"] Feb 02 13:53:48 crc kubenswrapper[4846]: I0202 13:53:48.061189 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-wxpwk"] Feb 02 13:53:49 crc kubenswrapper[4846]: I0202 13:53:49.438514 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5a1aebe-43c7-4791-a2ff-b81a10230d01" path="/var/lib/kubelet/pods/a5a1aebe-43c7-4791-a2ff-b81a10230d01/volumes" Feb 02 13:53:49 crc kubenswrapper[4846]: I0202 13:53:49.815798 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-769b6db55f-vttw8" podUID="1e35b053-4619-428e-82d2-46f4a12e475c" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.130:8004/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 13:53:54 crc kubenswrapper[4846]: E0202 13:53:54.504981 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 02 13:53:54 crc kubenswrapper[4846]: E0202 13:53:54.511055 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 02 13:53:54 crc kubenswrapper[4846]: E0202 13:53:54.513267 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 02 13:53:54 crc kubenswrapper[4846]: E0202 13:53:54.513408 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-677998f9bd-zfdvm" podUID="70b265c4-72ec-485b-a147-97e1a9ac0a1b" containerName="heat-engine" Feb 02 13:53:57 crc kubenswrapper[4846]: I0202 13:53:57.778377 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:57 crc kubenswrapper[4846]: I0202 13:53:57.937430 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-combined-ca-bundle\") pod \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " Feb 02 13:53:57 crc kubenswrapper[4846]: I0202 13:53:57.937836 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data\") pod \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " Feb 02 13:53:57 crc kubenswrapper[4846]: I0202 13:53:57.937890 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6vl6\" (UniqueName: \"kubernetes.io/projected/70b265c4-72ec-485b-a147-97e1a9ac0a1b-kube-api-access-w6vl6\") pod \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " Feb 02 13:53:57 crc kubenswrapper[4846]: I0202 13:53:57.937929 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data-custom\") pod \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\" (UID: \"70b265c4-72ec-485b-a147-97e1a9ac0a1b\") " Feb 02 13:53:57 crc kubenswrapper[4846]: I0202 13:53:57.943312 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70b265c4-72ec-485b-a147-97e1a9ac0a1b-kube-api-access-w6vl6" (OuterVolumeSpecName: "kube-api-access-w6vl6") pod "70b265c4-72ec-485b-a147-97e1a9ac0a1b" (UID: "70b265c4-72ec-485b-a147-97e1a9ac0a1b"). InnerVolumeSpecName "kube-api-access-w6vl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:53:57 crc kubenswrapper[4846]: I0202 13:53:57.943418 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "70b265c4-72ec-485b-a147-97e1a9ac0a1b" (UID: "70b265c4-72ec-485b-a147-97e1a9ac0a1b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:57 crc kubenswrapper[4846]: I0202 13:53:57.967269 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70b265c4-72ec-485b-a147-97e1a9ac0a1b" (UID: "70b265c4-72ec-485b-a147-97e1a9ac0a1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:57 crc kubenswrapper[4846]: I0202 13:53:57.992443 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data" (OuterVolumeSpecName: "config-data") pod "70b265c4-72ec-485b-a147-97e1a9ac0a1b" (UID: "70b265c4-72ec-485b-a147-97e1a9ac0a1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.020597 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d584cfd6-zscgj" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.121:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.121:8443: connect: connection refused" Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.039876 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.039911 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6vl6\" (UniqueName: \"kubernetes.io/projected/70b265c4-72ec-485b-a147-97e1a9ac0a1b-kube-api-access-w6vl6\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.039921 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.039929 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b265c4-72ec-485b-a147-97e1a9ac0a1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.277049 4846 generic.go:334] "Generic (PLEG): container finished" podID="70b265c4-72ec-485b-a147-97e1a9ac0a1b" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" exitCode=0 Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.277094 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-677998f9bd-zfdvm" Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.277109 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-677998f9bd-zfdvm" event={"ID":"70b265c4-72ec-485b-a147-97e1a9ac0a1b","Type":"ContainerDied","Data":"47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053"} Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.277156 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-677998f9bd-zfdvm" event={"ID":"70b265c4-72ec-485b-a147-97e1a9ac0a1b","Type":"ContainerDied","Data":"702733b699c3f65bd3b79776ce18f65048346f0835223c453039e58d9b6eb78d"} Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.277181 4846 scope.go:117] "RemoveContainer" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.309551 4846 scope.go:117] "RemoveContainer" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" Feb 02 13:53:58 crc kubenswrapper[4846]: E0202 13:53:58.310042 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053\": container with ID starting with 47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053 not found: ID does not exist" containerID="47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053" Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.310067 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053"} err="failed to get container status \"47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053\": rpc error: code = NotFound desc = could not find container \"47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053\": container with ID starting with 47bf8f2fa7cba7b1c73b77e67a4b021aeeb6b25e615da608ec131f45771d6053 not found: ID does not exist" Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.313107 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-677998f9bd-zfdvm"] Feb 02 13:53:58 crc kubenswrapper[4846]: I0202 13:53:58.322828 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-677998f9bd-zfdvm"] Feb 02 13:53:59 crc kubenswrapper[4846]: I0202 13:53:59.435938 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70b265c4-72ec-485b-a147-97e1a9ac0a1b" path="/var/lib/kubelet/pods/70b265c4-72ec-485b-a147-97e1a9ac0a1b/volumes" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.020074 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d584cfd6-zscgj" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.121:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.121:8443: connect: connection refused" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.020696 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.657528 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf"] Feb 02 13:54:08 crc kubenswrapper[4846]: E0202 13:54:08.658239 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e35b053-4619-428e-82d2-46f4a12e475c" containerName="heat-api" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658259 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e35b053-4619-428e-82d2-46f4a12e475c" containerName="heat-api" Feb 02 13:54:08 crc kubenswrapper[4846]: E0202 13:54:08.658278 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434d073d-68ad-4a28-97f0-8aab974ff3d9" containerName="heat-cfnapi" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658288 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="434d073d-68ad-4a28-97f0-8aab974ff3d9" containerName="heat-cfnapi" Feb 02 13:54:08 crc kubenswrapper[4846]: E0202 13:54:08.658303 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70b265c4-72ec-485b-a147-97e1a9ac0a1b" containerName="heat-engine" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658311 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="70b265c4-72ec-485b-a147-97e1a9ac0a1b" containerName="heat-engine" Feb 02 13:54:08 crc kubenswrapper[4846]: E0202 13:54:08.658344 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434d073d-68ad-4a28-97f0-8aab974ff3d9" containerName="heat-cfnapi" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658353 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="434d073d-68ad-4a28-97f0-8aab974ff3d9" containerName="heat-cfnapi" Feb 02 13:54:08 crc kubenswrapper[4846]: E0202 13:54:08.658363 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55343bd7-9622-44c6-81b2-9f140da38a90" containerName="heat-cfnapi" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658369 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="55343bd7-9622-44c6-81b2-9f140da38a90" containerName="heat-cfnapi" Feb 02 13:54:08 crc kubenswrapper[4846]: E0202 13:54:08.658390 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ecd283-215f-442b-9b13-39ddf4bb3be9" containerName="heat-api" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658396 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ecd283-215f-442b-9b13-39ddf4bb3be9" containerName="heat-api" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658566 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ecd283-215f-442b-9b13-39ddf4bb3be9" containerName="heat-api" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658584 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="70b265c4-72ec-485b-a147-97e1a9ac0a1b" containerName="heat-engine" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658595 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="434d073d-68ad-4a28-97f0-8aab974ff3d9" containerName="heat-cfnapi" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658609 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="55343bd7-9622-44c6-81b2-9f140da38a90" containerName="heat-cfnapi" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658653 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e35b053-4619-428e-82d2-46f4a12e475c" containerName="heat-api" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658673 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ecd283-215f-442b-9b13-39ddf4bb3be9" containerName="heat-api" Feb 02 13:54:08 crc kubenswrapper[4846]: E0202 13:54:08.658862 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ecd283-215f-442b-9b13-39ddf4bb3be9" containerName="heat-api" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.658880 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ecd283-215f-442b-9b13-39ddf4bb3be9" containerName="heat-api" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.659124 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="434d073d-68ad-4a28-97f0-8aab974ff3d9" containerName="heat-cfnapi" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.660207 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.664159 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.681374 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf"] Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.767024 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd747\" (UniqueName: \"kubernetes.io/projected/816eb531-874d-4167-8351-ceef3cd2704a-kube-api-access-xd747\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.767073 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.767122 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.868704 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.869293 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.869351 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd747\" (UniqueName: \"kubernetes.io/projected/816eb531-874d-4167-8351-ceef3cd2704a-kube-api-access-xd747\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.869396 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.869825 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.890416 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd747\" (UniqueName: \"kubernetes.io/projected/816eb531-874d-4167-8351-ceef3cd2704a-kube-api-access-xd747\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.959513 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:54:08 crc kubenswrapper[4846]: I0202 13:54:08.985403 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.072614 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccvc9\" (UniqueName: \"kubernetes.io/projected/510040f2-b874-48d8-bfb1-871b8abf0199-kube-api-access-ccvc9\") pod \"510040f2-b874-48d8-bfb1-871b8abf0199\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.072973 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-secret-key\") pod \"510040f2-b874-48d8-bfb1-871b8abf0199\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.073027 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-scripts\") pod \"510040f2-b874-48d8-bfb1-871b8abf0199\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.073429 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-config-data\") pod \"510040f2-b874-48d8-bfb1-871b8abf0199\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.073483 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/510040f2-b874-48d8-bfb1-871b8abf0199-logs\") pod \"510040f2-b874-48d8-bfb1-871b8abf0199\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.073541 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-combined-ca-bundle\") pod \"510040f2-b874-48d8-bfb1-871b8abf0199\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.073570 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-tls-certs\") pod \"510040f2-b874-48d8-bfb1-871b8abf0199\" (UID: \"510040f2-b874-48d8-bfb1-871b8abf0199\") " Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.075015 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/510040f2-b874-48d8-bfb1-871b8abf0199-logs" (OuterVolumeSpecName: "logs") pod "510040f2-b874-48d8-bfb1-871b8abf0199" (UID: "510040f2-b874-48d8-bfb1-871b8abf0199"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.078040 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/510040f2-b874-48d8-bfb1-871b8abf0199-kube-api-access-ccvc9" (OuterVolumeSpecName: "kube-api-access-ccvc9") pod "510040f2-b874-48d8-bfb1-871b8abf0199" (UID: "510040f2-b874-48d8-bfb1-871b8abf0199"). InnerVolumeSpecName "kube-api-access-ccvc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.080266 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "510040f2-b874-48d8-bfb1-871b8abf0199" (UID: "510040f2-b874-48d8-bfb1-871b8abf0199"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.102775 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-scripts" (OuterVolumeSpecName: "scripts") pod "510040f2-b874-48d8-bfb1-871b8abf0199" (UID: "510040f2-b874-48d8-bfb1-871b8abf0199"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.103224 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-config-data" (OuterVolumeSpecName: "config-data") pod "510040f2-b874-48d8-bfb1-871b8abf0199" (UID: "510040f2-b874-48d8-bfb1-871b8abf0199"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.111816 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "510040f2-b874-48d8-bfb1-871b8abf0199" (UID: "510040f2-b874-48d8-bfb1-871b8abf0199"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.132763 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "510040f2-b874-48d8-bfb1-871b8abf0199" (UID: "510040f2-b874-48d8-bfb1-871b8abf0199"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.176296 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccvc9\" (UniqueName: \"kubernetes.io/projected/510040f2-b874-48d8-bfb1-871b8abf0199-kube-api-access-ccvc9\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.176336 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.176349 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.176361 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/510040f2-b874-48d8-bfb1-871b8abf0199-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.176372 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/510040f2-b874-48d8-bfb1-871b8abf0199-logs\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.176382 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.176392 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/510040f2-b874-48d8-bfb1-871b8abf0199-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.453813 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf"] Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.463992 4846 generic.go:334] "Generic (PLEG): container finished" podID="510040f2-b874-48d8-bfb1-871b8abf0199" containerID="bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572" exitCode=137 Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.464046 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d584cfd6-zscgj" event={"ID":"510040f2-b874-48d8-bfb1-871b8abf0199","Type":"ContainerDied","Data":"bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572"} Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.464028 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d584cfd6-zscgj" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.464077 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d584cfd6-zscgj" event={"ID":"510040f2-b874-48d8-bfb1-871b8abf0199","Type":"ContainerDied","Data":"79842b909a357865e62bad261aad7a07f20c124f6847c50b9ab0bb31f90711d7"} Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.464095 4846 scope.go:117] "RemoveContainer" containerID="1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.501458 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d584cfd6-zscgj"] Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.512355 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d584cfd6-zscgj"] Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.671318 4846 scope.go:117] "RemoveContainer" containerID="bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.741872 4846 scope.go:117] "RemoveContainer" containerID="1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658" Feb 02 13:54:09 crc kubenswrapper[4846]: E0202 13:54:09.742398 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658\": container with ID starting with 1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658 not found: ID does not exist" containerID="1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.742428 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658"} err="failed to get container status \"1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658\": rpc error: code = NotFound desc = could not find container \"1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658\": container with ID starting with 1b1742b0382fcebb4889bdbd552e4eda79bb645c43fbcf5e038bfda1bdf93658 not found: ID does not exist" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.742449 4846 scope.go:117] "RemoveContainer" containerID="bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572" Feb 02 13:54:09 crc kubenswrapper[4846]: E0202 13:54:09.742892 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572\": container with ID starting with bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572 not found: ID does not exist" containerID="bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572" Feb 02 13:54:09 crc kubenswrapper[4846]: I0202 13:54:09.742944 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572"} err="failed to get container status \"bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572\": rpc error: code = NotFound desc = could not find container \"bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572\": container with ID starting with bf411550dc1962bbaf935ca2434d0936637e15e2f7a6b0e1166ab61138321572 not found: ID does not exist" Feb 02 13:54:10 crc kubenswrapper[4846]: I0202 13:54:10.476984 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" event={"ID":"816eb531-874d-4167-8351-ceef3cd2704a","Type":"ContainerStarted","Data":"4f3a95a6f215190ec4ecd151772d2e0448650178e8ef8323660fbb961da4d6fc"} Feb 02 13:54:10 crc kubenswrapper[4846]: I0202 13:54:10.477025 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" event={"ID":"816eb531-874d-4167-8351-ceef3cd2704a","Type":"ContainerStarted","Data":"58f5c808ac3e7fd77328ccc7bff456a177cce1845a6613ffef68682949a3facf"} Feb 02 13:54:11 crc kubenswrapper[4846]: I0202 13:54:11.440603 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" path="/var/lib/kubelet/pods/510040f2-b874-48d8-bfb1-871b8abf0199/volumes" Feb 02 13:54:11 crc kubenswrapper[4846]: I0202 13:54:11.489675 4846 generic.go:334] "Generic (PLEG): container finished" podID="816eb531-874d-4167-8351-ceef3cd2704a" containerID="4f3a95a6f215190ec4ecd151772d2e0448650178e8ef8323660fbb961da4d6fc" exitCode=0 Feb 02 13:54:11 crc kubenswrapper[4846]: I0202 13:54:11.489723 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" event={"ID":"816eb531-874d-4167-8351-ceef3cd2704a","Type":"ContainerDied","Data":"4f3a95a6f215190ec4ecd151772d2e0448650178e8ef8323660fbb961da4d6fc"} Feb 02 13:54:13 crc kubenswrapper[4846]: I0202 13:54:13.521792 4846 generic.go:334] "Generic (PLEG): container finished" podID="816eb531-874d-4167-8351-ceef3cd2704a" containerID="cbc0110e17b4b083c46d92c5f325b42a6f79e8600b91f0d759ea662c7e520ccb" exitCode=0 Feb 02 13:54:13 crc kubenswrapper[4846]: I0202 13:54:13.521847 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" event={"ID":"816eb531-874d-4167-8351-ceef3cd2704a","Type":"ContainerDied","Data":"cbc0110e17b4b083c46d92c5f325b42a6f79e8600b91f0d759ea662c7e520ccb"} Feb 02 13:54:14 crc kubenswrapper[4846]: I0202 13:54:14.539209 4846 generic.go:334] "Generic (PLEG): container finished" podID="816eb531-874d-4167-8351-ceef3cd2704a" containerID="df81ec2290d2a1fe230be5a6e742cc5ae357c0f894de2f67c0401e34b689de0f" exitCode=0 Feb 02 13:54:14 crc kubenswrapper[4846]: I0202 13:54:14.539315 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" event={"ID":"816eb531-874d-4167-8351-ceef3cd2704a","Type":"ContainerDied","Data":"df81ec2290d2a1fe230be5a6e742cc5ae357c0f894de2f67c0401e34b689de0f"} Feb 02 13:54:15 crc kubenswrapper[4846]: I0202 13:54:15.950722 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.132578 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd747\" (UniqueName: \"kubernetes.io/projected/816eb531-874d-4167-8351-ceef3cd2704a-kube-api-access-xd747\") pod \"816eb531-874d-4167-8351-ceef3cd2704a\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.133786 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-util\") pod \"816eb531-874d-4167-8351-ceef3cd2704a\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.134033 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-bundle\") pod \"816eb531-874d-4167-8351-ceef3cd2704a\" (UID: \"816eb531-874d-4167-8351-ceef3cd2704a\") " Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.136431 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-bundle" (OuterVolumeSpecName: "bundle") pod "816eb531-874d-4167-8351-ceef3cd2704a" (UID: "816eb531-874d-4167-8351-ceef3cd2704a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.141498 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-util" (OuterVolumeSpecName: "util") pod "816eb531-874d-4167-8351-ceef3cd2704a" (UID: "816eb531-874d-4167-8351-ceef3cd2704a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.143827 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/816eb531-874d-4167-8351-ceef3cd2704a-kube-api-access-xd747" (OuterVolumeSpecName: "kube-api-access-xd747") pod "816eb531-874d-4167-8351-ceef3cd2704a" (UID: "816eb531-874d-4167-8351-ceef3cd2704a"). InnerVolumeSpecName "kube-api-access-xd747". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.237089 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-util\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.237116 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/816eb531-874d-4167-8351-ceef3cd2704a-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.237126 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd747\" (UniqueName: \"kubernetes.io/projected/816eb531-874d-4167-8351-ceef3cd2704a-kube-api-access-xd747\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.562108 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" event={"ID":"816eb531-874d-4167-8351-ceef3cd2704a","Type":"ContainerDied","Data":"58f5c808ac3e7fd77328ccc7bff456a177cce1845a6613ffef68682949a3facf"} Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.562150 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58f5c808ac3e7fd77328ccc7bff456a177cce1845a6613ffef68682949a3facf" Feb 02 13:54:16 crc kubenswrapper[4846]: I0202 13:54:16.562223 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.167166 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k"] Feb 02 13:54:26 crc kubenswrapper[4846]: E0202 13:54:26.168160 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="816eb531-874d-4167-8351-ceef3cd2704a" containerName="pull" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.168177 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="816eb531-874d-4167-8351-ceef3cd2704a" containerName="pull" Feb 02 13:54:26 crc kubenswrapper[4846]: E0202 13:54:26.168191 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.168198 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon" Feb 02 13:54:26 crc kubenswrapper[4846]: E0202 13:54:26.168210 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon-log" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.168218 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon-log" Feb 02 13:54:26 crc kubenswrapper[4846]: E0202 13:54:26.168232 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="816eb531-874d-4167-8351-ceef3cd2704a" containerName="extract" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.168239 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="816eb531-874d-4167-8351-ceef3cd2704a" containerName="extract" Feb 02 13:54:26 crc kubenswrapper[4846]: E0202 13:54:26.168253 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="816eb531-874d-4167-8351-ceef3cd2704a" containerName="util" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.168260 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="816eb531-874d-4167-8351-ceef3cd2704a" containerName="util" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.168532 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon-log" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.168558 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="510040f2-b874-48d8-bfb1-871b8abf0199" containerName="horizon" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.168572 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="816eb531-874d-4167-8351-ceef3cd2704a" containerName="extract" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.169373 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.171285 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.171614 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.171784 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-2dhpj" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.180142 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k"] Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.247114 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqz9p\" (UniqueName: \"kubernetes.io/projected/80118cc5-e446-4a77-9b65-8d50ac60308f-kube-api-access-bqz9p\") pod \"obo-prometheus-operator-68bc856cb9-4d59k\" (UID: \"80118cc5-e446-4a77-9b65-8d50ac60308f\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.304714 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt"] Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.306126 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.310965 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-tlgnl" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.311053 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.312894 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt"] Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.314168 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.333102 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt"] Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.347984 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt"] Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.348997 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd224d4b-1d4c-403e-89e9-b7fda056a631-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt\" (UID: \"dd224d4b-1d4c-403e-89e9-b7fda056a631\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.349087 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqz9p\" (UniqueName: \"kubernetes.io/projected/80118cc5-e446-4a77-9b65-8d50ac60308f-kube-api-access-bqz9p\") pod \"obo-prometheus-operator-68bc856cb9-4d59k\" (UID: \"80118cc5-e446-4a77-9b65-8d50ac60308f\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.349117 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd224d4b-1d4c-403e-89e9-b7fda056a631-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt\" (UID: \"dd224d4b-1d4c-403e-89e9-b7fda056a631\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.368849 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqz9p\" (UniqueName: \"kubernetes.io/projected/80118cc5-e446-4a77-9b65-8d50ac60308f-kube-api-access-bqz9p\") pod \"obo-prometheus-operator-68bc856cb9-4d59k\" (UID: \"80118cc5-e446-4a77-9b65-8d50ac60308f\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.451169 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/87119693-d4c2-4f7c-bbc1-d73409d91508-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt\" (UID: \"87119693-d4c2-4f7c-bbc1-d73409d91508\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.451384 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd224d4b-1d4c-403e-89e9-b7fda056a631-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt\" (UID: \"dd224d4b-1d4c-403e-89e9-b7fda056a631\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.451461 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/87119693-d4c2-4f7c-bbc1-d73409d91508-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt\" (UID: \"87119693-d4c2-4f7c-bbc1-d73409d91508\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.451710 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd224d4b-1d4c-403e-89e9-b7fda056a631-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt\" (UID: \"dd224d4b-1d4c-403e-89e9-b7fda056a631\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.456088 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dd224d4b-1d4c-403e-89e9-b7fda056a631-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt\" (UID: \"dd224d4b-1d4c-403e-89e9-b7fda056a631\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.456271 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dd224d4b-1d4c-403e-89e9-b7fda056a631-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt\" (UID: \"dd224d4b-1d4c-403e-89e9-b7fda056a631\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.489914 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.539419 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-2fxhf"] Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.542084 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.546781 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-cbwfl" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.549451 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.554025 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/87119693-d4c2-4f7c-bbc1-d73409d91508-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt\" (UID: \"87119693-d4c2-4f7c-bbc1-d73409d91508\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.554408 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/87119693-d4c2-4f7c-bbc1-d73409d91508-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt\" (UID: \"87119693-d4c2-4f7c-bbc1-d73409d91508\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.558352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/87119693-d4c2-4f7c-bbc1-d73409d91508-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt\" (UID: \"87119693-d4c2-4f7c-bbc1-d73409d91508\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.560913 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/87119693-d4c2-4f7c-bbc1-d73409d91508-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt\" (UID: \"87119693-d4c2-4f7c-bbc1-d73409d91508\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.563730 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-2fxhf"] Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.635120 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.646077 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.655862 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l24xp\" (UniqueName: \"kubernetes.io/projected/ca5f1eb1-41d2-4db2-8f80-953bada112f4-kube-api-access-l24xp\") pod \"observability-operator-59bdc8b94-2fxhf\" (UID: \"ca5f1eb1-41d2-4db2-8f80-953bada112f4\") " pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.656016 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/ca5f1eb1-41d2-4db2-8f80-953bada112f4-observability-operator-tls\") pod \"observability-operator-59bdc8b94-2fxhf\" (UID: \"ca5f1eb1-41d2-4db2-8f80-953bada112f4\") " pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.762750 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l24xp\" (UniqueName: \"kubernetes.io/projected/ca5f1eb1-41d2-4db2-8f80-953bada112f4-kube-api-access-l24xp\") pod \"observability-operator-59bdc8b94-2fxhf\" (UID: \"ca5f1eb1-41d2-4db2-8f80-953bada112f4\") " pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.762940 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/ca5f1eb1-41d2-4db2-8f80-953bada112f4-observability-operator-tls\") pod \"observability-operator-59bdc8b94-2fxhf\" (UID: \"ca5f1eb1-41d2-4db2-8f80-953bada112f4\") " pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.770486 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-cb9p7"] Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.783039 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-cb9p7"] Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.783163 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.784974 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/ca5f1eb1-41d2-4db2-8f80-953bada112f4-observability-operator-tls\") pod \"observability-operator-59bdc8b94-2fxhf\" (UID: \"ca5f1eb1-41d2-4db2-8f80-953bada112f4\") " pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.798173 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-bpm9z" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.819896 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l24xp\" (UniqueName: \"kubernetes.io/projected/ca5f1eb1-41d2-4db2-8f80-953bada112f4-kube-api-access-l24xp\") pod \"observability-operator-59bdc8b94-2fxhf\" (UID: \"ca5f1eb1-41d2-4db2-8f80-953bada112f4\") " pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.865846 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c7fb2f0a-18d3-46ff-a910-de6aa48e71eb-openshift-service-ca\") pod \"perses-operator-5bf474d74f-cb9p7\" (UID: \"c7fb2f0a-18d3-46ff-a910-de6aa48e71eb\") " pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.866090 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7v2v\" (UniqueName: \"kubernetes.io/projected/c7fb2f0a-18d3-46ff-a910-de6aa48e71eb-kube-api-access-h7v2v\") pod \"perses-operator-5bf474d74f-cb9p7\" (UID: \"c7fb2f0a-18d3-46ff-a910-de6aa48e71eb\") " pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.971888 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7v2v\" (UniqueName: \"kubernetes.io/projected/c7fb2f0a-18d3-46ff-a910-de6aa48e71eb-kube-api-access-h7v2v\") pod \"perses-operator-5bf474d74f-cb9p7\" (UID: \"c7fb2f0a-18d3-46ff-a910-de6aa48e71eb\") " pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.971964 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c7fb2f0a-18d3-46ff-a910-de6aa48e71eb-openshift-service-ca\") pod \"perses-operator-5bf474d74f-cb9p7\" (UID: \"c7fb2f0a-18d3-46ff-a910-de6aa48e71eb\") " pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" Feb 02 13:54:26 crc kubenswrapper[4846]: I0202 13:54:26.973097 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c7fb2f0a-18d3-46ff-a910-de6aa48e71eb-openshift-service-ca\") pod \"perses-operator-5bf474d74f-cb9p7\" (UID: \"c7fb2f0a-18d3-46ff-a910-de6aa48e71eb\") " pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.003187 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.003415 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7v2v\" (UniqueName: \"kubernetes.io/projected/c7fb2f0a-18d3-46ff-a910-de6aa48e71eb-kube-api-access-h7v2v\") pod \"perses-operator-5bf474d74f-cb9p7\" (UID: \"c7fb2f0a-18d3-46ff-a910-de6aa48e71eb\") " pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.125098 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k"] Feb 02 13:54:27 crc kubenswrapper[4846]: W0202 13:54:27.137155 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80118cc5_e446_4a77_9b65_8d50ac60308f.slice/crio-e5dd4c14ad4cdc6248fca655c66b4e42649e4c17d6e021376e5eb25c58fa69dd WatchSource:0}: Error finding container e5dd4c14ad4cdc6248fca655c66b4e42649e4c17d6e021376e5eb25c58fa69dd: Status 404 returned error can't find the container with id e5dd4c14ad4cdc6248fca655c66b4e42649e4c17d6e021376e5eb25c58fa69dd Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.141208 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.388348 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt"] Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.530777 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt"] Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.631681 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-2fxhf"] Feb 02 13:54:27 crc kubenswrapper[4846]: W0202 13:54:27.645389 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca5f1eb1_41d2_4db2_8f80_953bada112f4.slice/crio-7a001d0b4112c517f41a00206381808d305e47a26d8a19ebc0791c7e9fa5a258 WatchSource:0}: Error finding container 7a001d0b4112c517f41a00206381808d305e47a26d8a19ebc0791c7e9fa5a258: Status 404 returned error can't find the container with id 7a001d0b4112c517f41a00206381808d305e47a26d8a19ebc0791c7e9fa5a258 Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.682996 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" event={"ID":"87119693-d4c2-4f7c-bbc1-d73409d91508","Type":"ContainerStarted","Data":"2c72097f2bb572215e54d2df75f045d8284bb2ed91e7a34e3c37271d2f7d9e71"} Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.690568 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" event={"ID":"ca5f1eb1-41d2-4db2-8f80-953bada112f4","Type":"ContainerStarted","Data":"7a001d0b4112c517f41a00206381808d305e47a26d8a19ebc0791c7e9fa5a258"} Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.696836 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" event={"ID":"dd224d4b-1d4c-403e-89e9-b7fda056a631","Type":"ContainerStarted","Data":"9be74704cc15db848c25bfac4d672197adb0c79a84ad4e02c38d941623baff0b"} Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.698466 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k" event={"ID":"80118cc5-e446-4a77-9b65-8d50ac60308f","Type":"ContainerStarted","Data":"e5dd4c14ad4cdc6248fca655c66b4e42649e4c17d6e021376e5eb25c58fa69dd"} Feb 02 13:54:27 crc kubenswrapper[4846]: W0202 13:54:27.733302 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7fb2f0a_18d3_46ff_a910_de6aa48e71eb.slice/crio-aaabbd55c486e259bf84370e0f58942cf99c9a345e36af1b2405534054e1fd8b WatchSource:0}: Error finding container aaabbd55c486e259bf84370e0f58942cf99c9a345e36af1b2405534054e1fd8b: Status 404 returned error can't find the container with id aaabbd55c486e259bf84370e0f58942cf99c9a345e36af1b2405534054e1fd8b Feb 02 13:54:27 crc kubenswrapper[4846]: I0202 13:54:27.733653 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-cb9p7"] Feb 02 13:54:28 crc kubenswrapper[4846]: I0202 13:54:28.723321 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" event={"ID":"c7fb2f0a-18d3-46ff-a910-de6aa48e71eb","Type":"ContainerStarted","Data":"aaabbd55c486e259bf84370e0f58942cf99c9a345e36af1b2405534054e1fd8b"} Feb 02 13:54:30 crc kubenswrapper[4846]: I0202 13:54:30.480387 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:54:30 crc kubenswrapper[4846]: I0202 13:54:30.480734 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.850661 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" event={"ID":"87119693-d4c2-4f7c-bbc1-d73409d91508","Type":"ContainerStarted","Data":"b80fb682b3430d13ab9b87a62c732ddeeae56647fd50401ee7e74d117a703eff"} Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.858741 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" event={"ID":"ca5f1eb1-41d2-4db2-8f80-953bada112f4","Type":"ContainerStarted","Data":"366ebbe7e2188623d649f2c2abd81628564427c9392902ef5112254031ec0575"} Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.859062 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.862995 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" event={"ID":"dd224d4b-1d4c-403e-89e9-b7fda056a631","Type":"ContainerStarted","Data":"86db7e54d93a5f6453487d99cc5bb0844e47b83b7d894948627610aa9b86be14"} Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.866637 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k" event={"ID":"80118cc5-e446-4a77-9b65-8d50ac60308f","Type":"ContainerStarted","Data":"191a4cd80c3ee43c2aca997f337fff850a5e5c8c2cb104acca81148ef696cca3"} Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.872916 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" event={"ID":"c7fb2f0a-18d3-46ff-a910-de6aa48e71eb","Type":"ContainerStarted","Data":"5dd40aff1def5fb7fdc4194922f5b6fb6562156d4099b63ff99e782899c5edbb"} Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.873152 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.874744 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt" podStartSLOduration=2.318450053 podStartE2EDuration="12.874727976s" podCreationTimestamp="2026-02-02 13:54:26 +0000 UTC" firstStartedPulling="2026-02-02 13:54:27.524983854 +0000 UTC m=+6298.753570717" lastFinishedPulling="2026-02-02 13:54:38.081261777 +0000 UTC m=+6309.309848640" observedRunningTime="2026-02-02 13:54:38.870526814 +0000 UTC m=+6310.099113687" watchObservedRunningTime="2026-02-02 13:54:38.874727976 +0000 UTC m=+6310.103314839" Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.909899 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-4d59k" podStartSLOduration=1.998759371 podStartE2EDuration="12.909876429s" podCreationTimestamp="2026-02-02 13:54:26 +0000 UTC" firstStartedPulling="2026-02-02 13:54:27.141906052 +0000 UTC m=+6298.370492905" lastFinishedPulling="2026-02-02 13:54:38.05302311 +0000 UTC m=+6309.281609963" observedRunningTime="2026-02-02 13:54:38.902022689 +0000 UTC m=+6310.130609552" watchObservedRunningTime="2026-02-02 13:54:38.909876429 +0000 UTC m=+6310.138463292" Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.940869 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" Feb 02 13:54:38 crc kubenswrapper[4846]: I0202 13:54:38.959422 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-2fxhf" podStartSLOduration=2.5492770289999998 podStartE2EDuration="12.959400972s" podCreationTimestamp="2026-02-02 13:54:26 +0000 UTC" firstStartedPulling="2026-02-02 13:54:27.647712485 +0000 UTC m=+6298.876299348" lastFinishedPulling="2026-02-02 13:54:38.057836428 +0000 UTC m=+6309.286423291" observedRunningTime="2026-02-02 13:54:38.947641706 +0000 UTC m=+6310.176228589" watchObservedRunningTime="2026-02-02 13:54:38.959400972 +0000 UTC m=+6310.187987845" Feb 02 13:54:39 crc kubenswrapper[4846]: I0202 13:54:39.002255 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt" podStartSLOduration=2.351406673 podStartE2EDuration="13.002230302s" podCreationTimestamp="2026-02-02 13:54:26 +0000 UTC" firstStartedPulling="2026-02-02 13:54:27.401969636 +0000 UTC m=+6298.630556499" lastFinishedPulling="2026-02-02 13:54:38.052793265 +0000 UTC m=+6309.281380128" observedRunningTime="2026-02-02 13:54:38.974761355 +0000 UTC m=+6310.203348238" watchObservedRunningTime="2026-02-02 13:54:39.002230302 +0000 UTC m=+6310.230817165" Feb 02 13:54:40 crc kubenswrapper[4846]: I0202 13:54:40.315664 4846 scope.go:117] "RemoveContainer" containerID="7b29da07bbab88e9e8a5d6b1cb8e79cc7b27f452040d70b52362efa8845f14a9" Feb 02 13:54:40 crc kubenswrapper[4846]: I0202 13:54:40.351823 4846 scope.go:117] "RemoveContainer" containerID="4197896df10ef7e7b0b43c79de5dc6cb20f0090a9950f89e32264e20708bb932" Feb 02 13:54:40 crc kubenswrapper[4846]: I0202 13:54:40.424046 4846 scope.go:117] "RemoveContainer" containerID="3482dc67a33cb3f7e8b5da214b372820fb55bcf723a7053e78d52e2e5d2be8f8" Feb 02 13:54:47 crc kubenswrapper[4846]: I0202 13:54:47.144212 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" Feb 02 13:54:47 crc kubenswrapper[4846]: I0202 13:54:47.176916 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-cb9p7" podStartSLOduration=10.860035099 podStartE2EDuration="21.176896817s" podCreationTimestamp="2026-02-02 13:54:26 +0000 UTC" firstStartedPulling="2026-02-02 13:54:27.736586803 +0000 UTC m=+6298.965173666" lastFinishedPulling="2026-02-02 13:54:38.053448511 +0000 UTC m=+6309.282035384" observedRunningTime="2026-02-02 13:54:39.052117054 +0000 UTC m=+6310.280703917" watchObservedRunningTime="2026-02-02 13:54:47.176896817 +0000 UTC m=+6318.405483680" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.328144 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.328977 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="143a8135-2e86-46cf-a748-b7429af615df" containerName="openstackclient" containerID="cri-o://495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd" gracePeriod=2 Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.337855 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.394298 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 02 13:54:50 crc kubenswrapper[4846]: E0202 13:54:50.394856 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="143a8135-2e86-46cf-a748-b7429af615df" containerName="openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.394882 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="143a8135-2e86-46cf-a748-b7429af615df" containerName="openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.395147 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="143a8135-2e86-46cf-a748-b7429af615df" containerName="openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.396117 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.406700 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.426962 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="143a8135-2e86-46cf-a748-b7429af615df" podUID="0f95d11b-ef19-4d01-b872-c43a03da644f" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.449095 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.449215 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhgk6\" (UniqueName: \"kubernetes.io/projected/0f95d11b-ef19-4d01-b872-c43a03da644f-kube-api-access-lhgk6\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.449275 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config-secret\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.449344 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.545226 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.548443 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.551241 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config-secret\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.551316 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.551432 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.551545 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhgk6\" (UniqueName: \"kubernetes.io/projected/0f95d11b-ef19-4d01-b872-c43a03da644f-kube-api-access-lhgk6\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.552453 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.560072 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config-secret\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.561962 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-n94kk" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.571843 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.627129 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhgk6\" (UniqueName: \"kubernetes.io/projected/0f95d11b-ef19-4d01-b872-c43a03da644f-kube-api-access-lhgk6\") pod \"openstackclient\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.627195 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.661436 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n47t6\" (UniqueName: \"kubernetes.io/projected/6b9d958f-3aab-4199-ad20-6fb2b318b7b4-kube-api-access-n47t6\") pod \"kube-state-metrics-0\" (UID: \"6b9d958f-3aab-4199-ad20-6fb2b318b7b4\") " pod="openstack/kube-state-metrics-0" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.725594 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.766903 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n47t6\" (UniqueName: \"kubernetes.io/projected/6b9d958f-3aab-4199-ad20-6fb2b318b7b4-kube-api-access-n47t6\") pod \"kube-state-metrics-0\" (UID: \"6b9d958f-3aab-4199-ad20-6fb2b318b7b4\") " pod="openstack/kube-state-metrics-0" Feb 02 13:54:50 crc kubenswrapper[4846]: I0202 13:54:50.834617 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n47t6\" (UniqueName: \"kubernetes.io/projected/6b9d958f-3aab-4199-ad20-6fb2b318b7b4-kube-api-access-n47t6\") pod \"kube-state-metrics-0\" (UID: \"6b9d958f-3aab-4199-ad20-6fb2b318b7b4\") " pod="openstack/kube-state-metrics-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.008982 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.117708 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-zb62r"] Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.159341 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b982-account-create-update-6p2b9"] Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.231708 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b982-account-create-update-6p2b9"] Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.353692 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-zb62r"] Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.489180 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="083851d9-5552-40fb-8b8f-0185a23e1f36" path="/var/lib/kubelet/pods/083851d9-5552-40fb-8b8f-0185a23e1f36/volumes" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.495218 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="752948fa-aec6-4754-baa0-c04dcd8dfb94" path="/var/lib/kubelet/pods/752948fa-aec6-4754-baa0-c04dcd8dfb94/volumes" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.533474 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.536850 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.545065 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.545257 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.545360 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.545457 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.556121 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-w8llp" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.612860 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.624965 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/ec35752b-6884-4a8e-b077-43cf63dd98c3-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.625093 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ec35752b-6884-4a8e-b077-43cf63dd98c3-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.625298 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ec35752b-6884-4a8e-b077-43cf63dd98c3-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.625323 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ec35752b-6884-4a8e-b077-43cf63dd98c3-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.625375 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khmpx\" (UniqueName: \"kubernetes.io/projected/ec35752b-6884-4a8e-b077-43cf63dd98c3-kube-api-access-khmpx\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.625402 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ec35752b-6884-4a8e-b077-43cf63dd98c3-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.625428 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ec35752b-6884-4a8e-b077-43cf63dd98c3-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.699276 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.728051 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ec35752b-6884-4a8e-b077-43cf63dd98c3-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.728125 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ec35752b-6884-4a8e-b077-43cf63dd98c3-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.728197 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/ec35752b-6884-4a8e-b077-43cf63dd98c3-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.728251 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ec35752b-6884-4a8e-b077-43cf63dd98c3-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.728387 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ec35752b-6884-4a8e-b077-43cf63dd98c3-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.728405 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ec35752b-6884-4a8e-b077-43cf63dd98c3-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.728457 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khmpx\" (UniqueName: \"kubernetes.io/projected/ec35752b-6884-4a8e-b077-43cf63dd98c3-kube-api-access-khmpx\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.730091 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ec35752b-6884-4a8e-b077-43cf63dd98c3-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.757565 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ec35752b-6884-4a8e-b077-43cf63dd98c3-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.767127 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/ec35752b-6884-4a8e-b077-43cf63dd98c3-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.769258 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ec35752b-6884-4a8e-b077-43cf63dd98c3-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.786095 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ec35752b-6884-4a8e-b077-43cf63dd98c3-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.786113 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ec35752b-6884-4a8e-b077-43cf63dd98c3-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:51 crc kubenswrapper[4846]: I0202 13:54:51.790532 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khmpx\" (UniqueName: \"kubernetes.io/projected/ec35752b-6884-4a8e-b077-43cf63dd98c3-kube-api-access-khmpx\") pod \"alertmanager-metric-storage-0\" (UID: \"ec35752b-6884-4a8e-b077-43cf63dd98c3\") " pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.016305 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.029157 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.049589 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.049847 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.049957 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.049971 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.050218 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.050350 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.050459 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.050578 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-nxnnq" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.050688 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.072459 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0f95d11b-ef19-4d01-b872-c43a03da644f","Type":"ContainerStarted","Data":"766c6b84438cf97215b24d07fcddbe99d51d25d47fa95ddb073bfac4fbb88e46"} Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.128367 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.174298 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fvs4\" (UniqueName: \"kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-kube-api-access-8fvs4\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.174382 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.174439 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.174514 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.174651 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.174681 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.174698 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca054c99-a1e7-4053-836e-d8885339d8be-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.174719 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.174873 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.174938 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.277385 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.277747 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.277780 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fvs4\" (UniqueName: \"kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-kube-api-access-8fvs4\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.277808 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.277842 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.277882 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.277944 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.278050 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.278097 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca054c99-a1e7-4053-836e-d8885339d8be-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.278122 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.279078 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.282643 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.283108 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.290343 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.291267 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-config\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.308605 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.308674 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3de1d7d06bd12c18bafded6cf43a4c4cec5027afab93e80aaa955b1f1f001fe0/globalmount\"" pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.310421 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca054c99-a1e7-4053-836e-d8885339d8be-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.318255 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.323780 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.325289 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fvs4\" (UniqueName: \"kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-kube-api-access-8fvs4\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.378521 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") pod \"prometheus-metric-storage-0\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: W0202 13:54:52.566995 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b9d958f_3aab_4199_ad20_6fb2b318b7b4.slice/crio-e9a7763118d06405b7e3ec081bcf72b9b6236520d9266c785abef69e0da8cafd WatchSource:0}: Error finding container e9a7763118d06405b7e3ec081bcf72b9b6236520d9266c785abef69e0da8cafd: Status 404 returned error can't find the container with id e9a7763118d06405b7e3ec081bcf72b9b6236520d9266c785abef69e0da8cafd Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.571679 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.682408 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.838552 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.897706 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.899558 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-combined-ca-bundle\") pod \"143a8135-2e86-46cf-a748-b7429af615df\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.899654 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-openstack-config-secret\") pod \"143a8135-2e86-46cf-a748-b7429af615df\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.899757 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/143a8135-2e86-46cf-a748-b7429af615df-openstack-config\") pod \"143a8135-2e86-46cf-a748-b7429af615df\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.899795 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scw2m\" (UniqueName: \"kubernetes.io/projected/143a8135-2e86-46cf-a748-b7429af615df-kube-api-access-scw2m\") pod \"143a8135-2e86-46cf-a748-b7429af615df\" (UID: \"143a8135-2e86-46cf-a748-b7429af615df\") " Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.905348 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/143a8135-2e86-46cf-a748-b7429af615df-kube-api-access-scw2m" (OuterVolumeSpecName: "kube-api-access-scw2m") pod "143a8135-2e86-46cf-a748-b7429af615df" (UID: "143a8135-2e86-46cf-a748-b7429af615df"). InnerVolumeSpecName "kube-api-access-scw2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.981912 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "143a8135-2e86-46cf-a748-b7429af615df" (UID: "143a8135-2e86-46cf-a748-b7429af615df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.983014 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/143a8135-2e86-46cf-a748-b7429af615df-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "143a8135-2e86-46cf-a748-b7429af615df" (UID: "143a8135-2e86-46cf-a748-b7429af615df"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:54:52 crc kubenswrapper[4846]: I0202 13:54:52.992254 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "143a8135-2e86-46cf-a748-b7429af615df" (UID: "143a8135-2e86-46cf-a748-b7429af615df"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.002728 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.002767 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/143a8135-2e86-46cf-a748-b7429af615df-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.002777 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scw2m\" (UniqueName: \"kubernetes.io/projected/143a8135-2e86-46cf-a748-b7429af615df-kube-api-access-scw2m\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.002785 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/143a8135-2e86-46cf-a748-b7429af615df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.099236 4846 generic.go:334] "Generic (PLEG): container finished" podID="143a8135-2e86-46cf-a748-b7429af615df" containerID="495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd" exitCode=137 Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.099435 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.099515 4846 scope.go:117] "RemoveContainer" containerID="495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.146100 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0f95d11b-ef19-4d01-b872-c43a03da644f","Type":"ContainerStarted","Data":"a87b5cbd9387dd85627b9ba6e6c08c08b5b8eefa4464afbe2bfaf0fdd49115eb"} Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.148065 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="143a8135-2e86-46cf-a748-b7429af615df" podUID="0f95d11b-ef19-4d01-b872-c43a03da644f" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.160687 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ec35752b-6884-4a8e-b077-43cf63dd98c3","Type":"ContainerStarted","Data":"5846278d12fb35e7f9b32500d3b02b7fb38c04b8ee7a4a35888aa7683de50b52"} Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.165978 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6b9d958f-3aab-4199-ad20-6fb2b318b7b4","Type":"ContainerStarted","Data":"e9a7763118d06405b7e3ec081bcf72b9b6236520d9266c785abef69e0da8cafd"} Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.180317 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.180301412 podStartE2EDuration="3.180301412s" podCreationTimestamp="2026-02-02 13:54:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:54:53.179204495 +0000 UTC m=+6324.407791358" watchObservedRunningTime="2026-02-02 13:54:53.180301412 +0000 UTC m=+6324.408888275" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.193181 4846 scope.go:117] "RemoveContainer" containerID="495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd" Feb 02 13:54:53 crc kubenswrapper[4846]: E0202 13:54:53.193597 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd\": container with ID starting with 495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd not found: ID does not exist" containerID="495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.193638 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd"} err="failed to get container status \"495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd\": rpc error: code = NotFound desc = could not find container \"495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd\": container with ID starting with 495f7a75aa50ff96e5cf0f620f0c792730667863b05a604e658a0c5fc56f8bfd not found: ID does not exist" Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.319426 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 02 13:54:53 crc kubenswrapper[4846]: I0202 13:54:53.436894 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="143a8135-2e86-46cf-a748-b7429af615df" path="/var/lib/kubelet/pods/143a8135-2e86-46cf-a748-b7429af615df/volumes" Feb 02 13:54:54 crc kubenswrapper[4846]: I0202 13:54:54.238753 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca054c99-a1e7-4053-836e-d8885339d8be","Type":"ContainerStarted","Data":"b52f9dec7e94e04c885265ab5aafa976cdea681d1343654f126ef0a6f86ee32b"} Feb 02 13:54:54 crc kubenswrapper[4846]: I0202 13:54:54.250908 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6b9d958f-3aab-4199-ad20-6fb2b318b7b4","Type":"ContainerStarted","Data":"c743bebd24bbf13a9176e62e172e920485827218baf086194a3ab262d2fef121"} Feb 02 13:54:54 crc kubenswrapper[4846]: I0202 13:54:54.251036 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 02 13:54:54 crc kubenswrapper[4846]: I0202 13:54:54.290788 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.85306352 podStartE2EDuration="4.2907672s" podCreationTimestamp="2026-02-02 13:54:50 +0000 UTC" firstStartedPulling="2026-02-02 13:54:52.569387135 +0000 UTC m=+6323.797973998" lastFinishedPulling="2026-02-02 13:54:53.007090815 +0000 UTC m=+6324.235677678" observedRunningTime="2026-02-02 13:54:54.27515212 +0000 UTC m=+6325.503738973" watchObservedRunningTime="2026-02-02 13:54:54.2907672 +0000 UTC m=+6325.519354063" Feb 02 13:54:59 crc kubenswrapper[4846]: I0202 13:54:59.039496 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-9mpqm"] Feb 02 13:54:59 crc kubenswrapper[4846]: I0202 13:54:59.051671 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-9mpqm"] Feb 02 13:54:59 crc kubenswrapper[4846]: I0202 13:54:59.439999 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="526b1562-1fa5-4ca0-80a5-738f483d37a4" path="/var/lib/kubelet/pods/526b1562-1fa5-4ca0-80a5-738f483d37a4/volumes" Feb 02 13:55:00 crc kubenswrapper[4846]: I0202 13:55:00.314054 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca054c99-a1e7-4053-836e-d8885339d8be","Type":"ContainerStarted","Data":"72a87804fa0c2907b3114a379966ea349582ef4b4707af2f7d1ed28b99b370ba"} Feb 02 13:55:00 crc kubenswrapper[4846]: I0202 13:55:00.316611 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ec35752b-6884-4a8e-b077-43cf63dd98c3","Type":"ContainerStarted","Data":"bd18c90f02afdf0c9e6c5193f1fafab50bdbc76f7ab30d9a521dcaebecf3a693"} Feb 02 13:55:00 crc kubenswrapper[4846]: I0202 13:55:00.479297 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:55:00 crc kubenswrapper[4846]: I0202 13:55:00.479387 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:55:01 crc kubenswrapper[4846]: I0202 13:55:01.015022 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 02 13:55:07 crc kubenswrapper[4846]: I0202 13:55:07.397420 4846 generic.go:334] "Generic (PLEG): container finished" podID="ca054c99-a1e7-4053-836e-d8885339d8be" containerID="72a87804fa0c2907b3114a379966ea349582ef4b4707af2f7d1ed28b99b370ba" exitCode=0 Feb 02 13:55:07 crc kubenswrapper[4846]: I0202 13:55:07.397496 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca054c99-a1e7-4053-836e-d8885339d8be","Type":"ContainerDied","Data":"72a87804fa0c2907b3114a379966ea349582ef4b4707af2f7d1ed28b99b370ba"} Feb 02 13:55:07 crc kubenswrapper[4846]: I0202 13:55:07.401213 4846 generic.go:334] "Generic (PLEG): container finished" podID="ec35752b-6884-4a8e-b077-43cf63dd98c3" containerID="bd18c90f02afdf0c9e6c5193f1fafab50bdbc76f7ab30d9a521dcaebecf3a693" exitCode=0 Feb 02 13:55:07 crc kubenswrapper[4846]: I0202 13:55:07.401248 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ec35752b-6884-4a8e-b077-43cf63dd98c3","Type":"ContainerDied","Data":"bd18c90f02afdf0c9e6c5193f1fafab50bdbc76f7ab30d9a521dcaebecf3a693"} Feb 02 13:55:10 crc kubenswrapper[4846]: I0202 13:55:10.450532 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ec35752b-6884-4a8e-b077-43cf63dd98c3","Type":"ContainerStarted","Data":"0d4cde2ce49a4448adeca4191efd7ef5267eab88effd9056319486db19f22e76"} Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.093091 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-glb7j"] Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.096457 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.109028 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-glb7j"] Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.147055 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-utilities\") pod \"community-operators-glb7j\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.147455 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-catalog-content\") pod \"community-operators-glb7j\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.147488 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd8z2\" (UniqueName: \"kubernetes.io/projected/1335710e-9a15-48cb-88c5-05abe9fc8c06-kube-api-access-nd8z2\") pod \"community-operators-glb7j\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.250356 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-utilities\") pod \"community-operators-glb7j\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.250487 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-catalog-content\") pod \"community-operators-glb7j\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.250524 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd8z2\" (UniqueName: \"kubernetes.io/projected/1335710e-9a15-48cb-88c5-05abe9fc8c06-kube-api-access-nd8z2\") pod \"community-operators-glb7j\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.250848 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-utilities\") pod \"community-operators-glb7j\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.250947 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-catalog-content\") pod \"community-operators-glb7j\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.275210 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd8z2\" (UniqueName: \"kubernetes.io/projected/1335710e-9a15-48cb-88c5-05abe9fc8c06-kube-api-access-nd8z2\") pod \"community-operators-glb7j\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.424530 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.547917 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ec35752b-6884-4a8e-b077-43cf63dd98c3","Type":"ContainerStarted","Data":"2023427f95b104e834d531ff30b25def215082b63d0e41abac26e3d9276a27c5"} Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.549790 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.555318 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Feb 02 13:55:15 crc kubenswrapper[4846]: I0202 13:55:15.602866 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=8.030069544 podStartE2EDuration="24.602847018s" podCreationTimestamp="2026-02-02 13:54:51 +0000 UTC" firstStartedPulling="2026-02-02 13:54:52.953703608 +0000 UTC m=+6324.182290471" lastFinishedPulling="2026-02-02 13:55:09.526481082 +0000 UTC m=+6340.755067945" observedRunningTime="2026-02-02 13:55:15.588743145 +0000 UTC m=+6346.817330008" watchObservedRunningTime="2026-02-02 13:55:15.602847018 +0000 UTC m=+6346.831433881" Feb 02 13:55:16 crc kubenswrapper[4846]: I0202 13:55:16.048757 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-glb7j"] Feb 02 13:55:16 crc kubenswrapper[4846]: I0202 13:55:16.556972 4846 generic.go:334] "Generic (PLEG): container finished" podID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerID="8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8" exitCode=0 Feb 02 13:55:16 crc kubenswrapper[4846]: I0202 13:55:16.557112 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glb7j" event={"ID":"1335710e-9a15-48cb-88c5-05abe9fc8c06","Type":"ContainerDied","Data":"8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8"} Feb 02 13:55:16 crc kubenswrapper[4846]: I0202 13:55:16.557163 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glb7j" event={"ID":"1335710e-9a15-48cb-88c5-05abe9fc8c06","Type":"ContainerStarted","Data":"9087a14a819d10549ccd0f15acf397c4f1d40741499f3cb7276833daa5e3598f"} Feb 02 13:55:16 crc kubenswrapper[4846]: I0202 13:55:16.561284 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca054c99-a1e7-4053-836e-d8885339d8be","Type":"ContainerStarted","Data":"fa2146e165ee0b4a4794e850c1d8284aad9e92232791742c9c4645eb688755e2"} Feb 02 13:55:17 crc kubenswrapper[4846]: I0202 13:55:17.573611 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glb7j" event={"ID":"1335710e-9a15-48cb-88c5-05abe9fc8c06","Type":"ContainerStarted","Data":"70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50"} Feb 02 13:55:19 crc kubenswrapper[4846]: I0202 13:55:19.596679 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca054c99-a1e7-4053-836e-d8885339d8be","Type":"ContainerStarted","Data":"77791c53aa0b512fb62c1ca3a69712de58b1c4ad4b7905c39e4ea3006d1b2322"} Feb 02 13:55:19 crc kubenswrapper[4846]: I0202 13:55:19.601063 4846 generic.go:334] "Generic (PLEG): container finished" podID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerID="70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50" exitCode=0 Feb 02 13:55:19 crc kubenswrapper[4846]: I0202 13:55:19.601107 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glb7j" event={"ID":"1335710e-9a15-48cb-88c5-05abe9fc8c06","Type":"ContainerDied","Data":"70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50"} Feb 02 13:55:20 crc kubenswrapper[4846]: I0202 13:55:20.617260 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glb7j" event={"ID":"1335710e-9a15-48cb-88c5-05abe9fc8c06","Type":"ContainerStarted","Data":"ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a"} Feb 02 13:55:20 crc kubenswrapper[4846]: I0202 13:55:20.652603 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-glb7j" podStartSLOduration=2.21835436 podStartE2EDuration="5.652580392s" podCreationTimestamp="2026-02-02 13:55:15 +0000 UTC" firstStartedPulling="2026-02-02 13:55:16.559291655 +0000 UTC m=+6347.787878528" lastFinishedPulling="2026-02-02 13:55:19.993517697 +0000 UTC m=+6351.222104560" observedRunningTime="2026-02-02 13:55:20.652055109 +0000 UTC m=+6351.880641972" watchObservedRunningTime="2026-02-02 13:55:20.652580392 +0000 UTC m=+6351.881167255" Feb 02 13:55:22 crc kubenswrapper[4846]: I0202 13:55:22.644349 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca054c99-a1e7-4053-836e-d8885339d8be","Type":"ContainerStarted","Data":"e98c4779f4d20d59db4f36e72779316924807e3e6d0cfae27d860d20945f6a02"} Feb 02 13:55:22 crc kubenswrapper[4846]: I0202 13:55:22.670534 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.795197915 podStartE2EDuration="32.670506718s" podCreationTimestamp="2026-02-02 13:54:50 +0000 UTC" firstStartedPulling="2026-02-02 13:54:53.327264171 +0000 UTC m=+6324.555851034" lastFinishedPulling="2026-02-02 13:55:22.202572974 +0000 UTC m=+6353.431159837" observedRunningTime="2026-02-02 13:55:22.6677061 +0000 UTC m=+6353.896292983" watchObservedRunningTime="2026-02-02 13:55:22.670506718 +0000 UTC m=+6353.899093601" Feb 02 13:55:22 crc kubenswrapper[4846]: I0202 13:55:22.683242 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:22 crc kubenswrapper[4846]: I0202 13:55:22.683450 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:22 crc kubenswrapper[4846]: I0202 13:55:22.686196 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:23 crc kubenswrapper[4846]: I0202 13:55:23.656676 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.438137 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.438567 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.529294 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.529575 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="0f95d11b-ef19-4d01-b872-c43a03da644f" containerName="openstackclient" containerID="cri-o://a87b5cbd9387dd85627b9ba6e6c08c08b5b8eefa4464afbe2bfaf0fdd49115eb" gracePeriod=2 Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.542522 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.561872 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 02 13:55:25 crc kubenswrapper[4846]: E0202 13:55:25.562536 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f95d11b-ef19-4d01-b872-c43a03da644f" containerName="openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.562561 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f95d11b-ef19-4d01-b872-c43a03da644f" containerName="openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.562885 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f95d11b-ef19-4d01-b872-c43a03da644f" containerName="openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.563936 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.578315 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.586877 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="0f95d11b-ef19-4d01-b872-c43a03da644f" podUID="b1f840c7-4bd5-427a-9f4b-b4fec6280665" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.589192 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b1f840c7-4bd5-427a-9f4b-b4fec6280665-openstack-config-secret\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.589250 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b1f840c7-4bd5-427a-9f4b-b4fec6280665-openstack-config\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.589318 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f840c7-4bd5-427a-9f4b-b4fec6280665-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.590646 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-779xn\" (UniqueName: \"kubernetes.io/projected/b1f840c7-4bd5-427a-9f4b-b4fec6280665-kube-api-access-779xn\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.692446 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b1f840c7-4bd5-427a-9f4b-b4fec6280665-openstack-config-secret\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.692818 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b1f840c7-4bd5-427a-9f4b-b4fec6280665-openstack-config\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.692990 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f840c7-4bd5-427a-9f4b-b4fec6280665-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.693238 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-779xn\" (UniqueName: \"kubernetes.io/projected/b1f840c7-4bd5-427a-9f4b-b4fec6280665-kube-api-access-779xn\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.693555 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b1f840c7-4bd5-427a-9f4b-b4fec6280665-openstack-config\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.700288 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b1f840c7-4bd5-427a-9f4b-b4fec6280665-openstack-config-secret\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.700387 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f840c7-4bd5-427a-9f4b-b4fec6280665-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.711124 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-779xn\" (UniqueName: \"kubernetes.io/projected/b1f840c7-4bd5-427a-9f4b-b4fec6280665-kube-api-access-779xn\") pod \"openstackclient\" (UID: \"b1f840c7-4bd5-427a-9f4b-b4fec6280665\") " pod="openstack/openstackclient" Feb 02 13:55:25 crc kubenswrapper[4846]: I0202 13:55:25.889193 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 13:55:26 crc kubenswrapper[4846]: I0202 13:55:26.476166 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 02 13:55:26 crc kubenswrapper[4846]: I0202 13:55:26.481099 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-glb7j" podUID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerName="registry-server" probeResult="failure" output=< Feb 02 13:55:26 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 13:55:26 crc kubenswrapper[4846]: > Feb 02 13:55:26 crc kubenswrapper[4846]: I0202 13:55:26.682866 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b1f840c7-4bd5-427a-9f4b-b4fec6280665","Type":"ContainerStarted","Data":"65ca67c0a9faa20f5e5a78c0e64f650710449647073f0b0b953cc430a4db2f23"} Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.053022 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.053358 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="prometheus" containerID="cri-o://fa2146e165ee0b4a4794e850c1d8284aad9e92232791742c9c4645eb688755e2" gracePeriod=600 Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.053418 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="thanos-sidecar" containerID="cri-o://e98c4779f4d20d59db4f36e72779316924807e3e6d0cfae27d860d20945f6a02" gracePeriod=600 Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.053538 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="config-reloader" containerID="cri-o://77791c53aa0b512fb62c1ca3a69712de58b1c4ad4b7905c39e4ea3006d1b2322" gracePeriod=600 Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.733382 4846 generic.go:334] "Generic (PLEG): container finished" podID="ca054c99-a1e7-4053-836e-d8885339d8be" containerID="e98c4779f4d20d59db4f36e72779316924807e3e6d0cfae27d860d20945f6a02" exitCode=0 Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.736008 4846 generic.go:334] "Generic (PLEG): container finished" podID="ca054c99-a1e7-4053-836e-d8885339d8be" containerID="77791c53aa0b512fb62c1ca3a69712de58b1c4ad4b7905c39e4ea3006d1b2322" exitCode=0 Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.736171 4846 generic.go:334] "Generic (PLEG): container finished" podID="ca054c99-a1e7-4053-836e-d8885339d8be" containerID="fa2146e165ee0b4a4794e850c1d8284aad9e92232791742c9c4645eb688755e2" exitCode=0 Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.734313 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca054c99-a1e7-4053-836e-d8885339d8be","Type":"ContainerDied","Data":"e98c4779f4d20d59db4f36e72779316924807e3e6d0cfae27d860d20945f6a02"} Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.736738 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca054c99-a1e7-4053-836e-d8885339d8be","Type":"ContainerDied","Data":"77791c53aa0b512fb62c1ca3a69712de58b1c4ad4b7905c39e4ea3006d1b2322"} Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.736774 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca054c99-a1e7-4053-836e-d8885339d8be","Type":"ContainerDied","Data":"fa2146e165ee0b4a4794e850c1d8284aad9e92232791742c9c4645eb688755e2"} Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.747962 4846 generic.go:334] "Generic (PLEG): container finished" podID="0f95d11b-ef19-4d01-b872-c43a03da644f" containerID="a87b5cbd9387dd85627b9ba6e6c08c08b5b8eefa4464afbe2bfaf0fdd49115eb" exitCode=137 Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.749114 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b1f840c7-4bd5-427a-9f4b-b4fec6280665","Type":"ContainerStarted","Data":"f3008d0f83b696e8698dd1069bbc1440bb2fc12bef5550060110bafede91c5be"} Feb 02 13:55:27 crc kubenswrapper[4846]: I0202 13:55:27.802434 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.802392557 podStartE2EDuration="2.802392557s" podCreationTimestamp="2026-02-02 13:55:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:55:27.798064532 +0000 UTC m=+6359.026651415" watchObservedRunningTime="2026-02-02 13:55:27.802392557 +0000 UTC m=+6359.030979420" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.077042 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.092221 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.279358 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-0\") pod \"ca054c99-a1e7-4053-836e-d8885339d8be\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280040 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-thanos-prometheus-http-client-file\") pod \"ca054c99-a1e7-4053-836e-d8885339d8be\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280307 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") pod \"ca054c99-a1e7-4053-836e-d8885339d8be\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280361 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-web-config\") pod \"ca054c99-a1e7-4053-836e-d8885339d8be\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280353 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "ca054c99-a1e7-4053-836e-d8885339d8be" (UID: "ca054c99-a1e7-4053-836e-d8885339d8be"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280469 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhgk6\" (UniqueName: \"kubernetes.io/projected/0f95d11b-ef19-4d01-b872-c43a03da644f-kube-api-access-lhgk6\") pod \"0f95d11b-ef19-4d01-b872-c43a03da644f\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280490 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-config\") pod \"ca054c99-a1e7-4053-836e-d8885339d8be\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280575 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config-secret\") pod \"0f95d11b-ef19-4d01-b872-c43a03da644f\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280597 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-tls-assets\") pod \"ca054c99-a1e7-4053-836e-d8885339d8be\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280705 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-2\") pod \"ca054c99-a1e7-4053-836e-d8885339d8be\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280748 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fvs4\" (UniqueName: \"kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-kube-api-access-8fvs4\") pod \"ca054c99-a1e7-4053-836e-d8885339d8be\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280797 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config\") pod \"0f95d11b-ef19-4d01-b872-c43a03da644f\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280890 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca054c99-a1e7-4053-836e-d8885339d8be-config-out\") pod \"ca054c99-a1e7-4053-836e-d8885339d8be\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.280924 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-1\") pod \"ca054c99-a1e7-4053-836e-d8885339d8be\" (UID: \"ca054c99-a1e7-4053-836e-d8885339d8be\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.281010 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-combined-ca-bundle\") pod \"0f95d11b-ef19-4d01-b872-c43a03da644f\" (UID: \"0f95d11b-ef19-4d01-b872-c43a03da644f\") " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.282492 4846 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.284643 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "ca054c99-a1e7-4053-836e-d8885339d8be" (UID: "ca054c99-a1e7-4053-836e-d8885339d8be"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.284874 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "ca054c99-a1e7-4053-836e-d8885339d8be" (UID: "ca054c99-a1e7-4053-836e-d8885339d8be"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.288460 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca054c99-a1e7-4053-836e-d8885339d8be-config-out" (OuterVolumeSpecName: "config-out") pod "ca054c99-a1e7-4053-836e-d8885339d8be" (UID: "ca054c99-a1e7-4053-836e-d8885339d8be"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.290766 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-config" (OuterVolumeSpecName: "config") pod "ca054c99-a1e7-4053-836e-d8885339d8be" (UID: "ca054c99-a1e7-4053-836e-d8885339d8be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.290968 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-kube-api-access-8fvs4" (OuterVolumeSpecName: "kube-api-access-8fvs4") pod "ca054c99-a1e7-4053-836e-d8885339d8be" (UID: "ca054c99-a1e7-4053-836e-d8885339d8be"). InnerVolumeSpecName "kube-api-access-8fvs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.305806 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "ca054c99-a1e7-4053-836e-d8885339d8be" (UID: "ca054c99-a1e7-4053-836e-d8885339d8be"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.315695 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "ca054c99-a1e7-4053-836e-d8885339d8be" (UID: "ca054c99-a1e7-4053-836e-d8885339d8be"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.317801 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f95d11b-ef19-4d01-b872-c43a03da644f-kube-api-access-lhgk6" (OuterVolumeSpecName: "kube-api-access-lhgk6") pod "0f95d11b-ef19-4d01-b872-c43a03da644f" (UID: "0f95d11b-ef19-4d01-b872-c43a03da644f"). InnerVolumeSpecName "kube-api-access-lhgk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.324475 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f95d11b-ef19-4d01-b872-c43a03da644f" (UID: "0f95d11b-ef19-4d01-b872-c43a03da644f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.344319 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-web-config" (OuterVolumeSpecName: "web-config") pod "ca054c99-a1e7-4053-836e-d8885339d8be" (UID: "ca054c99-a1e7-4053-836e-d8885339d8be"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.345913 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "ca054c99-a1e7-4053-836e-d8885339d8be" (UID: "ca054c99-a1e7-4053-836e-d8885339d8be"). InnerVolumeSpecName "pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.350404 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0f95d11b-ef19-4d01-b872-c43a03da644f" (UID: "0f95d11b-ef19-4d01-b872-c43a03da644f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384537 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") on node \"crc\" " Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384594 4846 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-web-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384611 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhgk6\" (UniqueName: \"kubernetes.io/projected/0f95d11b-ef19-4d01-b872-c43a03da644f-kube-api-access-lhgk6\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384652 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384668 4846 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-tls-assets\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384682 4846 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384696 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fvs4\" (UniqueName: \"kubernetes.io/projected/ca054c99-a1e7-4053-836e-d8885339d8be-kube-api-access-8fvs4\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384712 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384725 4846 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ca054c99-a1e7-4053-836e-d8885339d8be-config-out\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384739 4846 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ca054c99-a1e7-4053-836e-d8885339d8be-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384752 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.384766 4846 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ca054c99-a1e7-4053-836e-d8885339d8be-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.397099 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0f95d11b-ef19-4d01-b872-c43a03da644f" (UID: "0f95d11b-ef19-4d01-b872-c43a03da644f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.417584 4846 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.417825 4846 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515") on node "crc" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.487003 4846 reconciler_common.go:293] "Volume detached for volume \"pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.487047 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0f95d11b-ef19-4d01-b872-c43a03da644f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.769570 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ca054c99-a1e7-4053-836e-d8885339d8be","Type":"ContainerDied","Data":"b52f9dec7e94e04c885265ab5aafa976cdea681d1343654f126ef0a6f86ee32b"} Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.769755 4846 scope.go:117] "RemoveContainer" containerID="e98c4779f4d20d59db4f36e72779316924807e3e6d0cfae27d860d20945f6a02" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.769773 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.773475 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.816311 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="0f95d11b-ef19-4d01-b872-c43a03da644f" podUID="b1f840c7-4bd5-427a-9f4b-b4fec6280665" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.837910 4846 scope.go:117] "RemoveContainer" containerID="77791c53aa0b512fb62c1ca3a69712de58b1c4ad4b7905c39e4ea3006d1b2322" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.845600 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.856100 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.874192 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 02 13:55:28 crc kubenswrapper[4846]: E0202 13:55:28.874736 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="init-config-reloader" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.874763 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="init-config-reloader" Feb 02 13:55:28 crc kubenswrapper[4846]: E0202 13:55:28.874784 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="config-reloader" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.874794 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="config-reloader" Feb 02 13:55:28 crc kubenswrapper[4846]: E0202 13:55:28.874822 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="prometheus" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.874829 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="prometheus" Feb 02 13:55:28 crc kubenswrapper[4846]: E0202 13:55:28.874838 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="thanos-sidecar" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.874845 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="thanos-sidecar" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.875064 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="prometheus" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.875079 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="config-reloader" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.875099 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="thanos-sidecar" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.876972 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.880425 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.880837 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.881032 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.881216 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.881401 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.881775 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.883536 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-nxnnq" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.884995 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.887981 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.905288 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.914881 4846 scope.go:117] "RemoveContainer" containerID="fa2146e165ee0b4a4794e850c1d8284aad9e92232791742c9c4645eb688755e2" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.951273 4846 scope.go:117] "RemoveContainer" containerID="72a87804fa0c2907b3114a379966ea349582ef4b4707af2f7d1ed28b99b370ba" Feb 02 13:55:28 crc kubenswrapper[4846]: I0202 13:55:28.981877 4846 scope.go:117] "RemoveContainer" containerID="a87b5cbd9387dd85627b9ba6e6c08c08b5b8eefa4464afbe2bfaf0fdd49115eb" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.002432 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92f8c952-e300-4858-9bc5-c06350e65d6f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.002487 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92f8c952-e300-4858-9bc5-c06350e65d6f-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.002535 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-config\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.002690 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.002973 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vdgj\" (UniqueName: \"kubernetes.io/projected/92f8c952-e300-4858-9bc5-c06350e65d6f-kube-api-access-4vdgj\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.003130 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.003282 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.003343 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92f8c952-e300-4858-9bc5-c06350e65d6f-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.003401 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92f8c952-e300-4858-9bc5-c06350e65d6f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.003519 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92f8c952-e300-4858-9bc5-c06350e65d6f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.003802 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.004065 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.004135 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.051104 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c6fe-account-create-update-jbkbk"] Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.061442 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-gvd6t"] Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.071590 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c6fe-account-create-update-jbkbk"] Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.080837 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-gvd6t"] Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.107048 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vdgj\" (UniqueName: \"kubernetes.io/projected/92f8c952-e300-4858-9bc5-c06350e65d6f-kube-api-access-4vdgj\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.107132 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.107240 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.107305 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92f8c952-e300-4858-9bc5-c06350e65d6f-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.107328 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92f8c952-e300-4858-9bc5-c06350e65d6f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.107401 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92f8c952-e300-4858-9bc5-c06350e65d6f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.107447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.107522 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.107552 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.107906 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92f8c952-e300-4858-9bc5-c06350e65d6f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.108085 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92f8c952-e300-4858-9bc5-c06350e65d6f-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.108174 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-config\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.109147 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92f8c952-e300-4858-9bc5-c06350e65d6f-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.109373 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92f8c952-e300-4858-9bc5-c06350e65d6f-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.109668 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.110008 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.110032 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3de1d7d06bd12c18bafded6cf43a4c4cec5027afab93e80aaa955b1f1f001fe0/globalmount\"" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.112869 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92f8c952-e300-4858-9bc5-c06350e65d6f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.113125 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.113934 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-config\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.114770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92f8c952-e300-4858-9bc5-c06350e65d6f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.115249 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.115856 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.117226 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.118961 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92f8c952-e300-4858-9bc5-c06350e65d6f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.129260 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92f8c952-e300-4858-9bc5-c06350e65d6f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.138607 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vdgj\" (UniqueName: \"kubernetes.io/projected/92f8c952-e300-4858-9bc5-c06350e65d6f-kube-api-access-4vdgj\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.174106 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a38818e1-d653-4ae9-ac1e-54a7dc4fd515\") pod \"prometheus-metric-storage-0\" (UID: \"92f8c952-e300-4858-9bc5-c06350e65d6f\") " pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.203572 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.473863 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f95d11b-ef19-4d01-b872-c43a03da644f" path="/var/lib/kubelet/pods/0f95d11b-ef19-4d01-b872-c43a03da644f/volumes" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.475323 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4008e772-8019-48ec-8fe1-e49445af9f72" path="/var/lib/kubelet/pods/4008e772-8019-48ec-8fe1-e49445af9f72/volumes" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.480061 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" path="/var/lib/kubelet/pods/ca054c99-a1e7-4053-836e-d8885339d8be/volumes" Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.492284 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0fbe008-d516-47f7-ada7-b2e588ae3db4" path="/var/lib/kubelet/pods/d0fbe008-d516-47f7-ada7-b2e588ae3db4/volumes" Feb 02 13:55:29 crc kubenswrapper[4846]: W0202 13:55:29.747169 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92f8c952_e300_4858_9bc5_c06350e65d6f.slice/crio-2b3bd8907b28aa926da150c5bdcdaf1cca88a7e3aa37b37edeb97c054a9e111e WatchSource:0}: Error finding container 2b3bd8907b28aa926da150c5bdcdaf1cca88a7e3aa37b37edeb97c054a9e111e: Status 404 returned error can't find the container with id 2b3bd8907b28aa926da150c5bdcdaf1cca88a7e3aa37b37edeb97c054a9e111e Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.771467 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 02 13:55:29 crc kubenswrapper[4846]: I0202 13:55:29.795631 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92f8c952-e300-4858-9bc5-c06350e65d6f","Type":"ContainerStarted","Data":"2b3bd8907b28aa926da150c5bdcdaf1cca88a7e3aa37b37edeb97c054a9e111e"} Feb 02 13:55:30 crc kubenswrapper[4846]: I0202 13:55:30.479580 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 13:55:30 crc kubenswrapper[4846]: I0202 13:55:30.480172 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 13:55:30 crc kubenswrapper[4846]: I0202 13:55:30.480233 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 13:55:30 crc kubenswrapper[4846]: I0202 13:55:30.481407 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 13:55:30 crc kubenswrapper[4846]: I0202 13:55:30.481485 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" gracePeriod=600 Feb 02 13:55:30 crc kubenswrapper[4846]: E0202 13:55:30.665426 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:55:30 crc kubenswrapper[4846]: I0202 13:55:30.684823 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="ca054c99-a1e7-4053-836e-d8885339d8be" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.1.145:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 13:55:30 crc kubenswrapper[4846]: I0202 13:55:30.809885 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" exitCode=0 Feb 02 13:55:30 crc kubenswrapper[4846]: I0202 13:55:30.809926 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab"} Feb 02 13:55:30 crc kubenswrapper[4846]: I0202 13:55:30.809961 4846 scope.go:117] "RemoveContainer" containerID="fcae4c79abc512d4d8a5d4e3219b87abc67dad0f182ce8895f815c1ebd096f68" Feb 02 13:55:30 crc kubenswrapper[4846]: I0202 13:55:30.810767 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:55:30 crc kubenswrapper[4846]: E0202 13:55:30.811110 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.305430 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.309110 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.312218 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.312219 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.324788 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.459112 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-scripts\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.459499 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.459526 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-log-httpd\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.459583 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.459664 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsh6v\" (UniqueName: \"kubernetes.io/projected/2c1ac007-e978-4ff3-8fec-30310b62f5fc-kube-api-access-dsh6v\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.459699 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-config-data\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.459943 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-run-httpd\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.563265 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-scripts\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.563344 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.563370 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-log-httpd\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.563461 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.563511 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsh6v\" (UniqueName: \"kubernetes.io/projected/2c1ac007-e978-4ff3-8fec-30310b62f5fc-kube-api-access-dsh6v\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.563569 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-config-data\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.563711 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-run-httpd\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.564839 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-log-httpd\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.564882 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-run-httpd\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.570564 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.571310 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.572520 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-scripts\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.581701 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-config-data\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.586769 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsh6v\" (UniqueName: \"kubernetes.io/projected/2c1ac007-e978-4ff3-8fec-30310b62f5fc-kube-api-access-dsh6v\") pod \"ceilometer-0\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " pod="openstack/ceilometer-0" Feb 02 13:55:31 crc kubenswrapper[4846]: I0202 13:55:31.636048 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 13:55:32 crc kubenswrapper[4846]: I0202 13:55:32.145481 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:55:32 crc kubenswrapper[4846]: W0202 13:55:32.151173 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c1ac007_e978_4ff3_8fec_30310b62f5fc.slice/crio-32e132f338edd530dfcb842a71b769a4d8d2950b91c9b0dfb047994302f76ebf WatchSource:0}: Error finding container 32e132f338edd530dfcb842a71b769a4d8d2950b91c9b0dfb047994302f76ebf: Status 404 returned error can't find the container with id 32e132f338edd530dfcb842a71b769a4d8d2950b91c9b0dfb047994302f76ebf Feb 02 13:55:32 crc kubenswrapper[4846]: I0202 13:55:32.841726 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c1ac007-e978-4ff3-8fec-30310b62f5fc","Type":"ContainerStarted","Data":"32e132f338edd530dfcb842a71b769a4d8d2950b91c9b0dfb047994302f76ebf"} Feb 02 13:55:33 crc kubenswrapper[4846]: I0202 13:55:33.855808 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c1ac007-e978-4ff3-8fec-30310b62f5fc","Type":"ContainerStarted","Data":"af04bb887a71994e4781459d3133e5f42a8ba589ab2401aa883d1fab2619ab52"} Feb 02 13:55:33 crc kubenswrapper[4846]: I0202 13:55:33.856361 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c1ac007-e978-4ff3-8fec-30310b62f5fc","Type":"ContainerStarted","Data":"9c8d3f52a82434fd7412c182fc8d512d20dda21c64ecb81c027e230e6257efe8"} Feb 02 13:55:34 crc kubenswrapper[4846]: I0202 13:55:34.869651 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c1ac007-e978-4ff3-8fec-30310b62f5fc","Type":"ContainerStarted","Data":"9223b8d6f428658de474c93db680d9ead607ae4c5198c8d5e31d2fbd95ed6283"} Feb 02 13:55:34 crc kubenswrapper[4846]: I0202 13:55:34.871521 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92f8c952-e300-4858-9bc5-c06350e65d6f","Type":"ContainerStarted","Data":"e2aac9d09e4194f81dd0aac4dbb401c4f1d661e26dfc16ebe60f035f4b78712d"} Feb 02 13:55:35 crc kubenswrapper[4846]: I0202 13:55:35.041745 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-5dsdp"] Feb 02 13:55:35 crc kubenswrapper[4846]: I0202 13:55:35.054182 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-5dsdp"] Feb 02 13:55:35 crc kubenswrapper[4846]: I0202 13:55:35.440250 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59524ffd-8aca-4e7c-bef3-e838e3312129" path="/var/lib/kubelet/pods/59524ffd-8aca-4e7c-bef3-e838e3312129/volumes" Feb 02 13:55:35 crc kubenswrapper[4846]: I0202 13:55:35.501313 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:35 crc kubenswrapper[4846]: I0202 13:55:35.570261 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:35 crc kubenswrapper[4846]: I0202 13:55:35.740569 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-glb7j"] Feb 02 13:55:36 crc kubenswrapper[4846]: I0202 13:55:36.892475 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-glb7j" podUID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerName="registry-server" containerID="cri-o://ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a" gracePeriod=2 Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.415488 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.547347 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-catalog-content\") pod \"1335710e-9a15-48cb-88c5-05abe9fc8c06\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.547911 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd8z2\" (UniqueName: \"kubernetes.io/projected/1335710e-9a15-48cb-88c5-05abe9fc8c06-kube-api-access-nd8z2\") pod \"1335710e-9a15-48cb-88c5-05abe9fc8c06\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.547990 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-utilities\") pod \"1335710e-9a15-48cb-88c5-05abe9fc8c06\" (UID: \"1335710e-9a15-48cb-88c5-05abe9fc8c06\") " Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.548966 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-utilities" (OuterVolumeSpecName: "utilities") pod "1335710e-9a15-48cb-88c5-05abe9fc8c06" (UID: "1335710e-9a15-48cb-88c5-05abe9fc8c06"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.553446 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1335710e-9a15-48cb-88c5-05abe9fc8c06-kube-api-access-nd8z2" (OuterVolumeSpecName: "kube-api-access-nd8z2") pod "1335710e-9a15-48cb-88c5-05abe9fc8c06" (UID: "1335710e-9a15-48cb-88c5-05abe9fc8c06"). InnerVolumeSpecName "kube-api-access-nd8z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.625968 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1335710e-9a15-48cb-88c5-05abe9fc8c06" (UID: "1335710e-9a15-48cb-88c5-05abe9fc8c06"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.650336 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.650381 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd8z2\" (UniqueName: \"kubernetes.io/projected/1335710e-9a15-48cb-88c5-05abe9fc8c06-kube-api-access-nd8z2\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.650395 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1335710e-9a15-48cb-88c5-05abe9fc8c06-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.905332 4846 generic.go:334] "Generic (PLEG): container finished" podID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerID="ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a" exitCode=0 Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.905392 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-glb7j" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.905422 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glb7j" event={"ID":"1335710e-9a15-48cb-88c5-05abe9fc8c06","Type":"ContainerDied","Data":"ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a"} Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.905476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-glb7j" event={"ID":"1335710e-9a15-48cb-88c5-05abe9fc8c06","Type":"ContainerDied","Data":"9087a14a819d10549ccd0f15acf397c4f1d40741499f3cb7276833daa5e3598f"} Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.905500 4846 scope.go:117] "RemoveContainer" containerID="ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.911873 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c1ac007-e978-4ff3-8fec-30310b62f5fc","Type":"ContainerStarted","Data":"40a268cc4f98040ecf7197b7f236859697dd5ee20c1eefb00603aa283f68e085"} Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.936875 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.4338677300000002 podStartE2EDuration="6.936855436s" podCreationTimestamp="2026-02-02 13:55:31 +0000 UTC" firstStartedPulling="2026-02-02 13:55:32.154144531 +0000 UTC m=+6363.382731394" lastFinishedPulling="2026-02-02 13:55:36.657132237 +0000 UTC m=+6367.885719100" observedRunningTime="2026-02-02 13:55:37.931865705 +0000 UTC m=+6369.160452588" watchObservedRunningTime="2026-02-02 13:55:37.936855436 +0000 UTC m=+6369.165442299" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.952719 4846 scope.go:117] "RemoveContainer" containerID="70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50" Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.964462 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-glb7j"] Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.976964 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-glb7j"] Feb 02 13:55:37 crc kubenswrapper[4846]: I0202 13:55:37.992090 4846 scope.go:117] "RemoveContainer" containerID="8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8" Feb 02 13:55:38 crc kubenswrapper[4846]: I0202 13:55:38.028212 4846 scope.go:117] "RemoveContainer" containerID="ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a" Feb 02 13:55:38 crc kubenswrapper[4846]: E0202 13:55:38.028970 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a\": container with ID starting with ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a not found: ID does not exist" containerID="ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a" Feb 02 13:55:38 crc kubenswrapper[4846]: I0202 13:55:38.029023 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a"} err="failed to get container status \"ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a\": rpc error: code = NotFound desc = could not find container \"ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a\": container with ID starting with ca353bf6f7c3f947d24441b5a1119b5d816ec9274a16b90e1d71bcc7ed98ba9a not found: ID does not exist" Feb 02 13:55:38 crc kubenswrapper[4846]: I0202 13:55:38.029054 4846 scope.go:117] "RemoveContainer" containerID="70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50" Feb 02 13:55:38 crc kubenswrapper[4846]: E0202 13:55:38.029392 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50\": container with ID starting with 70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50 not found: ID does not exist" containerID="70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50" Feb 02 13:55:38 crc kubenswrapper[4846]: I0202 13:55:38.029427 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50"} err="failed to get container status \"70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50\": rpc error: code = NotFound desc = could not find container \"70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50\": container with ID starting with 70ebca7b2a933b37b1b460972d495a3b53864996e83d728456ea18c121721f50 not found: ID does not exist" Feb 02 13:55:38 crc kubenswrapper[4846]: I0202 13:55:38.029451 4846 scope.go:117] "RemoveContainer" containerID="8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8" Feb 02 13:55:38 crc kubenswrapper[4846]: E0202 13:55:38.029744 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8\": container with ID starting with 8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8 not found: ID does not exist" containerID="8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8" Feb 02 13:55:38 crc kubenswrapper[4846]: I0202 13:55:38.029779 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8"} err="failed to get container status \"8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8\": rpc error: code = NotFound desc = could not find container \"8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8\": container with ID starting with 8f74a6a390ca3a6c464239c3640cf0f8b7c2e86d7ac5e26a56be3dd3f9d5a6f8 not found: ID does not exist" Feb 02 13:55:38 crc kubenswrapper[4846]: I0202 13:55:38.922425 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 13:55:39 crc kubenswrapper[4846]: I0202 13:55:39.446383 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1335710e-9a15-48cb-88c5-05abe9fc8c06" path="/var/lib/kubelet/pods/1335710e-9a15-48cb-88c5-05abe9fc8c06/volumes" Feb 02 13:55:40 crc kubenswrapper[4846]: I0202 13:55:40.584012 4846 scope.go:117] "RemoveContainer" containerID="78715439dbb4defd7ed7a34b9d10c49e8b1b8c0d3000953fbbad0f003ef1f97c" Feb 02 13:55:40 crc kubenswrapper[4846]: I0202 13:55:40.636766 4846 scope.go:117] "RemoveContainer" containerID="f2eee019dc41b815eed61d6e84b86bf012d759d7822a6245d408781e923c3808" Feb 02 13:55:40 crc kubenswrapper[4846]: I0202 13:55:40.702517 4846 scope.go:117] "RemoveContainer" containerID="533f3721cb689883a09ff8d5d8bfa088d718faca9befee0385dc08870b204ae0" Feb 02 13:55:40 crc kubenswrapper[4846]: I0202 13:55:40.792206 4846 scope.go:117] "RemoveContainer" containerID="2aac42116c042c9e7bf73001ac3c3955e6da7b89ddcb55571d22166145af7b33" Feb 02 13:55:40 crc kubenswrapper[4846]: I0202 13:55:40.858910 4846 scope.go:117] "RemoveContainer" containerID="16aac1532f4b216f5978f276eb1f5873e8a7988440c11f635eb0a7dd25e5988c" Feb 02 13:55:40 crc kubenswrapper[4846]: I0202 13:55:40.913925 4846 scope.go:117] "RemoveContainer" containerID="6a11abe8177ab0d52fe13b49d6bf7f43d078f1dc4b9c14f25da0f913939da49a" Feb 02 13:55:40 crc kubenswrapper[4846]: I0202 13:55:40.968808 4846 scope.go:117] "RemoveContainer" containerID="cc31acf04483afec48f846a544006a56933563b4e24576cc654cfdd3cb045103" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.018168 4846 scope.go:117] "RemoveContainer" containerID="5233047f1e8062b98adf89a4397f34bc1f322c3eb1830130d6478a35aba0da7a" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.779862 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-9knl4"] Feb 02 13:55:41 crc kubenswrapper[4846]: E0202 13:55:41.780389 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerName="registry-server" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.780409 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerName="registry-server" Feb 02 13:55:41 crc kubenswrapper[4846]: E0202 13:55:41.780426 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerName="extract-content" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.780448 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerName="extract-content" Feb 02 13:55:41 crc kubenswrapper[4846]: E0202 13:55:41.780467 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerName="extract-utilities" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.780476 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerName="extract-utilities" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.780769 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1335710e-9a15-48cb-88c5-05abe9fc8c06" containerName="registry-server" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.781731 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9knl4" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.796357 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-9knl4"] Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.900716 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-5b76-account-create-update-px22h"] Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.902879 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5b76-account-create-update-px22h" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.908808 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.918269 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-5b76-account-create-update-px22h"] Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.952114 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6530545-8859-43e2-baee-8a1021830ba0-operator-scripts\") pod \"aodh-db-create-9knl4\" (UID: \"d6530545-8859-43e2-baee-8a1021830ba0\") " pod="openstack/aodh-db-create-9knl4" Feb 02 13:55:41 crc kubenswrapper[4846]: I0202 13:55:41.952265 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k84z\" (UniqueName: \"kubernetes.io/projected/d6530545-8859-43e2-baee-8a1021830ba0-kube-api-access-4k84z\") pod \"aodh-db-create-9knl4\" (UID: \"d6530545-8859-43e2-baee-8a1021830ba0\") " pod="openstack/aodh-db-create-9knl4" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.005967 4846 generic.go:334] "Generic (PLEG): container finished" podID="92f8c952-e300-4858-9bc5-c06350e65d6f" containerID="e2aac9d09e4194f81dd0aac4dbb401c4f1d661e26dfc16ebe60f035f4b78712d" exitCode=0 Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.006038 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92f8c952-e300-4858-9bc5-c06350e65d6f","Type":"ContainerDied","Data":"e2aac9d09e4194f81dd0aac4dbb401c4f1d661e26dfc16ebe60f035f4b78712d"} Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.053591 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6530545-8859-43e2-baee-8a1021830ba0-operator-scripts\") pod \"aodh-db-create-9knl4\" (UID: \"d6530545-8859-43e2-baee-8a1021830ba0\") " pod="openstack/aodh-db-create-9knl4" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.053728 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k84z\" (UniqueName: \"kubernetes.io/projected/d6530545-8859-43e2-baee-8a1021830ba0-kube-api-access-4k84z\") pod \"aodh-db-create-9knl4\" (UID: \"d6530545-8859-43e2-baee-8a1021830ba0\") " pod="openstack/aodh-db-create-9knl4" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.053862 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtsdt\" (UniqueName: \"kubernetes.io/projected/07b19d98-9059-4596-8ffb-b0264d3a6f14-kube-api-access-qtsdt\") pod \"aodh-5b76-account-create-update-px22h\" (UID: \"07b19d98-9059-4596-8ffb-b0264d3a6f14\") " pod="openstack/aodh-5b76-account-create-update-px22h" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.053923 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07b19d98-9059-4596-8ffb-b0264d3a6f14-operator-scripts\") pod \"aodh-5b76-account-create-update-px22h\" (UID: \"07b19d98-9059-4596-8ffb-b0264d3a6f14\") " pod="openstack/aodh-5b76-account-create-update-px22h" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.054736 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6530545-8859-43e2-baee-8a1021830ba0-operator-scripts\") pod \"aodh-db-create-9knl4\" (UID: \"d6530545-8859-43e2-baee-8a1021830ba0\") " pod="openstack/aodh-db-create-9knl4" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.072446 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k84z\" (UniqueName: \"kubernetes.io/projected/d6530545-8859-43e2-baee-8a1021830ba0-kube-api-access-4k84z\") pod \"aodh-db-create-9knl4\" (UID: \"d6530545-8859-43e2-baee-8a1021830ba0\") " pod="openstack/aodh-db-create-9knl4" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.105292 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9knl4" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.158785 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtsdt\" (UniqueName: \"kubernetes.io/projected/07b19d98-9059-4596-8ffb-b0264d3a6f14-kube-api-access-qtsdt\") pod \"aodh-5b76-account-create-update-px22h\" (UID: \"07b19d98-9059-4596-8ffb-b0264d3a6f14\") " pod="openstack/aodh-5b76-account-create-update-px22h" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.158887 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07b19d98-9059-4596-8ffb-b0264d3a6f14-operator-scripts\") pod \"aodh-5b76-account-create-update-px22h\" (UID: \"07b19d98-9059-4596-8ffb-b0264d3a6f14\") " pod="openstack/aodh-5b76-account-create-update-px22h" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.160508 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07b19d98-9059-4596-8ffb-b0264d3a6f14-operator-scripts\") pod \"aodh-5b76-account-create-update-px22h\" (UID: \"07b19d98-9059-4596-8ffb-b0264d3a6f14\") " pod="openstack/aodh-5b76-account-create-update-px22h" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.186720 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtsdt\" (UniqueName: \"kubernetes.io/projected/07b19d98-9059-4596-8ffb-b0264d3a6f14-kube-api-access-qtsdt\") pod \"aodh-5b76-account-create-update-px22h\" (UID: \"07b19d98-9059-4596-8ffb-b0264d3a6f14\") " pod="openstack/aodh-5b76-account-create-update-px22h" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.243404 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5b76-account-create-update-px22h" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.426936 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:55:42 crc kubenswrapper[4846]: E0202 13:55:42.427672 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.670090 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-9knl4"] Feb 02 13:55:42 crc kubenswrapper[4846]: W0202 13:55:42.683173 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6530545_8859_43e2_baee_8a1021830ba0.slice/crio-d80f1484fa51509dc1231d9e93807be531dac8a0daa7d0ab955b34460ecbc239 WatchSource:0}: Error finding container d80f1484fa51509dc1231d9e93807be531dac8a0daa7d0ab955b34460ecbc239: Status 404 returned error can't find the container with id d80f1484fa51509dc1231d9e93807be531dac8a0daa7d0ab955b34460ecbc239 Feb 02 13:55:42 crc kubenswrapper[4846]: I0202 13:55:42.957946 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-5b76-account-create-update-px22h"] Feb 02 13:55:43 crc kubenswrapper[4846]: I0202 13:55:43.034842 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92f8c952-e300-4858-9bc5-c06350e65d6f","Type":"ContainerStarted","Data":"9101c146b3e2944a067ef882997d452ab58d0c373136df974f1a518a5a46f658"} Feb 02 13:55:43 crc kubenswrapper[4846]: I0202 13:55:43.036762 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-5b76-account-create-update-px22h" event={"ID":"07b19d98-9059-4596-8ffb-b0264d3a6f14","Type":"ContainerStarted","Data":"6a05510f3da2e7e739ab4d28b62ef34b0054b64717f7d7a32e099342a6029b94"} Feb 02 13:55:43 crc kubenswrapper[4846]: I0202 13:55:43.041236 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9knl4" event={"ID":"d6530545-8859-43e2-baee-8a1021830ba0","Type":"ContainerStarted","Data":"3904af06e1634518b3a299fd9e9e0b9b00f923de5c964d7b162a72f1c5f4dfc7"} Feb 02 13:55:43 crc kubenswrapper[4846]: I0202 13:55:43.041273 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9knl4" event={"ID":"d6530545-8859-43e2-baee-8a1021830ba0","Type":"ContainerStarted","Data":"d80f1484fa51509dc1231d9e93807be531dac8a0daa7d0ab955b34460ecbc239"} Feb 02 13:55:43 crc kubenswrapper[4846]: I0202 13:55:43.072484 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-9knl4" podStartSLOduration=2.072459435 podStartE2EDuration="2.072459435s" podCreationTimestamp="2026-02-02 13:55:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:55:43.063888047 +0000 UTC m=+6374.292474910" watchObservedRunningTime="2026-02-02 13:55:43.072459435 +0000 UTC m=+6374.301046298" Feb 02 13:55:44 crc kubenswrapper[4846]: I0202 13:55:44.052095 4846 generic.go:334] "Generic (PLEG): container finished" podID="d6530545-8859-43e2-baee-8a1021830ba0" containerID="3904af06e1634518b3a299fd9e9e0b9b00f923de5c964d7b162a72f1c5f4dfc7" exitCode=0 Feb 02 13:55:44 crc kubenswrapper[4846]: I0202 13:55:44.052156 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9knl4" event={"ID":"d6530545-8859-43e2-baee-8a1021830ba0","Type":"ContainerDied","Data":"3904af06e1634518b3a299fd9e9e0b9b00f923de5c964d7b162a72f1c5f4dfc7"} Feb 02 13:55:44 crc kubenswrapper[4846]: I0202 13:55:44.054658 4846 generic.go:334] "Generic (PLEG): container finished" podID="07b19d98-9059-4596-8ffb-b0264d3a6f14" containerID="7b9245f9504a28fe84a9b2e9505168d9ea88623785a553ee63db7b001d883ce5" exitCode=0 Feb 02 13:55:44 crc kubenswrapper[4846]: I0202 13:55:44.054719 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-5b76-account-create-update-px22h" event={"ID":"07b19d98-9059-4596-8ffb-b0264d3a6f14","Type":"ContainerDied","Data":"7b9245f9504a28fe84a9b2e9505168d9ea88623785a553ee63db7b001d883ce5"} Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.719070 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9knl4" Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.727818 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5b76-account-create-update-px22h" Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.864986 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4k84z\" (UniqueName: \"kubernetes.io/projected/d6530545-8859-43e2-baee-8a1021830ba0-kube-api-access-4k84z\") pod \"d6530545-8859-43e2-baee-8a1021830ba0\" (UID: \"d6530545-8859-43e2-baee-8a1021830ba0\") " Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.865094 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6530545-8859-43e2-baee-8a1021830ba0-operator-scripts\") pod \"d6530545-8859-43e2-baee-8a1021830ba0\" (UID: \"d6530545-8859-43e2-baee-8a1021830ba0\") " Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.865182 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07b19d98-9059-4596-8ffb-b0264d3a6f14-operator-scripts\") pod \"07b19d98-9059-4596-8ffb-b0264d3a6f14\" (UID: \"07b19d98-9059-4596-8ffb-b0264d3a6f14\") " Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.865403 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtsdt\" (UniqueName: \"kubernetes.io/projected/07b19d98-9059-4596-8ffb-b0264d3a6f14-kube-api-access-qtsdt\") pod \"07b19d98-9059-4596-8ffb-b0264d3a6f14\" (UID: \"07b19d98-9059-4596-8ffb-b0264d3a6f14\") " Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.865693 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6530545-8859-43e2-baee-8a1021830ba0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d6530545-8859-43e2-baee-8a1021830ba0" (UID: "d6530545-8859-43e2-baee-8a1021830ba0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.865949 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07b19d98-9059-4596-8ffb-b0264d3a6f14-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "07b19d98-9059-4596-8ffb-b0264d3a6f14" (UID: "07b19d98-9059-4596-8ffb-b0264d3a6f14"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.867505 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6530545-8859-43e2-baee-8a1021830ba0-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.867533 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07b19d98-9059-4596-8ffb-b0264d3a6f14-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.873291 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6530545-8859-43e2-baee-8a1021830ba0-kube-api-access-4k84z" (OuterVolumeSpecName: "kube-api-access-4k84z") pod "d6530545-8859-43e2-baee-8a1021830ba0" (UID: "d6530545-8859-43e2-baee-8a1021830ba0"). InnerVolumeSpecName "kube-api-access-4k84z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.879775 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07b19d98-9059-4596-8ffb-b0264d3a6f14-kube-api-access-qtsdt" (OuterVolumeSpecName: "kube-api-access-qtsdt") pod "07b19d98-9059-4596-8ffb-b0264d3a6f14" (UID: "07b19d98-9059-4596-8ffb-b0264d3a6f14"). InnerVolumeSpecName "kube-api-access-qtsdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.971139 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtsdt\" (UniqueName: \"kubernetes.io/projected/07b19d98-9059-4596-8ffb-b0264d3a6f14-kube-api-access-qtsdt\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:45 crc kubenswrapper[4846]: I0202 13:55:45.971190 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4k84z\" (UniqueName: \"kubernetes.io/projected/d6530545-8859-43e2-baee-8a1021830ba0-kube-api-access-4k84z\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:46 crc kubenswrapper[4846]: I0202 13:55:46.079482 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9knl4" Feb 02 13:55:46 crc kubenswrapper[4846]: I0202 13:55:46.079517 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9knl4" event={"ID":"d6530545-8859-43e2-baee-8a1021830ba0","Type":"ContainerDied","Data":"d80f1484fa51509dc1231d9e93807be531dac8a0daa7d0ab955b34460ecbc239"} Feb 02 13:55:46 crc kubenswrapper[4846]: I0202 13:55:46.079558 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d80f1484fa51509dc1231d9e93807be531dac8a0daa7d0ab955b34460ecbc239" Feb 02 13:55:46 crc kubenswrapper[4846]: I0202 13:55:46.081720 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-5b76-account-create-update-px22h" event={"ID":"07b19d98-9059-4596-8ffb-b0264d3a6f14","Type":"ContainerDied","Data":"6a05510f3da2e7e739ab4d28b62ef34b0054b64717f7d7a32e099342a6029b94"} Feb 02 13:55:46 crc kubenswrapper[4846]: I0202 13:55:46.081748 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a05510f3da2e7e739ab4d28b62ef34b0054b64717f7d7a32e099342a6029b94" Feb 02 13:55:46 crc kubenswrapper[4846]: I0202 13:55:46.081773 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5b76-account-create-update-px22h" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.107057 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92f8c952-e300-4858-9bc5-c06350e65d6f","Type":"ContainerStarted","Data":"2685690f3a7b8bc6af78e262f4214bbd6a3a2dd7c9776a58a577e592a814cd42"} Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.286134 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-9rrdv"] Feb 02 13:55:47 crc kubenswrapper[4846]: E0202 13:55:47.286936 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07b19d98-9059-4596-8ffb-b0264d3a6f14" containerName="mariadb-account-create-update" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.286962 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="07b19d98-9059-4596-8ffb-b0264d3a6f14" containerName="mariadb-account-create-update" Feb 02 13:55:47 crc kubenswrapper[4846]: E0202 13:55:47.287002 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6530545-8859-43e2-baee-8a1021830ba0" containerName="mariadb-database-create" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.287009 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6530545-8859-43e2-baee-8a1021830ba0" containerName="mariadb-database-create" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.287194 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="07b19d98-9059-4596-8ffb-b0264d3a6f14" containerName="mariadb-account-create-update" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.287217 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6530545-8859-43e2-baee-8a1021830ba0" containerName="mariadb-database-create" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.287996 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.290578 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-cw9rk" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.298870 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.312173 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.312219 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.331276 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-9rrdv"] Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.410050 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-config-data\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.410132 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-scripts\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.410262 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7njx6\" (UniqueName: \"kubernetes.io/projected/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-kube-api-access-7njx6\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.410368 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-combined-ca-bundle\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.513877 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-combined-ca-bundle\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.515117 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-config-data\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.515660 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-scripts\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.515976 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7njx6\" (UniqueName: \"kubernetes.io/projected/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-kube-api-access-7njx6\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.519938 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-combined-ca-bundle\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.521253 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-scripts\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.521650 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-config-data\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.536977 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7njx6\" (UniqueName: \"kubernetes.io/projected/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-kube-api-access-7njx6\") pod \"aodh-db-sync-9rrdv\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:47 crc kubenswrapper[4846]: I0202 13:55:47.645297 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:48 crc kubenswrapper[4846]: I0202 13:55:48.125748 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92f8c952-e300-4858-9bc5-c06350e65d6f","Type":"ContainerStarted","Data":"38dd80131691d0ddd7afdccb6f104c182057c7cf2aceb4c74f72a2c31f816873"} Feb 02 13:55:48 crc kubenswrapper[4846]: I0202 13:55:48.177890 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=20.177853632 podStartE2EDuration="20.177853632s" podCreationTimestamp="2026-02-02 13:55:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:55:48.154141946 +0000 UTC m=+6379.382728819" watchObservedRunningTime="2026-02-02 13:55:48.177853632 +0000 UTC m=+6379.406440495" Feb 02 13:55:48 crc kubenswrapper[4846]: I0202 13:55:48.267558 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-9rrdv"] Feb 02 13:55:49 crc kubenswrapper[4846]: I0202 13:55:49.134236 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-9rrdv" event={"ID":"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b","Type":"ContainerStarted","Data":"0626a4ee4e4523b917c58d338adf1d306fc5402a686e00db4c69b0a1812b5608"} Feb 02 13:55:49 crc kubenswrapper[4846]: I0202 13:55:49.204575 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:53 crc kubenswrapper[4846]: I0202 13:55:53.181075 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-9rrdv" event={"ID":"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b","Type":"ContainerStarted","Data":"c3af41b11d7360f8f6f19e49f6e6f6db0b201758905d8db6835e47be2b26a19e"} Feb 02 13:55:53 crc kubenswrapper[4846]: I0202 13:55:53.199663 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-9rrdv" podStartSLOduration=1.644594427 podStartE2EDuration="6.199647036s" podCreationTimestamp="2026-02-02 13:55:47 +0000 UTC" firstStartedPulling="2026-02-02 13:55:48.267663703 +0000 UTC m=+6379.496250566" lastFinishedPulling="2026-02-02 13:55:52.822716312 +0000 UTC m=+6384.051303175" observedRunningTime="2026-02-02 13:55:53.197499334 +0000 UTC m=+6384.426086197" watchObservedRunningTime="2026-02-02 13:55:53.199647036 +0000 UTC m=+6384.428233899" Feb 02 13:55:56 crc kubenswrapper[4846]: I0202 13:55:56.223612 4846 generic.go:334] "Generic (PLEG): container finished" podID="68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b" containerID="c3af41b11d7360f8f6f19e49f6e6f6db0b201758905d8db6835e47be2b26a19e" exitCode=0 Feb 02 13:55:56 crc kubenswrapper[4846]: I0202 13:55:56.223648 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-9rrdv" event={"ID":"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b","Type":"ContainerDied","Data":"c3af41b11d7360f8f6f19e49f6e6f6db0b201758905d8db6835e47be2b26a19e"} Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.426395 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:55:57 crc kubenswrapper[4846]: E0202 13:55:57.427053 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.683986 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.781037 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-scripts\") pod \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.781098 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-combined-ca-bundle\") pod \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.781226 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-config-data\") pod \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.781382 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7njx6\" (UniqueName: \"kubernetes.io/projected/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-kube-api-access-7njx6\") pod \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\" (UID: \"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b\") " Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.787275 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-scripts" (OuterVolumeSpecName: "scripts") pod "68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b" (UID: "68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.787881 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-kube-api-access-7njx6" (OuterVolumeSpecName: "kube-api-access-7njx6") pod "68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b" (UID: "68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b"). InnerVolumeSpecName "kube-api-access-7njx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.819865 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b" (UID: "68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.820134 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-config-data" (OuterVolumeSpecName: "config-data") pod "68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b" (UID: "68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.884249 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7njx6\" (UniqueName: \"kubernetes.io/projected/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-kube-api-access-7njx6\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.884301 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.884311 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:57 crc kubenswrapper[4846]: I0202 13:55:57.884320 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:55:58 crc kubenswrapper[4846]: I0202 13:55:58.241781 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-9rrdv" event={"ID":"68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b","Type":"ContainerDied","Data":"0626a4ee4e4523b917c58d338adf1d306fc5402a686e00db4c69b0a1812b5608"} Feb 02 13:55:58 crc kubenswrapper[4846]: I0202 13:55:58.241987 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0626a4ee4e4523b917c58d338adf1d306fc5402a686e00db4c69b0a1812b5608" Feb 02 13:55:58 crc kubenswrapper[4846]: I0202 13:55:58.241854 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-9rrdv" Feb 02 13:55:58 crc kubenswrapper[4846]: E0202 13:55:58.431937 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68edd6cc_a80e_45e1_b8bf_bad5f82d8c8b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68edd6cc_a80e_45e1_b8bf_bad5f82d8c8b.slice/crio-0626a4ee4e4523b917c58d338adf1d306fc5402a686e00db4c69b0a1812b5608\": RecentStats: unable to find data in memory cache]" Feb 02 13:55:59 crc kubenswrapper[4846]: I0202 13:55:59.204139 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:59 crc kubenswrapper[4846]: I0202 13:55:59.219488 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Feb 02 13:55:59 crc kubenswrapper[4846]: I0202 13:55:59.259536 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.645537 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.856354 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Feb 02 13:56:01 crc kubenswrapper[4846]: E0202 13:56:01.856907 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b" containerName="aodh-db-sync" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.856942 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b" containerName="aodh-db-sync" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.857202 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b" containerName="aodh-db-sync" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.859665 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.866866 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-cw9rk" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.866905 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.867143 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.887955 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.982990 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pscjz\" (UniqueName: \"kubernetes.io/projected/0689d078-b506-4bf6-88e0-34924fe53ea8-kube-api-access-pscjz\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.983041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-config-data\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.983110 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-scripts\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:01 crc kubenswrapper[4846]: I0202 13:56:01.983135 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:02 crc kubenswrapper[4846]: I0202 13:56:02.085708 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-scripts\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:02 crc kubenswrapper[4846]: I0202 13:56:02.085780 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:02 crc kubenswrapper[4846]: I0202 13:56:02.086042 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pscjz\" (UniqueName: \"kubernetes.io/projected/0689d078-b506-4bf6-88e0-34924fe53ea8-kube-api-access-pscjz\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:02 crc kubenswrapper[4846]: I0202 13:56:02.086087 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-config-data\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:02 crc kubenswrapper[4846]: I0202 13:56:02.092082 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-scripts\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:02 crc kubenswrapper[4846]: I0202 13:56:02.092317 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-combined-ca-bundle\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:02 crc kubenswrapper[4846]: I0202 13:56:02.100759 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-config-data\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:02 crc kubenswrapper[4846]: I0202 13:56:02.109589 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pscjz\" (UniqueName: \"kubernetes.io/projected/0689d078-b506-4bf6-88e0-34924fe53ea8-kube-api-access-pscjz\") pod \"aodh-0\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " pod="openstack/aodh-0" Feb 02 13:56:02 crc kubenswrapper[4846]: I0202 13:56:02.198865 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 02 13:56:02 crc kubenswrapper[4846]: I0202 13:56:02.719730 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Feb 02 13:56:03 crc kubenswrapper[4846]: I0202 13:56:03.296220 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0689d078-b506-4bf6-88e0-34924fe53ea8","Type":"ContainerStarted","Data":"3780b13f555fc910efe45918e153d37db01d1a362eb4b5f8221c8a47c6453f81"} Feb 02 13:56:03 crc kubenswrapper[4846]: I0202 13:56:03.831398 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:03 crc kubenswrapper[4846]: I0202 13:56:03.831729 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="ceilometer-central-agent" containerID="cri-o://9c8d3f52a82434fd7412c182fc8d512d20dda21c64ecb81c027e230e6257efe8" gracePeriod=30 Feb 02 13:56:03 crc kubenswrapper[4846]: I0202 13:56:03.831867 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="sg-core" containerID="cri-o://9223b8d6f428658de474c93db680d9ead607ae4c5198c8d5e31d2fbd95ed6283" gracePeriod=30 Feb 02 13:56:03 crc kubenswrapper[4846]: I0202 13:56:03.831867 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="proxy-httpd" containerID="cri-o://40a268cc4f98040ecf7197b7f236859697dd5ee20c1eefb00603aa283f68e085" gracePeriod=30 Feb 02 13:56:03 crc kubenswrapper[4846]: I0202 13:56:03.831923 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="ceilometer-notification-agent" containerID="cri-o://af04bb887a71994e4781459d3133e5f42a8ba589ab2401aa883d1fab2619ab52" gracePeriod=30 Feb 02 13:56:04 crc kubenswrapper[4846]: I0202 13:56:04.315524 4846 generic.go:334] "Generic (PLEG): container finished" podID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerID="40a268cc4f98040ecf7197b7f236859697dd5ee20c1eefb00603aa283f68e085" exitCode=0 Feb 02 13:56:04 crc kubenswrapper[4846]: I0202 13:56:04.315914 4846 generic.go:334] "Generic (PLEG): container finished" podID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerID="9223b8d6f428658de474c93db680d9ead607ae4c5198c8d5e31d2fbd95ed6283" exitCode=2 Feb 02 13:56:04 crc kubenswrapper[4846]: I0202 13:56:04.315930 4846 generic.go:334] "Generic (PLEG): container finished" podID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerID="9c8d3f52a82434fd7412c182fc8d512d20dda21c64ecb81c027e230e6257efe8" exitCode=0 Feb 02 13:56:04 crc kubenswrapper[4846]: I0202 13:56:04.315842 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c1ac007-e978-4ff3-8fec-30310b62f5fc","Type":"ContainerDied","Data":"40a268cc4f98040ecf7197b7f236859697dd5ee20c1eefb00603aa283f68e085"} Feb 02 13:56:04 crc kubenswrapper[4846]: I0202 13:56:04.315999 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c1ac007-e978-4ff3-8fec-30310b62f5fc","Type":"ContainerDied","Data":"9223b8d6f428658de474c93db680d9ead607ae4c5198c8d5e31d2fbd95ed6283"} Feb 02 13:56:04 crc kubenswrapper[4846]: I0202 13:56:04.316016 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c1ac007-e978-4ff3-8fec-30310b62f5fc","Type":"ContainerDied","Data":"9c8d3f52a82434fd7412c182fc8d512d20dda21c64ecb81c027e230e6257efe8"} Feb 02 13:56:05 crc kubenswrapper[4846]: I0202 13:56:05.328013 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0689d078-b506-4bf6-88e0-34924fe53ea8","Type":"ContainerStarted","Data":"b849984a4c19f6f8718f93f1ae1ecb9e07825fd828ca14973dad2c55ca3f8916"} Feb 02 13:56:05 crc kubenswrapper[4846]: I0202 13:56:05.546130 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.353275 4846 generic.go:334] "Generic (PLEG): container finished" podID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerID="af04bb887a71994e4781459d3133e5f42a8ba589ab2401aa883d1fab2619ab52" exitCode=0 Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.354009 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c1ac007-e978-4ff3-8fec-30310b62f5fc","Type":"ContainerDied","Data":"af04bb887a71994e4781459d3133e5f42a8ba589ab2401aa883d1fab2619ab52"} Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.704136 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.816685 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-scripts\") pod \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.816725 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-combined-ca-bundle\") pod \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.816755 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-log-httpd\") pod \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.816775 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsh6v\" (UniqueName: \"kubernetes.io/projected/2c1ac007-e978-4ff3-8fec-30310b62f5fc-kube-api-access-dsh6v\") pod \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.816843 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-sg-core-conf-yaml\") pod \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.816998 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-config-data\") pod \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.817057 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-run-httpd\") pod \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\" (UID: \"2c1ac007-e978-4ff3-8fec-30310b62f5fc\") " Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.817426 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2c1ac007-e978-4ff3-8fec-30310b62f5fc" (UID: "2c1ac007-e978-4ff3-8fec-30310b62f5fc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.817522 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2c1ac007-e978-4ff3-8fec-30310b62f5fc" (UID: "2c1ac007-e978-4ff3-8fec-30310b62f5fc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.819595 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.819640 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c1ac007-e978-4ff3-8fec-30310b62f5fc-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.827862 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c1ac007-e978-4ff3-8fec-30310b62f5fc-kube-api-access-dsh6v" (OuterVolumeSpecName: "kube-api-access-dsh6v") pod "2c1ac007-e978-4ff3-8fec-30310b62f5fc" (UID: "2c1ac007-e978-4ff3-8fec-30310b62f5fc"). InnerVolumeSpecName "kube-api-access-dsh6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.827974 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-scripts" (OuterVolumeSpecName: "scripts") pod "2c1ac007-e978-4ff3-8fec-30310b62f5fc" (UID: "2c1ac007-e978-4ff3-8fec-30310b62f5fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.859738 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2c1ac007-e978-4ff3-8fec-30310b62f5fc" (UID: "2c1ac007-e978-4ff3-8fec-30310b62f5fc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.902162 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c1ac007-e978-4ff3-8fec-30310b62f5fc" (UID: "2c1ac007-e978-4ff3-8fec-30310b62f5fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.921431 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.921466 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.921478 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsh6v\" (UniqueName: \"kubernetes.io/projected/2c1ac007-e978-4ff3-8fec-30310b62f5fc-kube-api-access-dsh6v\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.921487 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:07 crc kubenswrapper[4846]: I0202 13:56:07.935070 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-config-data" (OuterVolumeSpecName: "config-data") pod "2c1ac007-e978-4ff3-8fec-30310b62f5fc" (UID: "2c1ac007-e978-4ff3-8fec-30310b62f5fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.024234 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c1ac007-e978-4ff3-8fec-30310b62f5fc-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.383943 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.383947 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c1ac007-e978-4ff3-8fec-30310b62f5fc","Type":"ContainerDied","Data":"32e132f338edd530dfcb842a71b769a4d8d2950b91c9b0dfb047994302f76ebf"} Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.384421 4846 scope.go:117] "RemoveContainer" containerID="40a268cc4f98040ecf7197b7f236859697dd5ee20c1eefb00603aa283f68e085" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.390997 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0689d078-b506-4bf6-88e0-34924fe53ea8","Type":"ContainerStarted","Data":"89d5042220604812d01c35ff839476320f6885886c0aecd34e3bf8bdd418b199"} Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.417986 4846 scope.go:117] "RemoveContainer" containerID="9223b8d6f428658de474c93db680d9ead607ae4c5198c8d5e31d2fbd95ed6283" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.425856 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:56:08 crc kubenswrapper[4846]: E0202 13:56:08.426193 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.446062 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.473863 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.496542 4846 scope.go:117] "RemoveContainer" containerID="af04bb887a71994e4781459d3133e5f42a8ba589ab2401aa883d1fab2619ab52" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.504303 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:08 crc kubenswrapper[4846]: E0202 13:56:08.504905 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="sg-core" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.504933 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="sg-core" Feb 02 13:56:08 crc kubenswrapper[4846]: E0202 13:56:08.504973 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="proxy-httpd" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.504981 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="proxy-httpd" Feb 02 13:56:08 crc kubenswrapper[4846]: E0202 13:56:08.505003 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="ceilometer-central-agent" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.505011 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="ceilometer-central-agent" Feb 02 13:56:08 crc kubenswrapper[4846]: E0202 13:56:08.505029 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="ceilometer-notification-agent" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.505036 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="ceilometer-notification-agent" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.505255 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="proxy-httpd" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.505281 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="ceilometer-notification-agent" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.505308 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="ceilometer-central-agent" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.505321 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" containerName="sg-core" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.507808 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.512197 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.512432 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.531274 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.534870 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-log-httpd\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.535032 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-scripts\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.535066 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.535143 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.535187 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-run-httpd\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.535382 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lqdr\" (UniqueName: \"kubernetes.io/projected/fcd2e199-5090-403a-ac17-c32ee4b2ff54-kube-api-access-8lqdr\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.535528 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-config-data\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.555102 4846 scope.go:117] "RemoveContainer" containerID="9c8d3f52a82434fd7412c182fc8d512d20dda21c64ecb81c027e230e6257efe8" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.637385 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.637487 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-run-httpd\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.637664 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lqdr\" (UniqueName: \"kubernetes.io/projected/fcd2e199-5090-403a-ac17-c32ee4b2ff54-kube-api-access-8lqdr\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.637768 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-config-data\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.637927 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-log-httpd\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.638094 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-scripts\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.638135 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.638449 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-run-httpd\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.639358 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-log-httpd\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.643178 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.643386 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-scripts\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.645714 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.660833 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-config-data\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.664729 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lqdr\" (UniqueName: \"kubernetes.io/projected/fcd2e199-5090-403a-ac17-c32ee4b2ff54-kube-api-access-8lqdr\") pod \"ceilometer-0\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " pod="openstack/ceilometer-0" Feb 02 13:56:08 crc kubenswrapper[4846]: I0202 13:56:08.853237 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 13:56:09 crc kubenswrapper[4846]: I0202 13:56:09.451680 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c1ac007-e978-4ff3-8fec-30310b62f5fc" path="/var/lib/kubelet/pods/2c1ac007-e978-4ff3-8fec-30310b62f5fc/volumes" Feb 02 13:56:09 crc kubenswrapper[4846]: I0202 13:56:09.529415 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:10 crc kubenswrapper[4846]: I0202 13:56:10.423698 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0689d078-b506-4bf6-88e0-34924fe53ea8","Type":"ContainerStarted","Data":"d42f8b3309202b1dd2ae39c153cedd1d43f360b8c575c8ac0d2dabd8c1d9e6c5"} Feb 02 13:56:10 crc kubenswrapper[4846]: I0202 13:56:10.425567 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcd2e199-5090-403a-ac17-c32ee4b2ff54","Type":"ContainerStarted","Data":"05fba50a8bbd31a88931cf72b6bb177958c74b97e8907a6f10a6f2f89637c811"} Feb 02 13:56:10 crc kubenswrapper[4846]: I0202 13:56:10.886774 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 13:56:10 crc kubenswrapper[4846]: I0202 13:56:10.887006 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="6b9d958f-3aab-4199-ad20-6fb2b318b7b4" containerName="kube-state-metrics" containerID="cri-o://c743bebd24bbf13a9176e62e172e920485827218baf086194a3ab262d2fef121" gracePeriod=30 Feb 02 13:56:11 crc kubenswrapper[4846]: I0202 13:56:11.010457 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="6b9d958f-3aab-4199-ad20-6fb2b318b7b4" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.1.143:8081/readyz\": dial tcp 10.217.1.143:8081: connect: connection refused" Feb 02 13:56:11 crc kubenswrapper[4846]: I0202 13:56:11.439926 4846 generic.go:334] "Generic (PLEG): container finished" podID="6b9d958f-3aab-4199-ad20-6fb2b318b7b4" containerID="c743bebd24bbf13a9176e62e172e920485827218baf086194a3ab262d2fef121" exitCode=2 Feb 02 13:56:11 crc kubenswrapper[4846]: I0202 13:56:11.440272 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6b9d958f-3aab-4199-ad20-6fb2b318b7b4","Type":"ContainerDied","Data":"c743bebd24bbf13a9176e62e172e920485827218baf086194a3ab262d2fef121"} Feb 02 13:56:11 crc kubenswrapper[4846]: I0202 13:56:11.442102 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcd2e199-5090-403a-ac17-c32ee4b2ff54","Type":"ContainerStarted","Data":"ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b"} Feb 02 13:56:11 crc kubenswrapper[4846]: I0202 13:56:11.653497 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 13:56:11 crc kubenswrapper[4846]: I0202 13:56:11.719068 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n47t6\" (UniqueName: \"kubernetes.io/projected/6b9d958f-3aab-4199-ad20-6fb2b318b7b4-kube-api-access-n47t6\") pod \"6b9d958f-3aab-4199-ad20-6fb2b318b7b4\" (UID: \"6b9d958f-3aab-4199-ad20-6fb2b318b7b4\") " Feb 02 13:56:11 crc kubenswrapper[4846]: I0202 13:56:11.734497 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b9d958f-3aab-4199-ad20-6fb2b318b7b4-kube-api-access-n47t6" (OuterVolumeSpecName: "kube-api-access-n47t6") pod "6b9d958f-3aab-4199-ad20-6fb2b318b7b4" (UID: "6b9d958f-3aab-4199-ad20-6fb2b318b7b4"). InnerVolumeSpecName "kube-api-access-n47t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:56:11 crc kubenswrapper[4846]: I0202 13:56:11.828225 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n47t6\" (UniqueName: \"kubernetes.io/projected/6b9d958f-3aab-4199-ad20-6fb2b318b7b4-kube-api-access-n47t6\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.455860 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0689d078-b506-4bf6-88e0-34924fe53ea8","Type":"ContainerStarted","Data":"2f80abe0a5193384654e993e006df38f8dae3eb2c8c5ab6eab5dd221b707724e"} Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.456227 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-api" containerID="cri-o://b849984a4c19f6f8718f93f1ae1ecb9e07825fd828ca14973dad2c55ca3f8916" gracePeriod=30 Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.456266 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-listener" containerID="cri-o://2f80abe0a5193384654e993e006df38f8dae3eb2c8c5ab6eab5dd221b707724e" gracePeriod=30 Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.456599 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-notifier" containerID="cri-o://d42f8b3309202b1dd2ae39c153cedd1d43f360b8c575c8ac0d2dabd8c1d9e6c5" gracePeriod=30 Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.456717 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-evaluator" containerID="cri-o://89d5042220604812d01c35ff839476320f6885886c0aecd34e3bf8bdd418b199" gracePeriod=30 Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.468982 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.470128 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6b9d958f-3aab-4199-ad20-6fb2b318b7b4","Type":"ContainerDied","Data":"e9a7763118d06405b7e3ec081bcf72b9b6236520d9266c785abef69e0da8cafd"} Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.470203 4846 scope.go:117] "RemoveContainer" containerID="c743bebd24bbf13a9176e62e172e920485827218baf086194a3ab262d2fef121" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.476291 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcd2e199-5090-403a-ac17-c32ee4b2ff54","Type":"ContainerStarted","Data":"29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2"} Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.481059 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.660316994 podStartE2EDuration="11.481040119s" podCreationTimestamp="2026-02-02 13:56:01 +0000 UTC" firstStartedPulling="2026-02-02 13:56:02.738198471 +0000 UTC m=+6393.966785334" lastFinishedPulling="2026-02-02 13:56:11.558921596 +0000 UTC m=+6402.787508459" observedRunningTime="2026-02-02 13:56:12.479237815 +0000 UTC m=+6403.707824678" watchObservedRunningTime="2026-02-02 13:56:12.481040119 +0000 UTC m=+6403.709626982" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.561358 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.627714 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.653697 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 13:56:12 crc kubenswrapper[4846]: E0202 13:56:12.654288 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b9d958f-3aab-4199-ad20-6fb2b318b7b4" containerName="kube-state-metrics" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.654318 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b9d958f-3aab-4199-ad20-6fb2b318b7b4" containerName="kube-state-metrics" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.654559 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b9d958f-3aab-4199-ad20-6fb2b318b7b4" containerName="kube-state-metrics" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.655727 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.658877 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.659203 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.680118 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.749920 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxlsx\" (UniqueName: \"kubernetes.io/projected/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-kube-api-access-bxlsx\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.750263 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.750297 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.750324 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.851793 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.851843 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.851873 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.852067 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxlsx\" (UniqueName: \"kubernetes.io/projected/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-kube-api-access-bxlsx\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.861268 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.862992 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.877522 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:12 crc kubenswrapper[4846]: I0202 13:56:12.893205 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxlsx\" (UniqueName: \"kubernetes.io/projected/f3d505b7-7c82-49d0-930d-8ba793ac0d4a-kube-api-access-bxlsx\") pod \"kube-state-metrics-0\" (UID: \"f3d505b7-7c82-49d0-930d-8ba793ac0d4a\") " pod="openstack/kube-state-metrics-0" Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.052333 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.368280 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.453094 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b9d958f-3aab-4199-ad20-6fb2b318b7b4" path="/var/lib/kubelet/pods/6b9d958f-3aab-4199-ad20-6fb2b318b7b4/volumes" Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.516326 4846 generic.go:334] "Generic (PLEG): container finished" podID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerID="d42f8b3309202b1dd2ae39c153cedd1d43f360b8c575c8ac0d2dabd8c1d9e6c5" exitCode=0 Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.516366 4846 generic.go:334] "Generic (PLEG): container finished" podID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerID="89d5042220604812d01c35ff839476320f6885886c0aecd34e3bf8bdd418b199" exitCode=0 Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.516377 4846 generic.go:334] "Generic (PLEG): container finished" podID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerID="b849984a4c19f6f8718f93f1ae1ecb9e07825fd828ca14973dad2c55ca3f8916" exitCode=0 Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.516501 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0689d078-b506-4bf6-88e0-34924fe53ea8","Type":"ContainerDied","Data":"d42f8b3309202b1dd2ae39c153cedd1d43f360b8c575c8ac0d2dabd8c1d9e6c5"} Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.516535 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0689d078-b506-4bf6-88e0-34924fe53ea8","Type":"ContainerDied","Data":"89d5042220604812d01c35ff839476320f6885886c0aecd34e3bf8bdd418b199"} Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.516547 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0689d078-b506-4bf6-88e0-34924fe53ea8","Type":"ContainerDied","Data":"b849984a4c19f6f8718f93f1ae1ecb9e07825fd828ca14973dad2c55ca3f8916"} Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.526426 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcd2e199-5090-403a-ac17-c32ee4b2ff54","Type":"ContainerStarted","Data":"b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69"} Feb 02 13:56:13 crc kubenswrapper[4846]: I0202 13:56:13.554386 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 02 13:56:14 crc kubenswrapper[4846]: I0202 13:56:14.541010 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f3d505b7-7c82-49d0-930d-8ba793ac0d4a","Type":"ContainerStarted","Data":"2e7bed0431626b006855b2ec0d40f4892f12eab67a34f06b4bf78eca7c8feb7a"} Feb 02 13:56:15 crc kubenswrapper[4846]: I0202 13:56:15.551581 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcd2e199-5090-403a-ac17-c32ee4b2ff54","Type":"ContainerStarted","Data":"65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f"} Feb 02 13:56:15 crc kubenswrapper[4846]: I0202 13:56:15.551732 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="ceilometer-central-agent" containerID="cri-o://ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b" gracePeriod=30 Feb 02 13:56:15 crc kubenswrapper[4846]: I0202 13:56:15.551796 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="sg-core" containerID="cri-o://b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69" gracePeriod=30 Feb 02 13:56:15 crc kubenswrapper[4846]: I0202 13:56:15.551792 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="proxy-httpd" containerID="cri-o://65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f" gracePeriod=30 Feb 02 13:56:15 crc kubenswrapper[4846]: I0202 13:56:15.551823 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="ceilometer-notification-agent" containerID="cri-o://29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2" gracePeriod=30 Feb 02 13:56:15 crc kubenswrapper[4846]: I0202 13:56:15.551910 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 13:56:15 crc kubenswrapper[4846]: I0202 13:56:15.553999 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f3d505b7-7c82-49d0-930d-8ba793ac0d4a","Type":"ContainerStarted","Data":"1d27ea15179de84ca1f9088aeb12d9a176edf4dd7a1fb663ce584474d9a6f86a"} Feb 02 13:56:15 crc kubenswrapper[4846]: I0202 13:56:15.554166 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 02 13:56:15 crc kubenswrapper[4846]: I0202 13:56:15.586190 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7682838739999998 podStartE2EDuration="7.586165068s" podCreationTimestamp="2026-02-02 13:56:08 +0000 UTC" firstStartedPulling="2026-02-02 13:56:09.790576101 +0000 UTC m=+6401.019162964" lastFinishedPulling="2026-02-02 13:56:14.608457295 +0000 UTC m=+6405.837044158" observedRunningTime="2026-02-02 13:56:15.578901812 +0000 UTC m=+6406.807488675" watchObservedRunningTime="2026-02-02 13:56:15.586165068 +0000 UTC m=+6406.814751931" Feb 02 13:56:15 crc kubenswrapper[4846]: I0202 13:56:15.599663 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.15511989 podStartE2EDuration="3.599640286s" podCreationTimestamp="2026-02-02 13:56:12 +0000 UTC" firstStartedPulling="2026-02-02 13:56:13.557647515 +0000 UTC m=+6404.786234378" lastFinishedPulling="2026-02-02 13:56:14.002167911 +0000 UTC m=+6405.230754774" observedRunningTime="2026-02-02 13:56:15.596055098 +0000 UTC m=+6406.824641961" watchObservedRunningTime="2026-02-02 13:56:15.599640286 +0000 UTC m=+6406.828227149" Feb 02 13:56:16 crc kubenswrapper[4846]: I0202 13:56:16.566638 4846 generic.go:334] "Generic (PLEG): container finished" podID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerID="65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f" exitCode=0 Feb 02 13:56:16 crc kubenswrapper[4846]: I0202 13:56:16.566975 4846 generic.go:334] "Generic (PLEG): container finished" podID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerID="b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69" exitCode=2 Feb 02 13:56:16 crc kubenswrapper[4846]: I0202 13:56:16.566987 4846 generic.go:334] "Generic (PLEG): container finished" podID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerID="29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2" exitCode=0 Feb 02 13:56:16 crc kubenswrapper[4846]: I0202 13:56:16.566655 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcd2e199-5090-403a-ac17-c32ee4b2ff54","Type":"ContainerDied","Data":"65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f"} Feb 02 13:56:16 crc kubenswrapper[4846]: I0202 13:56:16.567120 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcd2e199-5090-403a-ac17-c32ee4b2ff54","Type":"ContainerDied","Data":"b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69"} Feb 02 13:56:16 crc kubenswrapper[4846]: I0202 13:56:16.567137 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcd2e199-5090-403a-ac17-c32ee4b2ff54","Type":"ContainerDied","Data":"29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2"} Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.411681 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.594742 4846 generic.go:334] "Generic (PLEG): container finished" podID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerID="ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b" exitCode=0 Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.594794 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcd2e199-5090-403a-ac17-c32ee4b2ff54","Type":"ContainerDied","Data":"ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b"} Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.594820 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcd2e199-5090-403a-ac17-c32ee4b2ff54","Type":"ContainerDied","Data":"05fba50a8bbd31a88931cf72b6bb177958c74b97e8907a6f10a6f2f89637c811"} Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.594836 4846 scope.go:117] "RemoveContainer" containerID="65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.594958 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.613729 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lqdr\" (UniqueName: \"kubernetes.io/projected/fcd2e199-5090-403a-ac17-c32ee4b2ff54-kube-api-access-8lqdr\") pod \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.613842 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-combined-ca-bundle\") pod \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.613872 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-run-httpd\") pod \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.613917 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-log-httpd\") pod \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.613989 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-config-data\") pod \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.614038 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-sg-core-conf-yaml\") pod \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.614123 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-scripts\") pod \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\" (UID: \"fcd2e199-5090-403a-ac17-c32ee4b2ff54\") " Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.614455 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fcd2e199-5090-403a-ac17-c32ee4b2ff54" (UID: "fcd2e199-5090-403a-ac17-c32ee4b2ff54"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.615022 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fcd2e199-5090-403a-ac17-c32ee4b2ff54" (UID: "fcd2e199-5090-403a-ac17-c32ee4b2ff54"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.615032 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.621934 4846 scope.go:117] "RemoveContainer" containerID="b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.622900 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-scripts" (OuterVolumeSpecName: "scripts") pod "fcd2e199-5090-403a-ac17-c32ee4b2ff54" (UID: "fcd2e199-5090-403a-ac17-c32ee4b2ff54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.639230 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcd2e199-5090-403a-ac17-c32ee4b2ff54-kube-api-access-8lqdr" (OuterVolumeSpecName: "kube-api-access-8lqdr") pod "fcd2e199-5090-403a-ac17-c32ee4b2ff54" (UID: "fcd2e199-5090-403a-ac17-c32ee4b2ff54"). InnerVolumeSpecName "kube-api-access-8lqdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.650685 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fcd2e199-5090-403a-ac17-c32ee4b2ff54" (UID: "fcd2e199-5090-403a-ac17-c32ee4b2ff54"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.716343 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.716383 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lqdr\" (UniqueName: \"kubernetes.io/projected/fcd2e199-5090-403a-ac17-c32ee4b2ff54-kube-api-access-8lqdr\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.716396 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcd2e199-5090-403a-ac17-c32ee4b2ff54-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.716406 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.718111 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcd2e199-5090-403a-ac17-c32ee4b2ff54" (UID: "fcd2e199-5090-403a-ac17-c32ee4b2ff54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.740899 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-config-data" (OuterVolumeSpecName: "config-data") pod "fcd2e199-5090-403a-ac17-c32ee4b2ff54" (UID: "fcd2e199-5090-403a-ac17-c32ee4b2ff54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.768028 4846 scope.go:117] "RemoveContainer" containerID="29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.787430 4846 scope.go:117] "RemoveContainer" containerID="ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.808647 4846 scope.go:117] "RemoveContainer" containerID="65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f" Feb 02 13:56:19 crc kubenswrapper[4846]: E0202 13:56:19.809132 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f\": container with ID starting with 65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f not found: ID does not exist" containerID="65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.809175 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f"} err="failed to get container status \"65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f\": rpc error: code = NotFound desc = could not find container \"65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f\": container with ID starting with 65df9f7d58b0ce2021936358ad577ed1ca737c0a8d8988a624812195b5ddb59f not found: ID does not exist" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.809195 4846 scope.go:117] "RemoveContainer" containerID="b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69" Feb 02 13:56:19 crc kubenswrapper[4846]: E0202 13:56:19.809530 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69\": container with ID starting with b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69 not found: ID does not exist" containerID="b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.809547 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69"} err="failed to get container status \"b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69\": rpc error: code = NotFound desc = could not find container \"b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69\": container with ID starting with b35393d21345bc1714cb107638f430d6af3cc2957d5bc54d5fb5a02ad6f5cf69 not found: ID does not exist" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.809558 4846 scope.go:117] "RemoveContainer" containerID="29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2" Feb 02 13:56:19 crc kubenswrapper[4846]: E0202 13:56:19.809826 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2\": container with ID starting with 29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2 not found: ID does not exist" containerID="29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.809841 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2"} err="failed to get container status \"29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2\": rpc error: code = NotFound desc = could not find container \"29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2\": container with ID starting with 29a606ab7ea30b5175aa8de0b6c72b750b8aab1a1e8dfa9d32c55c38ce7e72a2 not found: ID does not exist" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.809852 4846 scope.go:117] "RemoveContainer" containerID="ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b" Feb 02 13:56:19 crc kubenswrapper[4846]: E0202 13:56:19.810023 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b\": container with ID starting with ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b not found: ID does not exist" containerID="ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.810038 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b"} err="failed to get container status \"ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b\": rpc error: code = NotFound desc = could not find container \"ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b\": container with ID starting with ef5f7fd19c8beb04ee3ef28c8b8e24c184a202264006e5e63b5fd65cad724f7b not found: ID does not exist" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.818374 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.818401 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcd2e199-5090-403a-ac17-c32ee4b2ff54-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.964827 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.976059 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.987184 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:19 crc kubenswrapper[4846]: E0202 13:56:19.987611 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="ceilometer-central-agent" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.987643 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="ceilometer-central-agent" Feb 02 13:56:19 crc kubenswrapper[4846]: E0202 13:56:19.987669 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="sg-core" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.987681 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="sg-core" Feb 02 13:56:19 crc kubenswrapper[4846]: E0202 13:56:19.987699 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="ceilometer-notification-agent" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.987705 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="ceilometer-notification-agent" Feb 02 13:56:19 crc kubenswrapper[4846]: E0202 13:56:19.987712 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="proxy-httpd" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.987719 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="proxy-httpd" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.987884 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="ceilometer-central-agent" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.987899 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="sg-core" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.987907 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="ceilometer-notification-agent" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.987918 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" containerName="proxy-httpd" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.989720 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.992412 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.992801 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 02 13:56:19 crc kubenswrapper[4846]: I0202 13:56:19.992865 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.002066 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.021543 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjp5k\" (UniqueName: \"kubernetes.io/projected/f6b5c3d0-71ce-441c-8423-bc2f285a782f-kube-api-access-tjp5k\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.021592 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.021707 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-scripts\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.021789 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-config-data\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.021843 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b5c3d0-71ce-441c-8423-bc2f285a782f-run-httpd\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.021861 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.021877 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.021893 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b5c3d0-71ce-441c-8423-bc2f285a782f-log-httpd\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.123341 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-config-data\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.123430 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b5c3d0-71ce-441c-8423-bc2f285a782f-run-httpd\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.123452 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.123470 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.123484 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b5c3d0-71ce-441c-8423-bc2f285a782f-log-httpd\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.123545 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjp5k\" (UniqueName: \"kubernetes.io/projected/f6b5c3d0-71ce-441c-8423-bc2f285a782f-kube-api-access-tjp5k\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.123574 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.123630 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-scripts\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.124352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b5c3d0-71ce-441c-8423-bc2f285a782f-log-httpd\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.124393 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6b5c3d0-71ce-441c-8423-bc2f285a782f-run-httpd\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.127014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.128102 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-scripts\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.128562 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.137479 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.137999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b5c3d0-71ce-441c-8423-bc2f285a782f-config-data\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.142440 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjp5k\" (UniqueName: \"kubernetes.io/projected/f6b5c3d0-71ce-441c-8423-bc2f285a782f-kube-api-access-tjp5k\") pod \"ceilometer-0\" (UID: \"f6b5c3d0-71ce-441c-8423-bc2f285a782f\") " pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.318144 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.434695 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:56:20 crc kubenswrapper[4846]: E0202 13:56:20.435116 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:56:20 crc kubenswrapper[4846]: I0202 13:56:20.803970 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 02 13:56:21 crc kubenswrapper[4846]: I0202 13:56:21.436506 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcd2e199-5090-403a-ac17-c32ee4b2ff54" path="/var/lib/kubelet/pods/fcd2e199-5090-403a-ac17-c32ee4b2ff54/volumes" Feb 02 13:56:21 crc kubenswrapper[4846]: I0202 13:56:21.620080 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b5c3d0-71ce-441c-8423-bc2f285a782f","Type":"ContainerStarted","Data":"500d2e1adcc78935ad6827fdb172d1b8fc77b517c4e0ad641806ba2537696555"} Feb 02 13:56:21 crc kubenswrapper[4846]: I0202 13:56:21.620144 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b5c3d0-71ce-441c-8423-bc2f285a782f","Type":"ContainerStarted","Data":"9b7961d30705388909df1cafd75339085ec52a7de322e3854035732d20411f19"} Feb 02 13:56:22 crc kubenswrapper[4846]: I0202 13:56:22.635385 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b5c3d0-71ce-441c-8423-bc2f285a782f","Type":"ContainerStarted","Data":"0837a11bea31bd9502b6b709f1aa1e9f03a9729a8228aad524c99c046067d494"} Feb 02 13:56:23 crc kubenswrapper[4846]: I0202 13:56:23.062770 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 02 13:56:23 crc kubenswrapper[4846]: I0202 13:56:23.646788 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b5c3d0-71ce-441c-8423-bc2f285a782f","Type":"ContainerStarted","Data":"28be5cac726fdfdd92bc83e908a7d7a6a04aad5e1d37a387d3c38f860d267bbe"} Feb 02 13:56:26 crc kubenswrapper[4846]: I0202 13:56:26.684151 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6b5c3d0-71ce-441c-8423-bc2f285a782f","Type":"ContainerStarted","Data":"94b9d6f3a7cb32ee2d5af5493269662a2c731c3a5f4e68ec87fc89d1856b842e"} Feb 02 13:56:26 crc kubenswrapper[4846]: I0202 13:56:26.684754 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 02 13:56:26 crc kubenswrapper[4846]: I0202 13:56:26.732478 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.131478143 podStartE2EDuration="7.73245796s" podCreationTimestamp="2026-02-02 13:56:19 +0000 UTC" firstStartedPulling="2026-02-02 13:56:20.814475269 +0000 UTC m=+6412.043062132" lastFinishedPulling="2026-02-02 13:56:25.415455086 +0000 UTC m=+6416.644041949" observedRunningTime="2026-02-02 13:56:26.720298554 +0000 UTC m=+6417.948885417" watchObservedRunningTime="2026-02-02 13:56:26.73245796 +0000 UTC m=+6417.961044823" Feb 02 13:56:34 crc kubenswrapper[4846]: I0202 13:56:34.047521 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-lkmcf"] Feb 02 13:56:34 crc kubenswrapper[4846]: I0202 13:56:34.060187 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-lkmcf"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.045848 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-t6jvf"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.057833 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-34fb-account-create-update-sdf2h"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.071879 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-2751-account-create-update-5dsmk"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.083212 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-0340-account-create-update-rcslh"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.091710 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-t6jvf"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.102664 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-6l9gs"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.114808 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-34fb-account-create-update-sdf2h"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.124197 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-2751-account-create-update-5dsmk"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.132819 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-0340-account-create-update-rcslh"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.141423 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-6l9gs"] Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.424427 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:56:35 crc kubenswrapper[4846]: E0202 13:56:35.425074 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.436714 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07556157-fc87-4ff5-9fba-2ee61250b2a5" path="/var/lib/kubelet/pods/07556157-fc87-4ff5-9fba-2ee61250b2a5/volumes" Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.438490 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5073120c-a1eb-4cf7-8b4c-a2b807027589" path="/var/lib/kubelet/pods/5073120c-a1eb-4cf7-8b4c-a2b807027589/volumes" Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.439719 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54e66fd9-2451-40c4-b317-783d545e1db9" path="/var/lib/kubelet/pods/54e66fd9-2451-40c4-b317-783d545e1db9/volumes" Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.441489 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d98e27a-a15d-49d4-a2ba-06accdf7770c" path="/var/lib/kubelet/pods/6d98e27a-a15d-49d4-a2ba-06accdf7770c/volumes" Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.443048 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ea7209b-7df1-4b9a-a666-cd8257cf8088" path="/var/lib/kubelet/pods/7ea7209b-7df1-4b9a-a666-cd8257cf8088/volumes" Feb 02 13:56:35 crc kubenswrapper[4846]: I0202 13:56:35.443765 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2cf8f76-47bb-4770-908b-723059ba5e13" path="/var/lib/kubelet/pods/f2cf8f76-47bb-4770-908b-723059ba5e13/volumes" Feb 02 13:56:41 crc kubenswrapper[4846]: I0202 13:56:41.374819 4846 scope.go:117] "RemoveContainer" containerID="f129d363d1dbdddac20c21c32b3245bac34e95f3d272ad4d1e2ff8d825e642fb" Feb 02 13:56:41 crc kubenswrapper[4846]: I0202 13:56:41.409282 4846 scope.go:117] "RemoveContainer" containerID="c5e00b3a85089e49d54b4fcbb634f4a71d45c3cadc78cc4584d3f28f246ef73e" Feb 02 13:56:41 crc kubenswrapper[4846]: I0202 13:56:41.471032 4846 scope.go:117] "RemoveContainer" containerID="567d80943db2be70626e34b3ff797b1eae044b0b8ede6c647397a824c261a48e" Feb 02 13:56:41 crc kubenswrapper[4846]: I0202 13:56:41.518128 4846 scope.go:117] "RemoveContainer" containerID="8893cbeaad2d8b9be4717112cb07b6e1b8a3d21172730b7a2da7125d168c166e" Feb 02 13:56:41 crc kubenswrapper[4846]: I0202 13:56:41.568542 4846 scope.go:117] "RemoveContainer" containerID="37020039b2fa96897ec7fba8404021b59483a98cc8dc8eda772c3424a2bd6350" Feb 02 13:56:41 crc kubenswrapper[4846]: I0202 13:56:41.620896 4846 scope.go:117] "RemoveContainer" containerID="764ebfdfa1d02b6a8ea75afc9431bf348d397e476a0d32ce5e0d2e93cd5c8748" Feb 02 13:56:42 crc kubenswrapper[4846]: I0202 13:56:42.843951 4846 generic.go:334] "Generic (PLEG): container finished" podID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerID="2f80abe0a5193384654e993e006df38f8dae3eb2c8c5ab6eab5dd221b707724e" exitCode=137 Feb 02 13:56:42 crc kubenswrapper[4846]: I0202 13:56:42.844042 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0689d078-b506-4bf6-88e0-34924fe53ea8","Type":"ContainerDied","Data":"2f80abe0a5193384654e993e006df38f8dae3eb2c8c5ab6eab5dd221b707724e"} Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.458547 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.543092 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-config-data\") pod \"0689d078-b506-4bf6-88e0-34924fe53ea8\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.543168 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pscjz\" (UniqueName: \"kubernetes.io/projected/0689d078-b506-4bf6-88e0-34924fe53ea8-kube-api-access-pscjz\") pod \"0689d078-b506-4bf6-88e0-34924fe53ea8\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.543366 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-scripts\") pod \"0689d078-b506-4bf6-88e0-34924fe53ea8\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.543397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-combined-ca-bundle\") pod \"0689d078-b506-4bf6-88e0-34924fe53ea8\" (UID: \"0689d078-b506-4bf6-88e0-34924fe53ea8\") " Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.550273 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0689d078-b506-4bf6-88e0-34924fe53ea8-kube-api-access-pscjz" (OuterVolumeSpecName: "kube-api-access-pscjz") pod "0689d078-b506-4bf6-88e0-34924fe53ea8" (UID: "0689d078-b506-4bf6-88e0-34924fe53ea8"). InnerVolumeSpecName "kube-api-access-pscjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.553696 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-scripts" (OuterVolumeSpecName: "scripts") pod "0689d078-b506-4bf6-88e0-34924fe53ea8" (UID: "0689d078-b506-4bf6-88e0-34924fe53ea8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.647272 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.647312 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pscjz\" (UniqueName: \"kubernetes.io/projected/0689d078-b506-4bf6-88e0-34924fe53ea8-kube-api-access-pscjz\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.912740 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0689d078-b506-4bf6-88e0-34924fe53ea8" (UID: "0689d078-b506-4bf6-88e0-34924fe53ea8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.913586 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-config-data" (OuterVolumeSpecName: "config-data") pod "0689d078-b506-4bf6-88e0-34924fe53ea8" (UID: "0689d078-b506-4bf6-88e0-34924fe53ea8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.927959 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0689d078-b506-4bf6-88e0-34924fe53ea8","Type":"ContainerDied","Data":"3780b13f555fc910efe45918e153d37db01d1a362eb4b5f8221c8a47c6453f81"} Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.928252 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.929014 4846 scope.go:117] "RemoveContainer" containerID="2f80abe0a5193384654e993e006df38f8dae3eb2c8c5ab6eab5dd221b707724e" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.978015 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.978048 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0689d078-b506-4bf6-88e0-34924fe53ea8-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.983832 4846 scope.go:117] "RemoveContainer" containerID="d42f8b3309202b1dd2ae39c153cedd1d43f360b8c575c8ac0d2dabd8c1d9e6c5" Feb 02 13:56:43 crc kubenswrapper[4846]: I0202 13:56:43.990086 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.007648 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.024956 4846 scope.go:117] "RemoveContainer" containerID="89d5042220604812d01c35ff839476320f6885886c0aecd34e3bf8bdd418b199" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.044374 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Feb 02 13:56:44 crc kubenswrapper[4846]: E0202 13:56:44.044922 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-listener" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.044946 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-listener" Feb 02 13:56:44 crc kubenswrapper[4846]: E0202 13:56:44.044979 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-evaluator" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.044990 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-evaluator" Feb 02 13:56:44 crc kubenswrapper[4846]: E0202 13:56:44.045000 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-notifier" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.045007 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-notifier" Feb 02 13:56:44 crc kubenswrapper[4846]: E0202 13:56:44.045045 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-api" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.045053 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-api" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.045294 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-notifier" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.045329 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-evaluator" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.045344 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-api" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.045362 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" containerName="aodh-listener" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.047748 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.048882 4846 scope.go:117] "RemoveContainer" containerID="b849984a4c19f6f8718f93f1ae1ecb9e07825fd828ca14973dad2c55ca3f8916" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.049737 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.052723 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.052740 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.052952 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-cw9rk" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.053084 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.064694 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.084201 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ck6nm"] Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.108013 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ck6nm"] Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.188218 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-combined-ca-bundle\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.188503 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-public-tls-certs\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.188550 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-internal-tls-certs\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.188739 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-scripts\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.188819 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-config-data\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.188909 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvhnk\" (UniqueName: \"kubernetes.io/projected/feca1cf6-5353-4280-9250-c233f7e2f888-kube-api-access-fvhnk\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.290353 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-scripts\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.290732 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-config-data\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.290874 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvhnk\" (UniqueName: \"kubernetes.io/projected/feca1cf6-5353-4280-9250-c233f7e2f888-kube-api-access-fvhnk\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.291080 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-combined-ca-bundle\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.291255 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-public-tls-certs\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.291340 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-internal-tls-certs\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.297194 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-public-tls-certs\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.299338 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-internal-tls-certs\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.299364 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-scripts\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.300325 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-config-data\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.302816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feca1cf6-5353-4280-9250-c233f7e2f888-combined-ca-bundle\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.308838 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvhnk\" (UniqueName: \"kubernetes.io/projected/feca1cf6-5353-4280-9250-c233f7e2f888-kube-api-access-fvhnk\") pod \"aodh-0\" (UID: \"feca1cf6-5353-4280-9250-c233f7e2f888\") " pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.388015 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 02 13:56:44 crc kubenswrapper[4846]: I0202 13:56:44.977097 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Feb 02 13:56:44 crc kubenswrapper[4846]: W0202 13:56:44.984714 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfeca1cf6_5353_4280_9250_c233f7e2f888.slice/crio-8db06de8b2cf779e7104dd449fba2010cca15c5e707cdd1bed14d143795813aa WatchSource:0}: Error finding container 8db06de8b2cf779e7104dd449fba2010cca15c5e707cdd1bed14d143795813aa: Status 404 returned error can't find the container with id 8db06de8b2cf779e7104dd449fba2010cca15c5e707cdd1bed14d143795813aa Feb 02 13:56:45 crc kubenswrapper[4846]: I0202 13:56:45.438387 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0689d078-b506-4bf6-88e0-34924fe53ea8" path="/var/lib/kubelet/pods/0689d078-b506-4bf6-88e0-34924fe53ea8/volumes" Feb 02 13:56:45 crc kubenswrapper[4846]: I0202 13:56:45.439852 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e8fb480-8530-42db-86ce-35326f86b8da" path="/var/lib/kubelet/pods/2e8fb480-8530-42db-86ce-35326f86b8da/volumes" Feb 02 13:56:45 crc kubenswrapper[4846]: I0202 13:56:45.956844 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"feca1cf6-5353-4280-9250-c233f7e2f888","Type":"ContainerStarted","Data":"21d92cf30285eac28fd3994094ba269e458896aa8681866893c7306a68a9f8c2"} Feb 02 13:56:45 crc kubenswrapper[4846]: I0202 13:56:45.956903 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"feca1cf6-5353-4280-9250-c233f7e2f888","Type":"ContainerStarted","Data":"8db06de8b2cf779e7104dd449fba2010cca15c5e707cdd1bed14d143795813aa"} Feb 02 13:56:46 crc kubenswrapper[4846]: I0202 13:56:46.977249 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"feca1cf6-5353-4280-9250-c233f7e2f888","Type":"ContainerStarted","Data":"06199b528bde2528a8a90228d5aaba9b1cad29cc57d371b39eeb22eac96f8167"} Feb 02 13:56:47 crc kubenswrapper[4846]: I0202 13:56:47.991189 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"feca1cf6-5353-4280-9250-c233f7e2f888","Type":"ContainerStarted","Data":"217395d9efb9923c2eca5da0e1bbcd11e0b1f1217c04169c895f8b9181868edb"} Feb 02 13:56:49 crc kubenswrapper[4846]: I0202 13:56:49.004504 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"feca1cf6-5353-4280-9250-c233f7e2f888","Type":"ContainerStarted","Data":"e138900a4be04019ac20bb18ab532eee25ceb6f1d526072349df4cd770676aed"} Feb 02 13:56:49 crc kubenswrapper[4846]: I0202 13:56:49.031596 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.892699969 podStartE2EDuration="6.031580268s" podCreationTimestamp="2026-02-02 13:56:43 +0000 UTC" firstStartedPulling="2026-02-02 13:56:44.987757022 +0000 UTC m=+6436.216343885" lastFinishedPulling="2026-02-02 13:56:48.126637321 +0000 UTC m=+6439.355224184" observedRunningTime="2026-02-02 13:56:49.031007674 +0000 UTC m=+6440.259594537" watchObservedRunningTime="2026-02-02 13:56:49.031580268 +0000 UTC m=+6440.260167131" Feb 02 13:56:49 crc kubenswrapper[4846]: I0202 13:56:49.435131 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:56:49 crc kubenswrapper[4846]: E0202 13:56:49.435763 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:56:50 crc kubenswrapper[4846]: I0202 13:56:50.340080 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.125809 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54cf44d4c9-mhtp5"] Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.128173 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.130772 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.140545 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54cf44d4c9-mhtp5"] Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.185437 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-config\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.185912 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-dns-svc\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.185999 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-sb\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.186316 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q5sr\" (UniqueName: \"kubernetes.io/projected/8a737962-b7d2-480d-ad6d-037a5efb9d31-kube-api-access-2q5sr\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.186546 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-openstack-cell1\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.186645 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-nb\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.289026 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-dns-svc\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.289082 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-sb\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.289155 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q5sr\" (UniqueName: \"kubernetes.io/projected/8a737962-b7d2-480d-ad6d-037a5efb9d31-kube-api-access-2q5sr\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.289224 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-openstack-cell1\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.289257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-nb\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.289349 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-config\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.290487 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-sb\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.290487 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-dns-svc\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.290594 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-nb\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.290743 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-openstack-cell1\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.291150 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-config\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.318170 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q5sr\" (UniqueName: \"kubernetes.io/projected/8a737962-b7d2-480d-ad6d-037a5efb9d31-kube-api-access-2q5sr\") pod \"dnsmasq-dns-54cf44d4c9-mhtp5\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:52 crc kubenswrapper[4846]: I0202 13:56:52.459601 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:53 crc kubenswrapper[4846]: I0202 13:56:53.086120 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54cf44d4c9-mhtp5"] Feb 02 13:56:54 crc kubenswrapper[4846]: I0202 13:56:54.063851 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a737962-b7d2-480d-ad6d-037a5efb9d31" containerID="ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b" exitCode=0 Feb 02 13:56:54 crc kubenswrapper[4846]: I0202 13:56:54.064433 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" event={"ID":"8a737962-b7d2-480d-ad6d-037a5efb9d31","Type":"ContainerDied","Data":"ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b"} Feb 02 13:56:54 crc kubenswrapper[4846]: I0202 13:56:54.064477 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" event={"ID":"8a737962-b7d2-480d-ad6d-037a5efb9d31","Type":"ContainerStarted","Data":"84d6deaf0b60c1cde1f15ed43eca91d134ce2f0f681bc31bcd3c78eec36e732a"} Feb 02 13:56:55 crc kubenswrapper[4846]: I0202 13:56:55.076038 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" event={"ID":"8a737962-b7d2-480d-ad6d-037a5efb9d31","Type":"ContainerStarted","Data":"308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b"} Feb 02 13:56:55 crc kubenswrapper[4846]: I0202 13:56:55.076695 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:56:55 crc kubenswrapper[4846]: I0202 13:56:55.095189 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" podStartSLOduration=3.095164414 podStartE2EDuration="3.095164414s" podCreationTimestamp="2026-02-02 13:56:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:56:55.095145584 +0000 UTC m=+6446.323732457" watchObservedRunningTime="2026-02-02 13:56:55.095164414 +0000 UTC m=+6446.323751277" Feb 02 13:57:00 crc kubenswrapper[4846]: I0202 13:57:00.423260 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:57:00 crc kubenswrapper[4846]: E0202 13:57:00.425046 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.063347 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-t45x7"] Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.076160 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-t45x7"] Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.460805 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.543828 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bbcfcc989-9vcqg"] Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.544077 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" podUID="3c239fe3-c133-4207-bc70-846b34c7859d" containerName="dnsmasq-dns" containerID="cri-o://610b151d0fabfd65c7b70a049a079ee2378e097c251eeebc0aaf8a165258dfa9" gracePeriod=10 Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.763971 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d4cd97679-9wtr9"] Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.769035 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.784169 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d4cd97679-9wtr9"] Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.850484 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg8ds\" (UniqueName: \"kubernetes.io/projected/eea30013-7cac-4157-b39e-da5f71999fba-kube-api-access-qg8ds\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.850549 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-config\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.850579 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-openstack-cell1\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.850602 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-ovsdbserver-sb\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.850979 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-ovsdbserver-nb\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.851516 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-dns-svc\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.956065 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg8ds\" (UniqueName: \"kubernetes.io/projected/eea30013-7cac-4157-b39e-da5f71999fba-kube-api-access-qg8ds\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.956157 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-config\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.956191 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-openstack-cell1\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.956219 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-ovsdbserver-sb\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.956288 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-ovsdbserver-nb\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.956379 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-dns-svc\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.958352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-dns-svc\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.960151 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-config\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.960888 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-openstack-cell1\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.962081 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-ovsdbserver-sb\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.963314 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eea30013-7cac-4157-b39e-da5f71999fba-ovsdbserver-nb\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:02 crc kubenswrapper[4846]: I0202 13:57:02.983354 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg8ds\" (UniqueName: \"kubernetes.io/projected/eea30013-7cac-4157-b39e-da5f71999fba-kube-api-access-qg8ds\") pod \"dnsmasq-dns-7d4cd97679-9wtr9\" (UID: \"eea30013-7cac-4157-b39e-da5f71999fba\") " pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.093194 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.158029 4846 generic.go:334] "Generic (PLEG): container finished" podID="3c239fe3-c133-4207-bc70-846b34c7859d" containerID="610b151d0fabfd65c7b70a049a079ee2378e097c251eeebc0aaf8a165258dfa9" exitCode=0 Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.158078 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" event={"ID":"3c239fe3-c133-4207-bc70-846b34c7859d","Type":"ContainerDied","Data":"610b151d0fabfd65c7b70a049a079ee2378e097c251eeebc0aaf8a165258dfa9"} Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.158110 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" event={"ID":"3c239fe3-c133-4207-bc70-846b34c7859d","Type":"ContainerDied","Data":"31d9a5f42b6af94b78d93773e076df22cc1db37a53a97496027d2118b8edc5f8"} Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.158126 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31d9a5f42b6af94b78d93773e076df22cc1db37a53a97496027d2118b8edc5f8" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.246480 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.364398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-sb\") pod \"3c239fe3-c133-4207-bc70-846b34c7859d\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.364510 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-dns-svc\") pod \"3c239fe3-c133-4207-bc70-846b34c7859d\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.364763 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsqsn\" (UniqueName: \"kubernetes.io/projected/3c239fe3-c133-4207-bc70-846b34c7859d-kube-api-access-rsqsn\") pod \"3c239fe3-c133-4207-bc70-846b34c7859d\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.364876 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-nb\") pod \"3c239fe3-c133-4207-bc70-846b34c7859d\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.364985 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-config\") pod \"3c239fe3-c133-4207-bc70-846b34c7859d\" (UID: \"3c239fe3-c133-4207-bc70-846b34c7859d\") " Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.439255 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c239fe3-c133-4207-bc70-846b34c7859d-kube-api-access-rsqsn" (OuterVolumeSpecName: "kube-api-access-rsqsn") pod "3c239fe3-c133-4207-bc70-846b34c7859d" (UID: "3c239fe3-c133-4207-bc70-846b34c7859d"). InnerVolumeSpecName "kube-api-access-rsqsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.467573 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsqsn\" (UniqueName: \"kubernetes.io/projected/3c239fe3-c133-4207-bc70-846b34c7859d-kube-api-access-rsqsn\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.550542 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="587b6a35-3ed8-4a91-8c38-daefcb8a244e" path="/var/lib/kubelet/pods/587b6a35-3ed8-4a91-8c38-daefcb8a244e/volumes" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.584376 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3c239fe3-c133-4207-bc70-846b34c7859d" (UID: "3c239fe3-c133-4207-bc70-846b34c7859d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.616828 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3c239fe3-c133-4207-bc70-846b34c7859d" (UID: "3c239fe3-c133-4207-bc70-846b34c7859d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.625443 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-config" (OuterVolumeSpecName: "config") pod "3c239fe3-c133-4207-bc70-846b34c7859d" (UID: "3c239fe3-c133-4207-bc70-846b34c7859d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.646761 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3c239fe3-c133-4207-bc70-846b34c7859d" (UID: "3c239fe3-c133-4207-bc70-846b34c7859d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.677019 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.677055 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.677065 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.677075 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c239fe3-c133-4207-bc70-846b34c7859d-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:03 crc kubenswrapper[4846]: I0202 13:57:03.780743 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d4cd97679-9wtr9"] Feb 02 13:57:04 crc kubenswrapper[4846]: I0202 13:57:04.045531 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-hgf4w"] Feb 02 13:57:04 crc kubenswrapper[4846]: I0202 13:57:04.059301 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-hgf4w"] Feb 02 13:57:04 crc kubenswrapper[4846]: I0202 13:57:04.170799 4846 generic.go:334] "Generic (PLEG): container finished" podID="eea30013-7cac-4157-b39e-da5f71999fba" containerID="63ef8ef1951edd75fd9545951d7c34b1353df1f7d3b95b95c16df7ae89c93160" exitCode=0 Feb 02 13:57:04 crc kubenswrapper[4846]: I0202 13:57:04.170895 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bbcfcc989-9vcqg" Feb 02 13:57:04 crc kubenswrapper[4846]: I0202 13:57:04.171925 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" event={"ID":"eea30013-7cac-4157-b39e-da5f71999fba","Type":"ContainerDied","Data":"63ef8ef1951edd75fd9545951d7c34b1353df1f7d3b95b95c16df7ae89c93160"} Feb 02 13:57:04 crc kubenswrapper[4846]: I0202 13:57:04.172021 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" event={"ID":"eea30013-7cac-4157-b39e-da5f71999fba","Type":"ContainerStarted","Data":"0544a8f454fb3a4a398293c0e0748494b68f2fa6f1261f0619e7f0bf2705f44b"} Feb 02 13:57:04 crc kubenswrapper[4846]: I0202 13:57:04.263889 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bbcfcc989-9vcqg"] Feb 02 13:57:04 crc kubenswrapper[4846]: I0202 13:57:04.274213 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bbcfcc989-9vcqg"] Feb 02 13:57:05 crc kubenswrapper[4846]: I0202 13:57:05.180781 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" event={"ID":"eea30013-7cac-4157-b39e-da5f71999fba","Type":"ContainerStarted","Data":"8a8efb773695652a2fbe3cb4d4cbe48b591332795130d0060dfeffaeb5e40127"} Feb 02 13:57:05 crc kubenswrapper[4846]: I0202 13:57:05.181915 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:05 crc kubenswrapper[4846]: I0202 13:57:05.208200 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" podStartSLOduration=3.208176802 podStartE2EDuration="3.208176802s" podCreationTimestamp="2026-02-02 13:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 13:57:05.197504012 +0000 UTC m=+6456.426090875" watchObservedRunningTime="2026-02-02 13:57:05.208176802 +0000 UTC m=+6456.436763665" Feb 02 13:57:05 crc kubenswrapper[4846]: I0202 13:57:05.438217 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c239fe3-c133-4207-bc70-846b34c7859d" path="/var/lib/kubelet/pods/3c239fe3-c133-4207-bc70-846b34c7859d/volumes" Feb 02 13:57:05 crc kubenswrapper[4846]: I0202 13:57:05.439194 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8904d23d-2e73-4eb8-b6f6-8d559d0365e6" path="/var/lib/kubelet/pods/8904d23d-2e73-4eb8-b6f6-8d559d0365e6/volumes" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.096526 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d4cd97679-9wtr9" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.168234 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54cf44d4c9-mhtp5"] Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.168872 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" podUID="8a737962-b7d2-480d-ad6d-037a5efb9d31" containerName="dnsmasq-dns" containerID="cri-o://308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b" gracePeriod=10 Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.753571 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.830226 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-dns-svc\") pod \"8a737962-b7d2-480d-ad6d-037a5efb9d31\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.830296 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-sb\") pod \"8a737962-b7d2-480d-ad6d-037a5efb9d31\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.830400 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-config\") pod \"8a737962-b7d2-480d-ad6d-037a5efb9d31\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.830442 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-openstack-cell1\") pod \"8a737962-b7d2-480d-ad6d-037a5efb9d31\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.830520 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-nb\") pod \"8a737962-b7d2-480d-ad6d-037a5efb9d31\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.830724 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2q5sr\" (UniqueName: \"kubernetes.io/projected/8a737962-b7d2-480d-ad6d-037a5efb9d31-kube-api-access-2q5sr\") pod \"8a737962-b7d2-480d-ad6d-037a5efb9d31\" (UID: \"8a737962-b7d2-480d-ad6d-037a5efb9d31\") " Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.835480 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a737962-b7d2-480d-ad6d-037a5efb9d31-kube-api-access-2q5sr" (OuterVolumeSpecName: "kube-api-access-2q5sr") pod "8a737962-b7d2-480d-ad6d-037a5efb9d31" (UID: "8a737962-b7d2-480d-ad6d-037a5efb9d31"). InnerVolumeSpecName "kube-api-access-2q5sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.893850 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "8a737962-b7d2-480d-ad6d-037a5efb9d31" (UID: "8a737962-b7d2-480d-ad6d-037a5efb9d31"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.904079 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-config" (OuterVolumeSpecName: "config") pod "8a737962-b7d2-480d-ad6d-037a5efb9d31" (UID: "8a737962-b7d2-480d-ad6d-037a5efb9d31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.904404 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8a737962-b7d2-480d-ad6d-037a5efb9d31" (UID: "8a737962-b7d2-480d-ad6d-037a5efb9d31"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.904473 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8a737962-b7d2-480d-ad6d-037a5efb9d31" (UID: "8a737962-b7d2-480d-ad6d-037a5efb9d31"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.904668 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8a737962-b7d2-480d-ad6d-037a5efb9d31" (UID: "8a737962-b7d2-480d-ad6d-037a5efb9d31"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.932877 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.932921 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.932937 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-config\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.932949 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.932959 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a737962-b7d2-480d-ad6d-037a5efb9d31-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:13 crc kubenswrapper[4846]: I0202 13:57:13.932969 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2q5sr\" (UniqueName: \"kubernetes.io/projected/8a737962-b7d2-480d-ad6d-037a5efb9d31-kube-api-access-2q5sr\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.270703 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a737962-b7d2-480d-ad6d-037a5efb9d31" containerID="308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b" exitCode=0 Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.270749 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" event={"ID":"8a737962-b7d2-480d-ad6d-037a5efb9d31","Type":"ContainerDied","Data":"308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b"} Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.270768 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.270790 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54cf44d4c9-mhtp5" event={"ID":"8a737962-b7d2-480d-ad6d-037a5efb9d31","Type":"ContainerDied","Data":"84d6deaf0b60c1cde1f15ed43eca91d134ce2f0f681bc31bcd3c78eec36e732a"} Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.270807 4846 scope.go:117] "RemoveContainer" containerID="308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b" Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.315790 4846 scope.go:117] "RemoveContainer" containerID="ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b" Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.389339 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54cf44d4c9-mhtp5"] Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.393972 4846 scope.go:117] "RemoveContainer" containerID="308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b" Feb 02 13:57:14 crc kubenswrapper[4846]: E0202 13:57:14.394700 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b\": container with ID starting with 308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b not found: ID does not exist" containerID="308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b" Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.394753 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b"} err="failed to get container status \"308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b\": rpc error: code = NotFound desc = could not find container \"308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b\": container with ID starting with 308184023fedb0672e2cd39a3c29e8459544fa8ac774ccf6c39cd62cc1585f6b not found: ID does not exist" Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.394787 4846 scope.go:117] "RemoveContainer" containerID="ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b" Feb 02 13:57:14 crc kubenswrapper[4846]: E0202 13:57:14.397266 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b\": container with ID starting with ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b not found: ID does not exist" containerID="ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b" Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.397309 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b"} err="failed to get container status \"ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b\": rpc error: code = NotFound desc = could not find container \"ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b\": container with ID starting with ef112924888d93eb798db45c1354f5aa3ed89e6ba88054987d10791ac96a3f2b not found: ID does not exist" Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.399424 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54cf44d4c9-mhtp5"] Feb 02 13:57:14 crc kubenswrapper[4846]: I0202 13:57:14.424296 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:57:14 crc kubenswrapper[4846]: E0202 13:57:14.424580 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:57:15 crc kubenswrapper[4846]: I0202 13:57:15.439341 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a737962-b7d2-480d-ad6d-037a5efb9d31" path="/var/lib/kubelet/pods/8a737962-b7d2-480d-ad6d-037a5efb9d31/volumes" Feb 02 13:57:20 crc kubenswrapper[4846]: I0202 13:57:20.044404 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-6b2jw"] Feb 02 13:57:20 crc kubenswrapper[4846]: I0202 13:57:20.054285 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-6b2jw"] Feb 02 13:57:21 crc kubenswrapper[4846]: I0202 13:57:21.454015 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fddbbaf0-81be-43a4-8441-dcdf4d9b83c0" path="/var/lib/kubelet/pods/fddbbaf0-81be-43a4-8441-dcdf4d9b83c0/volumes" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.677955 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn"] Feb 02 13:57:23 crc kubenswrapper[4846]: E0202 13:57:23.679279 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c239fe3-c133-4207-bc70-846b34c7859d" containerName="init" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.679300 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c239fe3-c133-4207-bc70-846b34c7859d" containerName="init" Feb 02 13:57:23 crc kubenswrapper[4846]: E0202 13:57:23.679330 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a737962-b7d2-480d-ad6d-037a5efb9d31" containerName="init" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.679338 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a737962-b7d2-480d-ad6d-037a5efb9d31" containerName="init" Feb 02 13:57:23 crc kubenswrapper[4846]: E0202 13:57:23.679359 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c239fe3-c133-4207-bc70-846b34c7859d" containerName="dnsmasq-dns" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.679367 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c239fe3-c133-4207-bc70-846b34c7859d" containerName="dnsmasq-dns" Feb 02 13:57:23 crc kubenswrapper[4846]: E0202 13:57:23.679381 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a737962-b7d2-480d-ad6d-037a5efb9d31" containerName="dnsmasq-dns" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.679388 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a737962-b7d2-480d-ad6d-037a5efb9d31" containerName="dnsmasq-dns" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.679712 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c239fe3-c133-4207-bc70-846b34c7859d" containerName="dnsmasq-dns" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.679754 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a737962-b7d2-480d-ad6d-037a5efb9d31" containerName="dnsmasq-dns" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.680680 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.684674 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.686136 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.686354 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.686515 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.698724 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn"] Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.771533 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.771680 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qm79\" (UniqueName: \"kubernetes.io/projected/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-kube-api-access-2qm79\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.771731 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.771795 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.874223 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qm79\" (UniqueName: \"kubernetes.io/projected/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-kube-api-access-2qm79\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.874310 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.874373 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.875646 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.881795 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.884250 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.897747 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:23 crc kubenswrapper[4846]: I0202 13:57:23.910475 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qm79\" (UniqueName: \"kubernetes.io/projected/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-kube-api-access-2qm79\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:24 crc kubenswrapper[4846]: I0202 13:57:24.038220 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:24 crc kubenswrapper[4846]: I0202 13:57:24.719224 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn"] Feb 02 13:57:25 crc kubenswrapper[4846]: I0202 13:57:25.393410 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" event={"ID":"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6","Type":"ContainerStarted","Data":"b268a223a6c99142b5e35e11add1fcae8c6630677e750c8c3708676dd8caf46b"} Feb 02 13:57:25 crc kubenswrapper[4846]: I0202 13:57:25.423499 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:57:25 crc kubenswrapper[4846]: E0202 13:57:25.423806 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:57:29 crc kubenswrapper[4846]: I0202 13:57:29.951926 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z2zw8"] Feb 02 13:57:29 crc kubenswrapper[4846]: I0202 13:57:29.955510 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.041410 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2zw8"] Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.143762 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-utilities\") pod \"certified-operators-z2zw8\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.143933 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-catalog-content\") pod \"certified-operators-z2zw8\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.144002 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n567j\" (UniqueName: \"kubernetes.io/projected/451fd5d7-ea92-4305-8c49-7d54646a8c78-kube-api-access-n567j\") pod \"certified-operators-z2zw8\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.245869 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-catalog-content\") pod \"certified-operators-z2zw8\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.245962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n567j\" (UniqueName: \"kubernetes.io/projected/451fd5d7-ea92-4305-8c49-7d54646a8c78-kube-api-access-n567j\") pod \"certified-operators-z2zw8\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.246138 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-utilities\") pod \"certified-operators-z2zw8\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.246827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-utilities\") pod \"certified-operators-z2zw8\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.246827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-catalog-content\") pod \"certified-operators-z2zw8\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.281716 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n567j\" (UniqueName: \"kubernetes.io/projected/451fd5d7-ea92-4305-8c49-7d54646a8c78-kube-api-access-n567j\") pod \"certified-operators-z2zw8\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:30 crc kubenswrapper[4846]: I0202 13:57:30.376932 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:36 crc kubenswrapper[4846]: I0202 13:57:36.110092 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 13:57:36 crc kubenswrapper[4846]: I0202 13:57:36.523253 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" event={"ID":"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6","Type":"ContainerStarted","Data":"1c1a2e6ced557e2cfccdb9bf8c5d7e0f1e13029ca78aff19e46955ade53a6728"} Feb 02 13:57:36 crc kubenswrapper[4846]: I0202 13:57:36.542926 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" podStartSLOduration=2.165537409 podStartE2EDuration="13.542904672s" podCreationTimestamp="2026-02-02 13:57:23 +0000 UTC" firstStartedPulling="2026-02-02 13:57:24.729870999 +0000 UTC m=+6475.958457862" lastFinishedPulling="2026-02-02 13:57:36.107238262 +0000 UTC m=+6487.335825125" observedRunningTime="2026-02-02 13:57:36.536465477 +0000 UTC m=+6487.765052400" watchObservedRunningTime="2026-02-02 13:57:36.542904672 +0000 UTC m=+6487.771491535" Feb 02 13:57:36 crc kubenswrapper[4846]: I0202 13:57:36.614724 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2zw8"] Feb 02 13:57:36 crc kubenswrapper[4846]: W0202 13:57:36.625869 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod451fd5d7_ea92_4305_8c49_7d54646a8c78.slice/crio-4ec5e552cc405445967975e185dc5e13dbd34a1044ac8bcf1b07f2cd280c4daa WatchSource:0}: Error finding container 4ec5e552cc405445967975e185dc5e13dbd34a1044ac8bcf1b07f2cd280c4daa: Status 404 returned error can't find the container with id 4ec5e552cc405445967975e185dc5e13dbd34a1044ac8bcf1b07f2cd280c4daa Feb 02 13:57:37 crc kubenswrapper[4846]: I0202 13:57:37.533257 4846 generic.go:334] "Generic (PLEG): container finished" podID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerID="aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14" exitCode=0 Feb 02 13:57:37 crc kubenswrapper[4846]: I0202 13:57:37.533315 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zw8" event={"ID":"451fd5d7-ea92-4305-8c49-7d54646a8c78","Type":"ContainerDied","Data":"aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14"} Feb 02 13:57:37 crc kubenswrapper[4846]: I0202 13:57:37.533550 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zw8" event={"ID":"451fd5d7-ea92-4305-8c49-7d54646a8c78","Type":"ContainerStarted","Data":"4ec5e552cc405445967975e185dc5e13dbd34a1044ac8bcf1b07f2cd280c4daa"} Feb 02 13:57:39 crc kubenswrapper[4846]: I0202 13:57:39.431467 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:57:39 crc kubenswrapper[4846]: E0202 13:57:39.432428 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:57:41 crc kubenswrapper[4846]: I0202 13:57:41.855096 4846 scope.go:117] "RemoveContainer" containerID="f15402751c443c63c555370f43747b6463389e32f6f111ab8230b1c238f9f2e2" Feb 02 13:57:41 crc kubenswrapper[4846]: I0202 13:57:41.923669 4846 scope.go:117] "RemoveContainer" containerID="7fd22820549e8a0ab310981d03f95edd3178f01653a9cc5680b5fbffd915d8b2" Feb 02 13:57:42 crc kubenswrapper[4846]: I0202 13:57:42.046714 4846 scope.go:117] "RemoveContainer" containerID="da476008128d2691003257fc23608aa5cc53e25a6d2fccac873c7220cbd57440" Feb 02 13:57:42 crc kubenswrapper[4846]: I0202 13:57:42.125632 4846 scope.go:117] "RemoveContainer" containerID="8d1edba65159d50673e4e593a3f7a608fd860b9d5b8b72df9c25acb668c97010" Feb 02 13:57:42 crc kubenswrapper[4846]: I0202 13:57:42.174724 4846 scope.go:117] "RemoveContainer" containerID="610b151d0fabfd65c7b70a049a079ee2378e097c251eeebc0aaf8a165258dfa9" Feb 02 13:57:42 crc kubenswrapper[4846]: I0202 13:57:42.219607 4846 scope.go:117] "RemoveContainer" containerID="c771d54ac604b5cf688d25d00e2566e75cdb141e7feca723b8a949b372c3f03a" Feb 02 13:57:42 crc kubenswrapper[4846]: I0202 13:57:42.581993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zw8" event={"ID":"451fd5d7-ea92-4305-8c49-7d54646a8c78","Type":"ContainerStarted","Data":"2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf"} Feb 02 13:57:44 crc kubenswrapper[4846]: I0202 13:57:44.606152 4846 generic.go:334] "Generic (PLEG): container finished" podID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerID="2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf" exitCode=0 Feb 02 13:57:44 crc kubenswrapper[4846]: I0202 13:57:44.606249 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zw8" event={"ID":"451fd5d7-ea92-4305-8c49-7d54646a8c78","Type":"ContainerDied","Data":"2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf"} Feb 02 13:57:45 crc kubenswrapper[4846]: I0202 13:57:45.618852 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zw8" event={"ID":"451fd5d7-ea92-4305-8c49-7d54646a8c78","Type":"ContainerStarted","Data":"9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e"} Feb 02 13:57:45 crc kubenswrapper[4846]: I0202 13:57:45.642178 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z2zw8" podStartSLOduration=9.1487191 podStartE2EDuration="16.642159421s" podCreationTimestamp="2026-02-02 13:57:29 +0000 UTC" firstStartedPulling="2026-02-02 13:57:37.535009796 +0000 UTC m=+6488.763596659" lastFinishedPulling="2026-02-02 13:57:45.028450117 +0000 UTC m=+6496.257036980" observedRunningTime="2026-02-02 13:57:45.636265138 +0000 UTC m=+6496.864852011" watchObservedRunningTime="2026-02-02 13:57:45.642159421 +0000 UTC m=+6496.870746284" Feb 02 13:57:49 crc kubenswrapper[4846]: I0202 13:57:49.655376 4846 generic.go:334] "Generic (PLEG): container finished" podID="1f33b8f6-a26a-4052-aa57-b0ce0d2924c6" containerID="1c1a2e6ced557e2cfccdb9bf8c5d7e0f1e13029ca78aff19e46955ade53a6728" exitCode=0 Feb 02 13:57:49 crc kubenswrapper[4846]: I0202 13:57:49.655466 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" event={"ID":"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6","Type":"ContainerDied","Data":"1c1a2e6ced557e2cfccdb9bf8c5d7e0f1e13029ca78aff19e46955ade53a6728"} Feb 02 13:57:50 crc kubenswrapper[4846]: I0202 13:57:50.378111 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:50 crc kubenswrapper[4846]: I0202 13:57:50.378821 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:50 crc kubenswrapper[4846]: I0202 13:57:50.429220 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:50 crc kubenswrapper[4846]: I0202 13:57:50.742929 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:50 crc kubenswrapper[4846]: I0202 13:57:50.791128 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2zw8"] Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.175040 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.371460 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-ssh-key-openstack-cell1\") pod \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.371521 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-inventory\") pod \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.371552 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qm79\" (UniqueName: \"kubernetes.io/projected/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-kube-api-access-2qm79\") pod \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.371574 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-pre-adoption-validation-combined-ca-bundle\") pod \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\" (UID: \"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6\") " Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.379922 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "1f33b8f6-a26a-4052-aa57-b0ce0d2924c6" (UID: "1f33b8f6-a26a-4052-aa57-b0ce0d2924c6"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.380060 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-kube-api-access-2qm79" (OuterVolumeSpecName: "kube-api-access-2qm79") pod "1f33b8f6-a26a-4052-aa57-b0ce0d2924c6" (UID: "1f33b8f6-a26a-4052-aa57-b0ce0d2924c6"). InnerVolumeSpecName "kube-api-access-2qm79". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.404589 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "1f33b8f6-a26a-4052-aa57-b0ce0d2924c6" (UID: "1f33b8f6-a26a-4052-aa57-b0ce0d2924c6"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.413681 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-inventory" (OuterVolumeSpecName: "inventory") pod "1f33b8f6-a26a-4052-aa57-b0ce0d2924c6" (UID: "1f33b8f6-a26a-4052-aa57-b0ce0d2924c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.474216 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.474255 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.474271 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qm79\" (UniqueName: \"kubernetes.io/projected/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-kube-api-access-2qm79\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.474286 4846 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f33b8f6-a26a-4052-aa57-b0ce0d2924c6-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.675801 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" event={"ID":"1f33b8f6-a26a-4052-aa57-b0ce0d2924c6","Type":"ContainerDied","Data":"b268a223a6c99142b5e35e11add1fcae8c6630677e750c8c3708676dd8caf46b"} Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.676371 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b268a223a6c99142b5e35e11add1fcae8c6630677e750c8c3708676dd8caf46b" Feb 02 13:57:51 crc kubenswrapper[4846]: I0202 13:57:51.675871 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn" Feb 02 13:57:52 crc kubenswrapper[4846]: I0202 13:57:52.682709 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z2zw8" podUID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerName="registry-server" containerID="cri-o://9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e" gracePeriod=2 Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.197531 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.313586 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n567j\" (UniqueName: \"kubernetes.io/projected/451fd5d7-ea92-4305-8c49-7d54646a8c78-kube-api-access-n567j\") pod \"451fd5d7-ea92-4305-8c49-7d54646a8c78\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.313806 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-catalog-content\") pod \"451fd5d7-ea92-4305-8c49-7d54646a8c78\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.314062 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-utilities\") pod \"451fd5d7-ea92-4305-8c49-7d54646a8c78\" (UID: \"451fd5d7-ea92-4305-8c49-7d54646a8c78\") " Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.314632 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-utilities" (OuterVolumeSpecName: "utilities") pod "451fd5d7-ea92-4305-8c49-7d54646a8c78" (UID: "451fd5d7-ea92-4305-8c49-7d54646a8c78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.315190 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.319683 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/451fd5d7-ea92-4305-8c49-7d54646a8c78-kube-api-access-n567j" (OuterVolumeSpecName: "kube-api-access-n567j") pod "451fd5d7-ea92-4305-8c49-7d54646a8c78" (UID: "451fd5d7-ea92-4305-8c49-7d54646a8c78"). InnerVolumeSpecName "kube-api-access-n567j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.366660 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "451fd5d7-ea92-4305-8c49-7d54646a8c78" (UID: "451fd5d7-ea92-4305-8c49-7d54646a8c78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.418680 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n567j\" (UniqueName: \"kubernetes.io/projected/451fd5d7-ea92-4305-8c49-7d54646a8c78-kube-api-access-n567j\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.418724 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/451fd5d7-ea92-4305-8c49-7d54646a8c78-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.692487 4846 generic.go:334] "Generic (PLEG): container finished" podID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerID="9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e" exitCode=0 Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.692536 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zw8" event={"ID":"451fd5d7-ea92-4305-8c49-7d54646a8c78","Type":"ContainerDied","Data":"9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e"} Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.692568 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zw8" event={"ID":"451fd5d7-ea92-4305-8c49-7d54646a8c78","Type":"ContainerDied","Data":"4ec5e552cc405445967975e185dc5e13dbd34a1044ac8bcf1b07f2cd280c4daa"} Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.692589 4846 scope.go:117] "RemoveContainer" containerID="9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.692598 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2zw8" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.724344 4846 scope.go:117] "RemoveContainer" containerID="2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.729273 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2zw8"] Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.740315 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z2zw8"] Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.748532 4846 scope.go:117] "RemoveContainer" containerID="aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.824298 4846 scope.go:117] "RemoveContainer" containerID="9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e" Feb 02 13:57:53 crc kubenswrapper[4846]: E0202 13:57:53.824936 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e\": container with ID starting with 9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e not found: ID does not exist" containerID="9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.825039 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e"} err="failed to get container status \"9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e\": rpc error: code = NotFound desc = could not find container \"9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e\": container with ID starting with 9e021f5a92b2334a09f5008989accb762acc7f5c8410668abbb4a69c948bee1e not found: ID does not exist" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.825075 4846 scope.go:117] "RemoveContainer" containerID="2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf" Feb 02 13:57:53 crc kubenswrapper[4846]: E0202 13:57:53.825478 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf\": container with ID starting with 2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf not found: ID does not exist" containerID="2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.825515 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf"} err="failed to get container status \"2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf\": rpc error: code = NotFound desc = could not find container \"2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf\": container with ID starting with 2c7e7c732bcfdce727aa16dd044c0d2355339830060d9ce96efa272bc66a45bf not found: ID does not exist" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.825535 4846 scope.go:117] "RemoveContainer" containerID="aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14" Feb 02 13:57:53 crc kubenswrapper[4846]: E0202 13:57:53.825840 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14\": container with ID starting with aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14 not found: ID does not exist" containerID="aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14" Feb 02 13:57:53 crc kubenswrapper[4846]: I0202 13:57:53.825866 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14"} err="failed to get container status \"aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14\": rpc error: code = NotFound desc = could not find container \"aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14\": container with ID starting with aa62d5b2e1a371453ab3bf7aa28401937c20d2074870e9eeec39f8d1f72dbb14 not found: ID does not exist" Feb 02 13:57:54 crc kubenswrapper[4846]: I0202 13:57:54.425015 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:57:54 crc kubenswrapper[4846]: E0202 13:57:54.425459 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:57:55 crc kubenswrapper[4846]: I0202 13:57:55.441998 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="451fd5d7-ea92-4305-8c49-7d54646a8c78" path="/var/lib/kubelet/pods/451fd5d7-ea92-4305-8c49-7d54646a8c78/volumes" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.356997 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw"] Feb 02 13:57:56 crc kubenswrapper[4846]: E0202 13:57:56.357549 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerName="registry-server" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.357572 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerName="registry-server" Feb 02 13:57:56 crc kubenswrapper[4846]: E0202 13:57:56.357591 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f33b8f6-a26a-4052-aa57-b0ce0d2924c6" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.357601 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f33b8f6-a26a-4052-aa57-b0ce0d2924c6" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Feb 02 13:57:56 crc kubenswrapper[4846]: E0202 13:57:56.357638 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerName="extract-utilities" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.357648 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerName="extract-utilities" Feb 02 13:57:56 crc kubenswrapper[4846]: E0202 13:57:56.357665 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerName="extract-content" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.357673 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerName="extract-content" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.357928 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f33b8f6-a26a-4052-aa57-b0ce0d2924c6" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.357965 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="451fd5d7-ea92-4305-8c49-7d54646a8c78" containerName="registry-server" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.359147 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.361982 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.362330 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.362766 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.363029 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.373530 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw"] Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.492454 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.492555 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.492856 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.493075 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzrdb\" (UniqueName: \"kubernetes.io/projected/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-kube-api-access-rzrdb\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.595174 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.595238 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.595333 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.595415 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzrdb\" (UniqueName: \"kubernetes.io/projected/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-kube-api-access-rzrdb\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.602549 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.603001 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.603359 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.618348 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzrdb\" (UniqueName: \"kubernetes.io/projected/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-kube-api-access-rzrdb\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:56 crc kubenswrapper[4846]: I0202 13:57:56.694472 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 13:57:57 crc kubenswrapper[4846]: I0202 13:57:57.272479 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw"] Feb 02 13:57:57 crc kubenswrapper[4846]: I0202 13:57:57.746963 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" event={"ID":"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef","Type":"ContainerStarted","Data":"92dc655a7274c2ac58b73d6f6fd85c228f18caeaf11d53e9781be1e922bc3b0c"} Feb 02 13:57:58 crc kubenswrapper[4846]: I0202 13:57:58.757414 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" event={"ID":"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef","Type":"ContainerStarted","Data":"587b813fae819d8b8fde8198cc0cc36b691b3e3dabd949fa8959d2c5255ebe2a"} Feb 02 13:57:58 crc kubenswrapper[4846]: I0202 13:57:58.778272 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" podStartSLOduration=2.361013332 podStartE2EDuration="2.778250465s" podCreationTimestamp="2026-02-02 13:57:56 +0000 UTC" firstStartedPulling="2026-02-02 13:57:57.271427642 +0000 UTC m=+6508.500014505" lastFinishedPulling="2026-02-02 13:57:57.688664775 +0000 UTC m=+6508.917251638" observedRunningTime="2026-02-02 13:57:58.769548394 +0000 UTC m=+6509.998135267" watchObservedRunningTime="2026-02-02 13:57:58.778250465 +0000 UTC m=+6510.006837328" Feb 02 13:58:07 crc kubenswrapper[4846]: I0202 13:58:07.423741 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:58:07 crc kubenswrapper[4846]: E0202 13:58:07.424694 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:58:20 crc kubenswrapper[4846]: I0202 13:58:20.424497 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:58:20 crc kubenswrapper[4846]: E0202 13:58:20.425828 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:58:35 crc kubenswrapper[4846]: I0202 13:58:35.424241 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:58:35 crc kubenswrapper[4846]: E0202 13:58:35.425378 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:58:42 crc kubenswrapper[4846]: I0202 13:58:42.421344 4846 scope.go:117] "RemoveContainer" containerID="5f9d52f18a37a63a4bc28d032601d1879bbf315c55d1fe5db1b5650199b27dce" Feb 02 13:58:42 crc kubenswrapper[4846]: I0202 13:58:42.627681 4846 scope.go:117] "RemoveContainer" containerID="ddfad6f8cdba946ad016e61bd42dc8962bec94675a6cdbbc07cc91eb2c9ebc9d" Feb 02 13:58:45 crc kubenswrapper[4846]: I0202 13:58:45.041105 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-9v7ph"] Feb 02 13:58:45 crc kubenswrapper[4846]: I0202 13:58:45.051303 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-9v7ph"] Feb 02 13:58:45 crc kubenswrapper[4846]: I0202 13:58:45.440240 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf" path="/var/lib/kubelet/pods/5715b2e0-db0a-4ccf-ba6f-a6bed9b716cf/volumes" Feb 02 13:58:46 crc kubenswrapper[4846]: I0202 13:58:46.031769 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-06a1-account-create-update-jj9jt"] Feb 02 13:58:46 crc kubenswrapper[4846]: I0202 13:58:46.044784 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-06a1-account-create-update-jj9jt"] Feb 02 13:58:47 crc kubenswrapper[4846]: I0202 13:58:47.441340 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94bb9e87-02ea-4165-bb0a-00acfd0c4c77" path="/var/lib/kubelet/pods/94bb9e87-02ea-4165-bb0a-00acfd0c4c77/volumes" Feb 02 13:58:48 crc kubenswrapper[4846]: I0202 13:58:48.424367 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:58:48 crc kubenswrapper[4846]: E0202 13:58:48.424663 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:58:51 crc kubenswrapper[4846]: I0202 13:58:51.034229 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-nr6g7"] Feb 02 13:58:51 crc kubenswrapper[4846]: I0202 13:58:51.044329 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-nr6g7"] Feb 02 13:58:51 crc kubenswrapper[4846]: I0202 13:58:51.435777 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c30f1bf7-8638-4199-a977-fbab41f1dad2" path="/var/lib/kubelet/pods/c30f1bf7-8638-4199-a977-fbab41f1dad2/volumes" Feb 02 13:58:52 crc kubenswrapper[4846]: I0202 13:58:52.037757 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-21cb-account-create-update-p9275"] Feb 02 13:58:52 crc kubenswrapper[4846]: I0202 13:58:52.049392 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-21cb-account-create-update-p9275"] Feb 02 13:58:53 crc kubenswrapper[4846]: I0202 13:58:53.437582 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9155b41a-6ea6-4ff8-81d7-18363415052a" path="/var/lib/kubelet/pods/9155b41a-6ea6-4ff8-81d7-18363415052a/volumes" Feb 02 13:58:59 crc kubenswrapper[4846]: I0202 13:58:59.423733 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:58:59 crc kubenswrapper[4846]: E0202 13:58:59.424802 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:59:14 crc kubenswrapper[4846]: I0202 13:59:14.423783 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:59:14 crc kubenswrapper[4846]: E0202 13:59:14.425434 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:59:25 crc kubenswrapper[4846]: I0202 13:59:25.423902 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:59:25 crc kubenswrapper[4846]: E0202 13:59:25.425383 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:59:35 crc kubenswrapper[4846]: I0202 13:59:35.054050 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-ml2mq"] Feb 02 13:59:35 crc kubenswrapper[4846]: I0202 13:59:35.062846 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-ml2mq"] Feb 02 13:59:35 crc kubenswrapper[4846]: I0202 13:59:35.444294 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27f64e70-e2a4-4d28-85fa-32421976a607" path="/var/lib/kubelet/pods/27f64e70-e2a4-4d28-85fa-32421976a607/volumes" Feb 02 13:59:38 crc kubenswrapper[4846]: I0202 13:59:38.423588 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:59:38 crc kubenswrapper[4846]: E0202 13:59:38.424251 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 13:59:42 crc kubenswrapper[4846]: I0202 13:59:42.705390 4846 scope.go:117] "RemoveContainer" containerID="f32c406473cbf1428acf134df8ba011325e01b9f9ce31974b90488b8d9e66468" Feb 02 13:59:42 crc kubenswrapper[4846]: I0202 13:59:42.750024 4846 scope.go:117] "RemoveContainer" containerID="01df9df9171d818929a91e1f45eba8202cd2797e9cbcd282919bd1d1c1bb521c" Feb 02 13:59:42 crc kubenswrapper[4846]: I0202 13:59:42.812527 4846 scope.go:117] "RemoveContainer" containerID="80eb979d842bcaa49dbe868238d121d7d74f6f1ecc1a5f2a2ca20f0ef0009967" Feb 02 13:59:42 crc kubenswrapper[4846]: I0202 13:59:42.869277 4846 scope.go:117] "RemoveContainer" containerID="b261b51c3bc340880b1917f93aa56e2877153da942e93fe1a361ae5c7dfc2ff5" Feb 02 13:59:42 crc kubenswrapper[4846]: I0202 13:59:42.914550 4846 scope.go:117] "RemoveContainer" containerID="5a64a2f0dbfcc27fcbf3658c317b5037567325030e77d5bb98568fae0d3b49b1" Feb 02 13:59:42 crc kubenswrapper[4846]: I0202 13:59:42.960742 4846 scope.go:117] "RemoveContainer" containerID="a2a9bcd1bf2742b52894d74d287ed03a580068e20aaf6d39d232ca19f9bfee85" Feb 02 13:59:51 crc kubenswrapper[4846]: I0202 13:59:51.424260 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 13:59:51 crc kubenswrapper[4846]: E0202 13:59:51.425107 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.167081 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz"] Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.172480 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.179057 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.179811 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.194869 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz"] Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.218894 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/82d62939-2d72-495d-a677-4dad13b4fab1-secret-volume\") pod \"collect-profiles-29500680-j9xzz\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.219121 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcgvq\" (UniqueName: \"kubernetes.io/projected/82d62939-2d72-495d-a677-4dad13b4fab1-kube-api-access-pcgvq\") pod \"collect-profiles-29500680-j9xzz\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.219199 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82d62939-2d72-495d-a677-4dad13b4fab1-config-volume\") pod \"collect-profiles-29500680-j9xzz\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.321553 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcgvq\" (UniqueName: \"kubernetes.io/projected/82d62939-2d72-495d-a677-4dad13b4fab1-kube-api-access-pcgvq\") pod \"collect-profiles-29500680-j9xzz\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.321687 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82d62939-2d72-495d-a677-4dad13b4fab1-config-volume\") pod \"collect-profiles-29500680-j9xzz\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.321862 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/82d62939-2d72-495d-a677-4dad13b4fab1-secret-volume\") pod \"collect-profiles-29500680-j9xzz\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.322794 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82d62939-2d72-495d-a677-4dad13b4fab1-config-volume\") pod \"collect-profiles-29500680-j9xzz\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.329027 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/82d62939-2d72-495d-a677-4dad13b4fab1-secret-volume\") pod \"collect-profiles-29500680-j9xzz\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.348684 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcgvq\" (UniqueName: \"kubernetes.io/projected/82d62939-2d72-495d-a677-4dad13b4fab1-kube-api-access-pcgvq\") pod \"collect-profiles-29500680-j9xzz\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.500717 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:00 crc kubenswrapper[4846]: I0202 14:00:00.978542 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz"] Feb 02 14:00:00 crc kubenswrapper[4846]: W0202 14:00:00.984390 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82d62939_2d72_495d_a677_4dad13b4fab1.slice/crio-7a3ced271b9c1388405a99bd51c41af5d147da4c0d0269e8bdaf8bdec7abd976 WatchSource:0}: Error finding container 7a3ced271b9c1388405a99bd51c41af5d147da4c0d0269e8bdaf8bdec7abd976: Status 404 returned error can't find the container with id 7a3ced271b9c1388405a99bd51c41af5d147da4c0d0269e8bdaf8bdec7abd976 Feb 02 14:00:01 crc kubenswrapper[4846]: I0202 14:00:01.954127 4846 generic.go:334] "Generic (PLEG): container finished" podID="82d62939-2d72-495d-a677-4dad13b4fab1" containerID="4bd03e5ad50707aa0e71408b7e83edd214369278933e9d4dede1351c4ea61f87" exitCode=0 Feb 02 14:00:01 crc kubenswrapper[4846]: I0202 14:00:01.954296 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" event={"ID":"82d62939-2d72-495d-a677-4dad13b4fab1","Type":"ContainerDied","Data":"4bd03e5ad50707aa0e71408b7e83edd214369278933e9d4dede1351c4ea61f87"} Feb 02 14:00:01 crc kubenswrapper[4846]: I0202 14:00:01.954423 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" event={"ID":"82d62939-2d72-495d-a677-4dad13b4fab1","Type":"ContainerStarted","Data":"7a3ced271b9c1388405a99bd51c41af5d147da4c0d0269e8bdaf8bdec7abd976"} Feb 02 14:00:02 crc kubenswrapper[4846]: I0202 14:00:02.424045 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 14:00:02 crc kubenswrapper[4846]: E0202 14:00:02.424514 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.338982 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.490315 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/82d62939-2d72-495d-a677-4dad13b4fab1-secret-volume\") pod \"82d62939-2d72-495d-a677-4dad13b4fab1\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.490389 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82d62939-2d72-495d-a677-4dad13b4fab1-config-volume\") pod \"82d62939-2d72-495d-a677-4dad13b4fab1\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.490441 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcgvq\" (UniqueName: \"kubernetes.io/projected/82d62939-2d72-495d-a677-4dad13b4fab1-kube-api-access-pcgvq\") pod \"82d62939-2d72-495d-a677-4dad13b4fab1\" (UID: \"82d62939-2d72-495d-a677-4dad13b4fab1\") " Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.491094 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82d62939-2d72-495d-a677-4dad13b4fab1-config-volume" (OuterVolumeSpecName: "config-volume") pod "82d62939-2d72-495d-a677-4dad13b4fab1" (UID: "82d62939-2d72-495d-a677-4dad13b4fab1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.492254 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82d62939-2d72-495d-a677-4dad13b4fab1-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.497225 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82d62939-2d72-495d-a677-4dad13b4fab1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "82d62939-2d72-495d-a677-4dad13b4fab1" (UID: "82d62939-2d72-495d-a677-4dad13b4fab1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.499587 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82d62939-2d72-495d-a677-4dad13b4fab1-kube-api-access-pcgvq" (OuterVolumeSpecName: "kube-api-access-pcgvq") pod "82d62939-2d72-495d-a677-4dad13b4fab1" (UID: "82d62939-2d72-495d-a677-4dad13b4fab1"). InnerVolumeSpecName "kube-api-access-pcgvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.596486 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/82d62939-2d72-495d-a677-4dad13b4fab1-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.596529 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcgvq\" (UniqueName: \"kubernetes.io/projected/82d62939-2d72-495d-a677-4dad13b4fab1-kube-api-access-pcgvq\") on node \"crc\" DevicePath \"\"" Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.997113 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.997117 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz" event={"ID":"82d62939-2d72-495d-a677-4dad13b4fab1","Type":"ContainerDied","Data":"7a3ced271b9c1388405a99bd51c41af5d147da4c0d0269e8bdaf8bdec7abd976"} Feb 02 14:00:03 crc kubenswrapper[4846]: I0202 14:00:03.997234 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a3ced271b9c1388405a99bd51c41af5d147da4c0d0269e8bdaf8bdec7abd976" Feb 02 14:00:04 crc kubenswrapper[4846]: I0202 14:00:04.424286 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm"] Feb 02 14:00:04 crc kubenswrapper[4846]: I0202 14:00:04.435537 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500635-89vhm"] Feb 02 14:00:05 crc kubenswrapper[4846]: I0202 14:00:05.435908 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b" path="/var/lib/kubelet/pods/cf2ecf4f-2f9c-48bd-a024-db8e0c1beb0b/volumes" Feb 02 14:00:14 crc kubenswrapper[4846]: I0202 14:00:14.423468 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 14:00:14 crc kubenswrapper[4846]: E0202 14:00:14.424269 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:00:29 crc kubenswrapper[4846]: I0202 14:00:29.431251 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 14:00:29 crc kubenswrapper[4846]: E0202 14:00:29.432106 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:00:40 crc kubenswrapper[4846]: I0202 14:00:40.423666 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 14:00:41 crc kubenswrapper[4846]: I0202 14:00:41.376212 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"b0feaef9b428cd0c86a2e50ab4652a789c6551e6f5340ab71f273dd3ca08238a"} Feb 02 14:00:43 crc kubenswrapper[4846]: I0202 14:00:43.120539 4846 scope.go:117] "RemoveContainer" containerID="ddc740203fd5cf82960549078d5f6395a8cb9e8a6df3d7e01a826068e292b53f" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.165050 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29500681-ln5z2"] Feb 02 14:01:00 crc kubenswrapper[4846]: E0202 14:01:00.166230 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82d62939-2d72-495d-a677-4dad13b4fab1" containerName="collect-profiles" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.166248 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="82d62939-2d72-495d-a677-4dad13b4fab1" containerName="collect-profiles" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.166497 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="82d62939-2d72-495d-a677-4dad13b4fab1" containerName="collect-profiles" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.167600 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.177741 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29500681-ln5z2"] Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.304140 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-combined-ca-bundle\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.304199 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-config-data\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.304218 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-fernet-keys\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.304267 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtgls\" (UniqueName: \"kubernetes.io/projected/32950bad-c06b-4cd1-bae5-1fd20646ce59-kube-api-access-mtgls\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.406916 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-combined-ca-bundle\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.406977 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-config-data\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.407007 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-fernet-keys\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.407034 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtgls\" (UniqueName: \"kubernetes.io/projected/32950bad-c06b-4cd1-bae5-1fd20646ce59-kube-api-access-mtgls\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.415606 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-config-data\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.415757 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-fernet-keys\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.416660 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-combined-ca-bundle\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.425959 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtgls\" (UniqueName: \"kubernetes.io/projected/32950bad-c06b-4cd1-bae5-1fd20646ce59-kube-api-access-mtgls\") pod \"keystone-cron-29500681-ln5z2\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:00 crc kubenswrapper[4846]: I0202 14:01:00.508426 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:01 crc kubenswrapper[4846]: I0202 14:01:01.007291 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29500681-ln5z2"] Feb 02 14:01:01 crc kubenswrapper[4846]: I0202 14:01:01.624104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29500681-ln5z2" event={"ID":"32950bad-c06b-4cd1-bae5-1fd20646ce59","Type":"ContainerStarted","Data":"9825a65af8b8e903761a6c872bda3d25069293fbd3b8fec78e549ba0081b0534"} Feb 02 14:01:01 crc kubenswrapper[4846]: I0202 14:01:01.624446 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29500681-ln5z2" event={"ID":"32950bad-c06b-4cd1-bae5-1fd20646ce59","Type":"ContainerStarted","Data":"689fd7c309a680269e76a314cca0ec3226141a396d5eb463de9aefb7fb7d02bc"} Feb 02 14:01:01 crc kubenswrapper[4846]: I0202 14:01:01.642008 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29500681-ln5z2" podStartSLOduration=1.641988877 podStartE2EDuration="1.641988877s" podCreationTimestamp="2026-02-02 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 14:01:01.636735709 +0000 UTC m=+6692.865322572" watchObservedRunningTime="2026-02-02 14:01:01.641988877 +0000 UTC m=+6692.870575740" Feb 02 14:01:04 crc kubenswrapper[4846]: I0202 14:01:04.649292 4846 generic.go:334] "Generic (PLEG): container finished" podID="32950bad-c06b-4cd1-bae5-1fd20646ce59" containerID="9825a65af8b8e903761a6c872bda3d25069293fbd3b8fec78e549ba0081b0534" exitCode=0 Feb 02 14:01:04 crc kubenswrapper[4846]: I0202 14:01:04.649377 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29500681-ln5z2" event={"ID":"32950bad-c06b-4cd1-bae5-1fd20646ce59","Type":"ContainerDied","Data":"9825a65af8b8e903761a6c872bda3d25069293fbd3b8fec78e549ba0081b0534"} Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.068316 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.256042 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-combined-ca-bundle\") pod \"32950bad-c06b-4cd1-bae5-1fd20646ce59\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.256302 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtgls\" (UniqueName: \"kubernetes.io/projected/32950bad-c06b-4cd1-bae5-1fd20646ce59-kube-api-access-mtgls\") pod \"32950bad-c06b-4cd1-bae5-1fd20646ce59\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.256523 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-fernet-keys\") pod \"32950bad-c06b-4cd1-bae5-1fd20646ce59\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.256595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-config-data\") pod \"32950bad-c06b-4cd1-bae5-1fd20646ce59\" (UID: \"32950bad-c06b-4cd1-bae5-1fd20646ce59\") " Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.261845 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "32950bad-c06b-4cd1-bae5-1fd20646ce59" (UID: "32950bad-c06b-4cd1-bae5-1fd20646ce59"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.261857 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32950bad-c06b-4cd1-bae5-1fd20646ce59-kube-api-access-mtgls" (OuterVolumeSpecName: "kube-api-access-mtgls") pod "32950bad-c06b-4cd1-bae5-1fd20646ce59" (UID: "32950bad-c06b-4cd1-bae5-1fd20646ce59"). InnerVolumeSpecName "kube-api-access-mtgls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.286155 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32950bad-c06b-4cd1-bae5-1fd20646ce59" (UID: "32950bad-c06b-4cd1-bae5-1fd20646ce59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.317530 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-config-data" (OuterVolumeSpecName: "config-data") pod "32950bad-c06b-4cd1-bae5-1fd20646ce59" (UID: "32950bad-c06b-4cd1-bae5-1fd20646ce59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.359279 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.359332 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.359342 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32950bad-c06b-4cd1-bae5-1fd20646ce59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.359352 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtgls\" (UniqueName: \"kubernetes.io/projected/32950bad-c06b-4cd1-bae5-1fd20646ce59-kube-api-access-mtgls\") on node \"crc\" DevicePath \"\"" Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.669716 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29500681-ln5z2" event={"ID":"32950bad-c06b-4cd1-bae5-1fd20646ce59","Type":"ContainerDied","Data":"689fd7c309a680269e76a314cca0ec3226141a396d5eb463de9aefb7fb7d02bc"} Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.669759 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="689fd7c309a680269e76a314cca0ec3226141a396d5eb463de9aefb7fb7d02bc" Feb 02 14:01:06 crc kubenswrapper[4846]: I0202 14:01:06.669798 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29500681-ln5z2" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.206179 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-skwkr"] Feb 02 14:01:20 crc kubenswrapper[4846]: E0202 14:01:20.207282 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32950bad-c06b-4cd1-bae5-1fd20646ce59" containerName="keystone-cron" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.207300 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="32950bad-c06b-4cd1-bae5-1fd20646ce59" containerName="keystone-cron" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.207534 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="32950bad-c06b-4cd1-bae5-1fd20646ce59" containerName="keystone-cron" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.209341 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.218515 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-skwkr"] Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.308111 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-catalog-content\") pod \"redhat-marketplace-skwkr\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.308443 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-utilities\") pod \"redhat-marketplace-skwkr\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.308526 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd6cg\" (UniqueName: \"kubernetes.io/projected/bcd69d08-9350-47b3-9df6-632ab5a9116f-kube-api-access-xd6cg\") pod \"redhat-marketplace-skwkr\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.410586 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-utilities\") pod \"redhat-marketplace-skwkr\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.410718 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd6cg\" (UniqueName: \"kubernetes.io/projected/bcd69d08-9350-47b3-9df6-632ab5a9116f-kube-api-access-xd6cg\") pod \"redhat-marketplace-skwkr\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.410853 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-catalog-content\") pod \"redhat-marketplace-skwkr\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.411050 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-utilities\") pod \"redhat-marketplace-skwkr\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.411422 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-catalog-content\") pod \"redhat-marketplace-skwkr\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.432092 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd6cg\" (UniqueName: \"kubernetes.io/projected/bcd69d08-9350-47b3-9df6-632ab5a9116f-kube-api-access-xd6cg\") pod \"redhat-marketplace-skwkr\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:20 crc kubenswrapper[4846]: I0202 14:01:20.529999 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:21 crc kubenswrapper[4846]: I0202 14:01:21.137387 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-skwkr"] Feb 02 14:01:21 crc kubenswrapper[4846]: I0202 14:01:21.815468 4846 generic.go:334] "Generic (PLEG): container finished" podID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerID="3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5" exitCode=0 Feb 02 14:01:21 crc kubenswrapper[4846]: I0202 14:01:21.815550 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skwkr" event={"ID":"bcd69d08-9350-47b3-9df6-632ab5a9116f","Type":"ContainerDied","Data":"3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5"} Feb 02 14:01:21 crc kubenswrapper[4846]: I0202 14:01:21.815914 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skwkr" event={"ID":"bcd69d08-9350-47b3-9df6-632ab5a9116f","Type":"ContainerStarted","Data":"211681d80bb719fa804708761582aaad056c71c8e82876fe0d597cefc1c7b1b5"} Feb 02 14:01:21 crc kubenswrapper[4846]: I0202 14:01:21.818811 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 14:01:23 crc kubenswrapper[4846]: I0202 14:01:23.839328 4846 generic.go:334] "Generic (PLEG): container finished" podID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerID="ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba" exitCode=0 Feb 02 14:01:23 crc kubenswrapper[4846]: I0202 14:01:23.839478 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skwkr" event={"ID":"bcd69d08-9350-47b3-9df6-632ab5a9116f","Type":"ContainerDied","Data":"ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba"} Feb 02 14:01:24 crc kubenswrapper[4846]: I0202 14:01:24.851470 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skwkr" event={"ID":"bcd69d08-9350-47b3-9df6-632ab5a9116f","Type":"ContainerStarted","Data":"936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa"} Feb 02 14:01:24 crc kubenswrapper[4846]: I0202 14:01:24.882755 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-skwkr" podStartSLOduration=2.410692381 podStartE2EDuration="4.882733146s" podCreationTimestamp="2026-02-02 14:01:20 +0000 UTC" firstStartedPulling="2026-02-02 14:01:21.818452828 +0000 UTC m=+6713.047039691" lastFinishedPulling="2026-02-02 14:01:24.290493593 +0000 UTC m=+6715.519080456" observedRunningTime="2026-02-02 14:01:24.872097457 +0000 UTC m=+6716.100684330" watchObservedRunningTime="2026-02-02 14:01:24.882733146 +0000 UTC m=+6716.111320009" Feb 02 14:01:30 crc kubenswrapper[4846]: I0202 14:01:30.530983 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:30 crc kubenswrapper[4846]: I0202 14:01:30.531597 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:30 crc kubenswrapper[4846]: I0202 14:01:30.579872 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:30 crc kubenswrapper[4846]: I0202 14:01:30.968067 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:31 crc kubenswrapper[4846]: I0202 14:01:31.016968 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-skwkr"] Feb 02 14:01:32 crc kubenswrapper[4846]: I0202 14:01:32.940129 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-skwkr" podUID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerName="registry-server" containerID="cri-o://936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa" gracePeriod=2 Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.448233 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.606240 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-utilities\") pod \"bcd69d08-9350-47b3-9df6-632ab5a9116f\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.606397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd6cg\" (UniqueName: \"kubernetes.io/projected/bcd69d08-9350-47b3-9df6-632ab5a9116f-kube-api-access-xd6cg\") pod \"bcd69d08-9350-47b3-9df6-632ab5a9116f\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.606520 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-catalog-content\") pod \"bcd69d08-9350-47b3-9df6-632ab5a9116f\" (UID: \"bcd69d08-9350-47b3-9df6-632ab5a9116f\") " Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.607180 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-utilities" (OuterVolumeSpecName: "utilities") pod "bcd69d08-9350-47b3-9df6-632ab5a9116f" (UID: "bcd69d08-9350-47b3-9df6-632ab5a9116f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.613085 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcd69d08-9350-47b3-9df6-632ab5a9116f-kube-api-access-xd6cg" (OuterVolumeSpecName: "kube-api-access-xd6cg") pod "bcd69d08-9350-47b3-9df6-632ab5a9116f" (UID: "bcd69d08-9350-47b3-9df6-632ab5a9116f"). InnerVolumeSpecName "kube-api-access-xd6cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.629475 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bcd69d08-9350-47b3-9df6-632ab5a9116f" (UID: "bcd69d08-9350-47b3-9df6-632ab5a9116f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.709889 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.709938 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcd69d08-9350-47b3-9df6-632ab5a9116f-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.709952 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd6cg\" (UniqueName: \"kubernetes.io/projected/bcd69d08-9350-47b3-9df6-632ab5a9116f-kube-api-access-xd6cg\") on node \"crc\" DevicePath \"\"" Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.956083 4846 generic.go:334] "Generic (PLEG): container finished" podID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerID="936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa" exitCode=0 Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.956151 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skwkr" event={"ID":"bcd69d08-9350-47b3-9df6-632ab5a9116f","Type":"ContainerDied","Data":"936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa"} Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.956186 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skwkr" event={"ID":"bcd69d08-9350-47b3-9df6-632ab5a9116f","Type":"ContainerDied","Data":"211681d80bb719fa804708761582aaad056c71c8e82876fe0d597cefc1c7b1b5"} Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.956226 4846 scope.go:117] "RemoveContainer" containerID="936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa" Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.956510 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-skwkr" Feb 02 14:01:33 crc kubenswrapper[4846]: I0202 14:01:33.979833 4846 scope.go:117] "RemoveContainer" containerID="ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba" Feb 02 14:01:34 crc kubenswrapper[4846]: I0202 14:01:34.002267 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-skwkr"] Feb 02 14:01:34 crc kubenswrapper[4846]: I0202 14:01:34.014197 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-skwkr"] Feb 02 14:01:34 crc kubenswrapper[4846]: I0202 14:01:34.016395 4846 scope.go:117] "RemoveContainer" containerID="3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5" Feb 02 14:01:34 crc kubenswrapper[4846]: I0202 14:01:34.071414 4846 scope.go:117] "RemoveContainer" containerID="936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa" Feb 02 14:01:34 crc kubenswrapper[4846]: E0202 14:01:34.072302 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa\": container with ID starting with 936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa not found: ID does not exist" containerID="936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa" Feb 02 14:01:34 crc kubenswrapper[4846]: I0202 14:01:34.072339 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa"} err="failed to get container status \"936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa\": rpc error: code = NotFound desc = could not find container \"936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa\": container with ID starting with 936a4e680ac90fa2ddfdb300242ef755199aaedf0127985276264c382faa11fa not found: ID does not exist" Feb 02 14:01:34 crc kubenswrapper[4846]: I0202 14:01:34.072365 4846 scope.go:117] "RemoveContainer" containerID="ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba" Feb 02 14:01:34 crc kubenswrapper[4846]: E0202 14:01:34.072728 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba\": container with ID starting with ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba not found: ID does not exist" containerID="ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba" Feb 02 14:01:34 crc kubenswrapper[4846]: I0202 14:01:34.072769 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba"} err="failed to get container status \"ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba\": rpc error: code = NotFound desc = could not find container \"ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba\": container with ID starting with ada80826f3f32d6b14a5cbb1060e2ea1c3fab23d1bab55445a4338ebf9122cba not found: ID does not exist" Feb 02 14:01:34 crc kubenswrapper[4846]: I0202 14:01:34.072799 4846 scope.go:117] "RemoveContainer" containerID="3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5" Feb 02 14:01:34 crc kubenswrapper[4846]: E0202 14:01:34.073127 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5\": container with ID starting with 3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5 not found: ID does not exist" containerID="3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5" Feb 02 14:01:34 crc kubenswrapper[4846]: I0202 14:01:34.073155 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5"} err="failed to get container status \"3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5\": rpc error: code = NotFound desc = could not find container \"3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5\": container with ID starting with 3c69d990b59a46b86b696ad66a25df8a4c0ab0e69ac3996a506423df87b629e5 not found: ID does not exist" Feb 02 14:01:35 crc kubenswrapper[4846]: I0202 14:01:35.433871 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcd69d08-9350-47b3-9df6-632ab5a9116f" path="/var/lib/kubelet/pods/bcd69d08-9350-47b3-9df6-632ab5a9116f/volumes" Feb 02 14:03:00 crc kubenswrapper[4846]: I0202 14:03:00.479464 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:03:00 crc kubenswrapper[4846]: I0202 14:03:00.480021 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:03:09 crc kubenswrapper[4846]: I0202 14:03:09.043341 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-tms4z"] Feb 02 14:03:09 crc kubenswrapper[4846]: I0202 14:03:09.056364 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-fb24-account-create-update-6hv7c"] Feb 02 14:03:09 crc kubenswrapper[4846]: I0202 14:03:09.065157 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-tms4z"] Feb 02 14:03:09 crc kubenswrapper[4846]: I0202 14:03:09.074687 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-fb24-account-create-update-6hv7c"] Feb 02 14:03:09 crc kubenswrapper[4846]: I0202 14:03:09.442786 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ea28d1f-bc99-4cd3-9f37-922870704b52" path="/var/lib/kubelet/pods/5ea28d1f-bc99-4cd3-9f37-922870704b52/volumes" Feb 02 14:03:09 crc kubenswrapper[4846]: I0202 14:03:09.443828 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86ba6367-a1ef-4568-96e2-cddbcab98c68" path="/var/lib/kubelet/pods/86ba6367-a1ef-4568-96e2-cddbcab98c68/volumes" Feb 02 14:03:23 crc kubenswrapper[4846]: I0202 14:03:23.051933 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-g6986"] Feb 02 14:03:23 crc kubenswrapper[4846]: I0202 14:03:23.066573 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-g6986"] Feb 02 14:03:23 crc kubenswrapper[4846]: I0202 14:03:23.435928 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="337292e8-0928-4f18-86d6-e433f406af6b" path="/var/lib/kubelet/pods/337292e8-0928-4f18-86d6-e433f406af6b/volumes" Feb 02 14:03:30 crc kubenswrapper[4846]: I0202 14:03:30.478529 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:03:30 crc kubenswrapper[4846]: I0202 14:03:30.479132 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:03:43 crc kubenswrapper[4846]: I0202 14:03:43.245497 4846 scope.go:117] "RemoveContainer" containerID="a0a6c3dd28e50d762ec4e4bde884d757f3d597a1560d821b6ff6cb795d82dc0d" Feb 02 14:03:43 crc kubenswrapper[4846]: I0202 14:03:43.275935 4846 scope.go:117] "RemoveContainer" containerID="bdc95a26941ce1d40c00fc2320b558f64a8062946be22fdd0e03417633000c42" Feb 02 14:03:43 crc kubenswrapper[4846]: I0202 14:03:43.363607 4846 scope.go:117] "RemoveContainer" containerID="7650dafd98d9fd70ffe724a2db03c28b9c0d9052eb70c0e3042f1828171adc1a" Feb 02 14:04:00 crc kubenswrapper[4846]: I0202 14:04:00.479703 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:04:00 crc kubenswrapper[4846]: I0202 14:04:00.480335 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:04:00 crc kubenswrapper[4846]: I0202 14:04:00.480384 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 14:04:00 crc kubenswrapper[4846]: I0202 14:04:00.481265 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b0feaef9b428cd0c86a2e50ab4652a789c6551e6f5340ab71f273dd3ca08238a"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 14:04:00 crc kubenswrapper[4846]: I0202 14:04:00.481333 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://b0feaef9b428cd0c86a2e50ab4652a789c6551e6f5340ab71f273dd3ca08238a" gracePeriod=600 Feb 02 14:04:01 crc kubenswrapper[4846]: I0202 14:04:01.486742 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="b0feaef9b428cd0c86a2e50ab4652a789c6551e6f5340ab71f273dd3ca08238a" exitCode=0 Feb 02 14:04:01 crc kubenswrapper[4846]: I0202 14:04:01.486833 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"b0feaef9b428cd0c86a2e50ab4652a789c6551e6f5340ab71f273dd3ca08238a"} Feb 02 14:04:01 crc kubenswrapper[4846]: I0202 14:04:01.487663 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17"} Feb 02 14:04:01 crc kubenswrapper[4846]: I0202 14:04:01.487693 4846 scope.go:117] "RemoveContainer" containerID="9dcd41bf6282b8afa4df7e5903d2ab916fa8963089e6240eeb55e8fdf3fd20ab" Feb 02 14:05:46 crc kubenswrapper[4846]: I0202 14:05:46.061302 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-5b76-account-create-update-px22h"] Feb 02 14:05:46 crc kubenswrapper[4846]: I0202 14:05:46.072763 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-9knl4"] Feb 02 14:05:46 crc kubenswrapper[4846]: I0202 14:05:46.087694 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-5b76-account-create-update-px22h"] Feb 02 14:05:46 crc kubenswrapper[4846]: I0202 14:05:46.106008 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-9knl4"] Feb 02 14:05:47 crc kubenswrapper[4846]: I0202 14:05:47.437714 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07b19d98-9059-4596-8ffb-b0264d3a6f14" path="/var/lib/kubelet/pods/07b19d98-9059-4596-8ffb-b0264d3a6f14/volumes" Feb 02 14:05:47 crc kubenswrapper[4846]: I0202 14:05:47.440049 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6530545-8859-43e2-baee-8a1021830ba0" path="/var/lib/kubelet/pods/d6530545-8859-43e2-baee-8a1021830ba0/volumes" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.000012 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hjzb7"] Feb 02 14:05:55 crc kubenswrapper[4846]: E0202 14:05:55.004778 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerName="extract-content" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.004802 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerName="extract-content" Feb 02 14:05:55 crc kubenswrapper[4846]: E0202 14:05:55.004822 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerName="extract-utilities" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.004829 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerName="extract-utilities" Feb 02 14:05:55 crc kubenswrapper[4846]: E0202 14:05:55.004846 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerName="registry-server" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.004852 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerName="registry-server" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.005066 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcd69d08-9350-47b3-9df6-632ab5a9116f" containerName="registry-server" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.006552 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.100814 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hjzb7"] Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.125971 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdpfx\" (UniqueName: \"kubernetes.io/projected/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-kube-api-access-qdpfx\") pod \"community-operators-hjzb7\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.127313 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-catalog-content\") pod \"community-operators-hjzb7\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.127939 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-utilities\") pod \"community-operators-hjzb7\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.229701 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-utilities\") pod \"community-operators-hjzb7\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.229779 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdpfx\" (UniqueName: \"kubernetes.io/projected/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-kube-api-access-qdpfx\") pod \"community-operators-hjzb7\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.229798 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-catalog-content\") pod \"community-operators-hjzb7\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.230189 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-utilities\") pod \"community-operators-hjzb7\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.230300 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-catalog-content\") pod \"community-operators-hjzb7\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.254288 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdpfx\" (UniqueName: \"kubernetes.io/projected/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-kube-api-access-qdpfx\") pod \"community-operators-hjzb7\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.330709 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:05:55 crc kubenswrapper[4846]: I0202 14:05:55.891437 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hjzb7"] Feb 02 14:05:56 crc kubenswrapper[4846]: I0202 14:05:56.602024 4846 generic.go:334] "Generic (PLEG): container finished" podID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerID="a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac" exitCode=0 Feb 02 14:05:56 crc kubenswrapper[4846]: I0202 14:05:56.602083 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjzb7" event={"ID":"98c2a60e-d6de-48d6-9f0b-f40fd252ed42","Type":"ContainerDied","Data":"a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac"} Feb 02 14:05:56 crc kubenswrapper[4846]: I0202 14:05:56.602359 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjzb7" event={"ID":"98c2a60e-d6de-48d6-9f0b-f40fd252ed42","Type":"ContainerStarted","Data":"c48f55eb8302576653181898235fa1a0efbfd0647a89865e077e77f12e5960cf"} Feb 02 14:05:57 crc kubenswrapper[4846]: I0202 14:05:57.614829 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjzb7" event={"ID":"98c2a60e-d6de-48d6-9f0b-f40fd252ed42","Type":"ContainerStarted","Data":"3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116"} Feb 02 14:05:58 crc kubenswrapper[4846]: I0202 14:05:58.036410 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-9rrdv"] Feb 02 14:05:58 crc kubenswrapper[4846]: I0202 14:05:58.046696 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-9rrdv"] Feb 02 14:05:59 crc kubenswrapper[4846]: I0202 14:05:59.444894 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b" path="/var/lib/kubelet/pods/68edd6cc-a80e-45e1-b8bf-bad5f82d8c8b/volumes" Feb 02 14:05:59 crc kubenswrapper[4846]: I0202 14:05:59.637465 4846 generic.go:334] "Generic (PLEG): container finished" podID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerID="3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116" exitCode=0 Feb 02 14:05:59 crc kubenswrapper[4846]: I0202 14:05:59.637519 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjzb7" event={"ID":"98c2a60e-d6de-48d6-9f0b-f40fd252ed42","Type":"ContainerDied","Data":"3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116"} Feb 02 14:06:00 crc kubenswrapper[4846]: I0202 14:06:00.479145 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:06:00 crc kubenswrapper[4846]: I0202 14:06:00.479525 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:06:00 crc kubenswrapper[4846]: I0202 14:06:00.650508 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjzb7" event={"ID":"98c2a60e-d6de-48d6-9f0b-f40fd252ed42","Type":"ContainerStarted","Data":"62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411"} Feb 02 14:06:00 crc kubenswrapper[4846]: I0202 14:06:00.673480 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hjzb7" podStartSLOduration=3.018687735 podStartE2EDuration="6.673458402s" podCreationTimestamp="2026-02-02 14:05:54 +0000 UTC" firstStartedPulling="2026-02-02 14:05:56.604013045 +0000 UTC m=+6987.832599908" lastFinishedPulling="2026-02-02 14:06:00.258783712 +0000 UTC m=+6991.487370575" observedRunningTime="2026-02-02 14:06:00.670078881 +0000 UTC m=+6991.898665744" watchObservedRunningTime="2026-02-02 14:06:00.673458402 +0000 UTC m=+6991.902045265" Feb 02 14:06:05 crc kubenswrapper[4846]: I0202 14:06:05.331963 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:06:05 crc kubenswrapper[4846]: I0202 14:06:05.333202 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:06:05 crc kubenswrapper[4846]: I0202 14:06:05.382759 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:06:05 crc kubenswrapper[4846]: I0202 14:06:05.756326 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:06:05 crc kubenswrapper[4846]: I0202 14:06:05.804165 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hjzb7"] Feb 02 14:06:07 crc kubenswrapper[4846]: I0202 14:06:07.732586 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hjzb7" podUID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerName="registry-server" containerID="cri-o://62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411" gracePeriod=2 Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.258837 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.435358 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdpfx\" (UniqueName: \"kubernetes.io/projected/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-kube-api-access-qdpfx\") pod \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.435846 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-catalog-content\") pod \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.436057 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-utilities\") pod \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\" (UID: \"98c2a60e-d6de-48d6-9f0b-f40fd252ed42\") " Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.436605 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-utilities" (OuterVolumeSpecName: "utilities") pod "98c2a60e-d6de-48d6-9f0b-f40fd252ed42" (UID: "98c2a60e-d6de-48d6-9f0b-f40fd252ed42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.436990 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.442108 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-kube-api-access-qdpfx" (OuterVolumeSpecName: "kube-api-access-qdpfx") pod "98c2a60e-d6de-48d6-9f0b-f40fd252ed42" (UID: "98c2a60e-d6de-48d6-9f0b-f40fd252ed42"). InnerVolumeSpecName "kube-api-access-qdpfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.488132 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98c2a60e-d6de-48d6-9f0b-f40fd252ed42" (UID: "98c2a60e-d6de-48d6-9f0b-f40fd252ed42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.539419 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdpfx\" (UniqueName: \"kubernetes.io/projected/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-kube-api-access-qdpfx\") on node \"crc\" DevicePath \"\"" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.539458 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c2a60e-d6de-48d6-9f0b-f40fd252ed42-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.743290 4846 generic.go:334] "Generic (PLEG): container finished" podID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerID="62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411" exitCode=0 Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.743339 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjzb7" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.743357 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjzb7" event={"ID":"98c2a60e-d6de-48d6-9f0b-f40fd252ed42","Type":"ContainerDied","Data":"62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411"} Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.743746 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjzb7" event={"ID":"98c2a60e-d6de-48d6-9f0b-f40fd252ed42","Type":"ContainerDied","Data":"c48f55eb8302576653181898235fa1a0efbfd0647a89865e077e77f12e5960cf"} Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.743767 4846 scope.go:117] "RemoveContainer" containerID="62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.773219 4846 scope.go:117] "RemoveContainer" containerID="3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.777835 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hjzb7"] Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.788577 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hjzb7"] Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.794011 4846 scope.go:117] "RemoveContainer" containerID="a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.839971 4846 scope.go:117] "RemoveContainer" containerID="62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411" Feb 02 14:06:08 crc kubenswrapper[4846]: E0202 14:06:08.840430 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411\": container with ID starting with 62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411 not found: ID does not exist" containerID="62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.840480 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411"} err="failed to get container status \"62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411\": rpc error: code = NotFound desc = could not find container \"62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411\": container with ID starting with 62c8ea839a2eaca3ff9f568135da4a66abb6351e8eecf2b28f44d281f71ce411 not found: ID does not exist" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.840512 4846 scope.go:117] "RemoveContainer" containerID="3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116" Feb 02 14:06:08 crc kubenswrapper[4846]: E0202 14:06:08.841018 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116\": container with ID starting with 3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116 not found: ID does not exist" containerID="3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.841052 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116"} err="failed to get container status \"3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116\": rpc error: code = NotFound desc = could not find container \"3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116\": container with ID starting with 3eef86711aef8723881f8b88cc72718cd487f82581ed8cf40712970487b42116 not found: ID does not exist" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.841075 4846 scope.go:117] "RemoveContainer" containerID="a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac" Feb 02 14:06:08 crc kubenswrapper[4846]: E0202 14:06:08.841394 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac\": container with ID starting with a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac not found: ID does not exist" containerID="a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac" Feb 02 14:06:08 crc kubenswrapper[4846]: I0202 14:06:08.841419 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac"} err="failed to get container status \"a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac\": rpc error: code = NotFound desc = could not find container \"a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac\": container with ID starting with a682216092870d9ae0cfa53ee4af543fc2aebff31e280a78ccd4414368ff4fac not found: ID does not exist" Feb 02 14:06:09 crc kubenswrapper[4846]: I0202 14:06:09.435261 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" path="/var/lib/kubelet/pods/98c2a60e-d6de-48d6-9f0b-f40fd252ed42/volumes" Feb 02 14:06:30 crc kubenswrapper[4846]: I0202 14:06:30.478639 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:06:30 crc kubenswrapper[4846]: I0202 14:06:30.479191 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:06:43 crc kubenswrapper[4846]: I0202 14:06:43.512971 4846 scope.go:117] "RemoveContainer" containerID="7b9245f9504a28fe84a9b2e9505168d9ea88623785a553ee63db7b001d883ce5" Feb 02 14:06:43 crc kubenswrapper[4846]: I0202 14:06:43.542523 4846 scope.go:117] "RemoveContainer" containerID="c3af41b11d7360f8f6f19e49f6e6f6db0b201758905d8db6835e47be2b26a19e" Feb 02 14:06:43 crc kubenswrapper[4846]: I0202 14:06:43.593227 4846 scope.go:117] "RemoveContainer" containerID="3904af06e1634518b3a299fd9e9e0b9b00f923de5c964d7b162a72f1c5f4dfc7" Feb 02 14:07:00 crc kubenswrapper[4846]: I0202 14:07:00.479008 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:07:00 crc kubenswrapper[4846]: I0202 14:07:00.479544 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:07:00 crc kubenswrapper[4846]: I0202 14:07:00.479594 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 14:07:00 crc kubenswrapper[4846]: I0202 14:07:00.480443 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 14:07:00 crc kubenswrapper[4846]: I0202 14:07:00.480493 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" gracePeriod=600 Feb 02 14:07:00 crc kubenswrapper[4846]: E0202 14:07:00.607053 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:07:01 crc kubenswrapper[4846]: I0202 14:07:01.244931 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" exitCode=0 Feb 02 14:07:01 crc kubenswrapper[4846]: I0202 14:07:01.244991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17"} Feb 02 14:07:01 crc kubenswrapper[4846]: I0202 14:07:01.245275 4846 scope.go:117] "RemoveContainer" containerID="b0feaef9b428cd0c86a2e50ab4652a789c6551e6f5340ab71f273dd3ca08238a" Feb 02 14:07:01 crc kubenswrapper[4846]: I0202 14:07:01.246020 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:07:01 crc kubenswrapper[4846]: E0202 14:07:01.246298 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:07:14 crc kubenswrapper[4846]: I0202 14:07:14.424210 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:07:14 crc kubenswrapper[4846]: E0202 14:07:14.425028 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:07:26 crc kubenswrapper[4846]: I0202 14:07:26.423931 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:07:26 crc kubenswrapper[4846]: E0202 14:07:26.425130 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.386930 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mxp8p"] Feb 02 14:07:39 crc kubenswrapper[4846]: E0202 14:07:39.387890 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerName="registry-server" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.387908 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerName="registry-server" Feb 02 14:07:39 crc kubenswrapper[4846]: E0202 14:07:39.387942 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerName="extract-utilities" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.387950 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerName="extract-utilities" Feb 02 14:07:39 crc kubenswrapper[4846]: E0202 14:07:39.387977 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerName="extract-content" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.387984 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerName="extract-content" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.388244 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c2a60e-d6de-48d6-9f0b-f40fd252ed42" containerName="registry-server" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.390090 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.426557 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:07:39 crc kubenswrapper[4846]: E0202 14:07:39.426918 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.476366 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mxp8p"] Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.510328 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjtnl\" (UniqueName: \"kubernetes.io/projected/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-kube-api-access-xjtnl\") pod \"certified-operators-mxp8p\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.510384 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-catalog-content\") pod \"certified-operators-mxp8p\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.510530 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-utilities\") pod \"certified-operators-mxp8p\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.612812 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjtnl\" (UniqueName: \"kubernetes.io/projected/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-kube-api-access-xjtnl\") pod \"certified-operators-mxp8p\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.612911 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-catalog-content\") pod \"certified-operators-mxp8p\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.613112 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-utilities\") pod \"certified-operators-mxp8p\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.613536 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-catalog-content\") pod \"certified-operators-mxp8p\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.614650 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-utilities\") pod \"certified-operators-mxp8p\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.640553 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjtnl\" (UniqueName: \"kubernetes.io/projected/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-kube-api-access-xjtnl\") pod \"certified-operators-mxp8p\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:39 crc kubenswrapper[4846]: I0202 14:07:39.728607 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:40 crc kubenswrapper[4846]: I0202 14:07:40.295525 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mxp8p"] Feb 02 14:07:40 crc kubenswrapper[4846]: I0202 14:07:40.632483 4846 generic.go:334] "Generic (PLEG): container finished" podID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerID="1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d" exitCode=0 Feb 02 14:07:40 crc kubenswrapper[4846]: I0202 14:07:40.632589 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxp8p" event={"ID":"f1e2e1a9-487b-45aa-abad-dfac6028b1e9","Type":"ContainerDied","Data":"1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d"} Feb 02 14:07:40 crc kubenswrapper[4846]: I0202 14:07:40.632952 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxp8p" event={"ID":"f1e2e1a9-487b-45aa-abad-dfac6028b1e9","Type":"ContainerStarted","Data":"6b74e5486e222e98d6fa3f268928d4531cb2b23aa3eea5a033d59328d661adb9"} Feb 02 14:07:40 crc kubenswrapper[4846]: I0202 14:07:40.636116 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 14:07:41 crc kubenswrapper[4846]: I0202 14:07:41.646332 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxp8p" event={"ID":"f1e2e1a9-487b-45aa-abad-dfac6028b1e9","Type":"ContainerStarted","Data":"1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef"} Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.368210 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k4l5h"] Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.370839 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.381469 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k4l5h"] Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.475776 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-utilities\") pod \"redhat-operators-k4l5h\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.475857 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk2hb\" (UniqueName: \"kubernetes.io/projected/d100bd2b-ca02-4dd2-abec-394722bee8a1-kube-api-access-sk2hb\") pod \"redhat-operators-k4l5h\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.475954 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-catalog-content\") pod \"redhat-operators-k4l5h\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.578273 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk2hb\" (UniqueName: \"kubernetes.io/projected/d100bd2b-ca02-4dd2-abec-394722bee8a1-kube-api-access-sk2hb\") pod \"redhat-operators-k4l5h\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.578471 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-catalog-content\") pod \"redhat-operators-k4l5h\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.578792 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-utilities\") pod \"redhat-operators-k4l5h\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.579344 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-utilities\") pod \"redhat-operators-k4l5h\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.579877 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-catalog-content\") pod \"redhat-operators-k4l5h\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.604827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk2hb\" (UniqueName: \"kubernetes.io/projected/d100bd2b-ca02-4dd2-abec-394722bee8a1-kube-api-access-sk2hb\") pod \"redhat-operators-k4l5h\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:42 crc kubenswrapper[4846]: I0202 14:07:42.693338 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:43 crc kubenswrapper[4846]: I0202 14:07:43.207421 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k4l5h"] Feb 02 14:07:43 crc kubenswrapper[4846]: I0202 14:07:43.666823 4846 generic.go:334] "Generic (PLEG): container finished" podID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerID="1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef" exitCode=0 Feb 02 14:07:43 crc kubenswrapper[4846]: I0202 14:07:43.667009 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxp8p" event={"ID":"f1e2e1a9-487b-45aa-abad-dfac6028b1e9","Type":"ContainerDied","Data":"1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef"} Feb 02 14:07:43 crc kubenswrapper[4846]: I0202 14:07:43.669351 4846 generic.go:334] "Generic (PLEG): container finished" podID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerID="4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209" exitCode=0 Feb 02 14:07:43 crc kubenswrapper[4846]: I0202 14:07:43.669390 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4l5h" event={"ID":"d100bd2b-ca02-4dd2-abec-394722bee8a1","Type":"ContainerDied","Data":"4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209"} Feb 02 14:07:43 crc kubenswrapper[4846]: I0202 14:07:43.669415 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4l5h" event={"ID":"d100bd2b-ca02-4dd2-abec-394722bee8a1","Type":"ContainerStarted","Data":"454524e03a9da411db92ec7ece771b6183162efdf6f082704dfb1a3bc025997e"} Feb 02 14:07:44 crc kubenswrapper[4846]: I0202 14:07:44.679502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4l5h" event={"ID":"d100bd2b-ca02-4dd2-abec-394722bee8a1","Type":"ContainerStarted","Data":"a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b"} Feb 02 14:07:44 crc kubenswrapper[4846]: I0202 14:07:44.684777 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxp8p" event={"ID":"f1e2e1a9-487b-45aa-abad-dfac6028b1e9","Type":"ContainerStarted","Data":"54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68"} Feb 02 14:07:44 crc kubenswrapper[4846]: I0202 14:07:44.725394 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mxp8p" podStartSLOduration=2.304543042 podStartE2EDuration="5.725376197s" podCreationTimestamp="2026-02-02 14:07:39 +0000 UTC" firstStartedPulling="2026-02-02 14:07:40.635904573 +0000 UTC m=+7091.864491436" lastFinishedPulling="2026-02-02 14:07:44.056737728 +0000 UTC m=+7095.285324591" observedRunningTime="2026-02-02 14:07:44.718544151 +0000 UTC m=+7095.947131064" watchObservedRunningTime="2026-02-02 14:07:44.725376197 +0000 UTC m=+7095.953963060" Feb 02 14:07:49 crc kubenswrapper[4846]: I0202 14:07:49.729735 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:49 crc kubenswrapper[4846]: I0202 14:07:49.730353 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:49 crc kubenswrapper[4846]: I0202 14:07:49.781472 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:50 crc kubenswrapper[4846]: I0202 14:07:50.736561 4846 generic.go:334] "Generic (PLEG): container finished" podID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerID="a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b" exitCode=0 Feb 02 14:07:50 crc kubenswrapper[4846]: I0202 14:07:50.736650 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4l5h" event={"ID":"d100bd2b-ca02-4dd2-abec-394722bee8a1","Type":"ContainerDied","Data":"a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b"} Feb 02 14:07:50 crc kubenswrapper[4846]: I0202 14:07:50.789407 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:51 crc kubenswrapper[4846]: I0202 14:07:51.557741 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mxp8p"] Feb 02 14:07:51 crc kubenswrapper[4846]: I0202 14:07:51.747388 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4l5h" event={"ID":"d100bd2b-ca02-4dd2-abec-394722bee8a1","Type":"ContainerStarted","Data":"bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc"} Feb 02 14:07:51 crc kubenswrapper[4846]: I0202 14:07:51.764177 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k4l5h" podStartSLOduration=2.212862131 podStartE2EDuration="9.764156666s" podCreationTimestamp="2026-02-02 14:07:42 +0000 UTC" firstStartedPulling="2026-02-02 14:07:43.671193596 +0000 UTC m=+7094.899780469" lastFinishedPulling="2026-02-02 14:07:51.222488141 +0000 UTC m=+7102.451075004" observedRunningTime="2026-02-02 14:07:51.762165888 +0000 UTC m=+7102.990752751" watchObservedRunningTime="2026-02-02 14:07:51.764156666 +0000 UTC m=+7102.992743529" Feb 02 14:07:52 crc kubenswrapper[4846]: I0202 14:07:52.693496 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:52 crc kubenswrapper[4846]: I0202 14:07:52.693873 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:07:52 crc kubenswrapper[4846]: I0202 14:07:52.755430 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mxp8p" podUID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerName="registry-server" containerID="cri-o://54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68" gracePeriod=2 Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.262285 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.358678 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-catalog-content\") pod \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.358751 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjtnl\" (UniqueName: \"kubernetes.io/projected/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-kube-api-access-xjtnl\") pod \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.359543 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-utilities" (OuterVolumeSpecName: "utilities") pod "f1e2e1a9-487b-45aa-abad-dfac6028b1e9" (UID: "f1e2e1a9-487b-45aa-abad-dfac6028b1e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.359646 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-utilities\") pod \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\" (UID: \"f1e2e1a9-487b-45aa-abad-dfac6028b1e9\") " Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.360118 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.368325 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-kube-api-access-xjtnl" (OuterVolumeSpecName: "kube-api-access-xjtnl") pod "f1e2e1a9-487b-45aa-abad-dfac6028b1e9" (UID: "f1e2e1a9-487b-45aa-abad-dfac6028b1e9"). InnerVolumeSpecName "kube-api-access-xjtnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.403955 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1e2e1a9-487b-45aa-abad-dfac6028b1e9" (UID: "f1e2e1a9-487b-45aa-abad-dfac6028b1e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.424155 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:07:53 crc kubenswrapper[4846]: E0202 14:07:53.424556 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.462264 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.462302 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjtnl\" (UniqueName: \"kubernetes.io/projected/f1e2e1a9-487b-45aa-abad-dfac6028b1e9-kube-api-access-xjtnl\") on node \"crc\" DevicePath \"\"" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.743546 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k4l5h" podUID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerName="registry-server" probeResult="failure" output=< Feb 02 14:07:53 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:07:53 crc kubenswrapper[4846]: > Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.766721 4846 generic.go:334] "Generic (PLEG): container finished" podID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerID="54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68" exitCode=0 Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.766798 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxp8p" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.766795 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxp8p" event={"ID":"f1e2e1a9-487b-45aa-abad-dfac6028b1e9","Type":"ContainerDied","Data":"54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68"} Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.766981 4846 scope.go:117] "RemoveContainer" containerID="54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.768770 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxp8p" event={"ID":"f1e2e1a9-487b-45aa-abad-dfac6028b1e9","Type":"ContainerDied","Data":"6b74e5486e222e98d6fa3f268928d4531cb2b23aa3eea5a033d59328d661adb9"} Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.808531 4846 scope.go:117] "RemoveContainer" containerID="1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.811308 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mxp8p"] Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.825726 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mxp8p"] Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.837500 4846 scope.go:117] "RemoveContainer" containerID="1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.883055 4846 scope.go:117] "RemoveContainer" containerID="54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68" Feb 02 14:07:53 crc kubenswrapper[4846]: E0202 14:07:53.883664 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68\": container with ID starting with 54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68 not found: ID does not exist" containerID="54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.883702 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68"} err="failed to get container status \"54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68\": rpc error: code = NotFound desc = could not find container \"54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68\": container with ID starting with 54d6f7f2e6dbccfb6ccd0e3a2d854a7b5e4f0a9bc3a104da348d8a5c3a900a68 not found: ID does not exist" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.883728 4846 scope.go:117] "RemoveContainer" containerID="1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef" Feb 02 14:07:53 crc kubenswrapper[4846]: E0202 14:07:53.884246 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef\": container with ID starting with 1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef not found: ID does not exist" containerID="1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.884345 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef"} err="failed to get container status \"1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef\": rpc error: code = NotFound desc = could not find container \"1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef\": container with ID starting with 1666b55756e4b5db437af2218b3f4d4f5ec6c5159b981e2476b41a371a4361ef not found: ID does not exist" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.884439 4846 scope.go:117] "RemoveContainer" containerID="1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d" Feb 02 14:07:53 crc kubenswrapper[4846]: E0202 14:07:53.884870 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d\": container with ID starting with 1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d not found: ID does not exist" containerID="1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d" Feb 02 14:07:53 crc kubenswrapper[4846]: I0202 14:07:53.884897 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d"} err="failed to get container status \"1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d\": rpc error: code = NotFound desc = could not find container \"1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d\": container with ID starting with 1ea7f0ccb228a93bf382e93307454b4f8831d2a3a9979254c7d7c6659497436d not found: ID does not exist" Feb 02 14:07:55 crc kubenswrapper[4846]: I0202 14:07:55.438940 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" path="/var/lib/kubelet/pods/f1e2e1a9-487b-45aa-abad-dfac6028b1e9/volumes" Feb 02 14:08:02 crc kubenswrapper[4846]: I0202 14:08:02.746631 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:08:02 crc kubenswrapper[4846]: I0202 14:08:02.801588 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:08:02 crc kubenswrapper[4846]: I0202 14:08:02.992418 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k4l5h"] Feb 02 14:08:03 crc kubenswrapper[4846]: I0202 14:08:03.868819 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k4l5h" podUID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerName="registry-server" containerID="cri-o://bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc" gracePeriod=2 Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.382879 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.513742 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-utilities\") pod \"d100bd2b-ca02-4dd2-abec-394722bee8a1\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.514574 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-utilities" (OuterVolumeSpecName: "utilities") pod "d100bd2b-ca02-4dd2-abec-394722bee8a1" (UID: "d100bd2b-ca02-4dd2-abec-394722bee8a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.514653 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-catalog-content\") pod \"d100bd2b-ca02-4dd2-abec-394722bee8a1\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.514728 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk2hb\" (UniqueName: \"kubernetes.io/projected/d100bd2b-ca02-4dd2-abec-394722bee8a1-kube-api-access-sk2hb\") pod \"d100bd2b-ca02-4dd2-abec-394722bee8a1\" (UID: \"d100bd2b-ca02-4dd2-abec-394722bee8a1\") " Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.515215 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.520187 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d100bd2b-ca02-4dd2-abec-394722bee8a1-kube-api-access-sk2hb" (OuterVolumeSpecName: "kube-api-access-sk2hb") pod "d100bd2b-ca02-4dd2-abec-394722bee8a1" (UID: "d100bd2b-ca02-4dd2-abec-394722bee8a1"). InnerVolumeSpecName "kube-api-access-sk2hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.617270 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk2hb\" (UniqueName: \"kubernetes.io/projected/d100bd2b-ca02-4dd2-abec-394722bee8a1-kube-api-access-sk2hb\") on node \"crc\" DevicePath \"\"" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.648737 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d100bd2b-ca02-4dd2-abec-394722bee8a1" (UID: "d100bd2b-ca02-4dd2-abec-394722bee8a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.720071 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d100bd2b-ca02-4dd2-abec-394722bee8a1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.882001 4846 generic.go:334] "Generic (PLEG): container finished" podID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerID="bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc" exitCode=0 Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.882053 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4l5h" event={"ID":"d100bd2b-ca02-4dd2-abec-394722bee8a1","Type":"ContainerDied","Data":"bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc"} Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.882091 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k4l5h" event={"ID":"d100bd2b-ca02-4dd2-abec-394722bee8a1","Type":"ContainerDied","Data":"454524e03a9da411db92ec7ece771b6183162efdf6f082704dfb1a3bc025997e"} Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.882099 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k4l5h" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.882112 4846 scope.go:117] "RemoveContainer" containerID="bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.902362 4846 scope.go:117] "RemoveContainer" containerID="a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.932691 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k4l5h"] Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.938523 4846 scope.go:117] "RemoveContainer" containerID="4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.943835 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k4l5h"] Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.998675 4846 scope.go:117] "RemoveContainer" containerID="bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc" Feb 02 14:08:04 crc kubenswrapper[4846]: E0202 14:08:04.999479 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc\": container with ID starting with bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc not found: ID does not exist" containerID="bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.999522 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc"} err="failed to get container status \"bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc\": rpc error: code = NotFound desc = could not find container \"bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc\": container with ID starting with bcd07d2d906e37480bc5fdaf1dced515a4218e084447bfc5da3b65a0b49a36fc not found: ID does not exist" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.999551 4846 scope.go:117] "RemoveContainer" containerID="a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b" Feb 02 14:08:04 crc kubenswrapper[4846]: E0202 14:08:04.999875 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b\": container with ID starting with a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b not found: ID does not exist" containerID="a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.999907 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b"} err="failed to get container status \"a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b\": rpc error: code = NotFound desc = could not find container \"a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b\": container with ID starting with a84f49fbeb101ad6d6576e043a567e96148c8fa4b14ebc8704083766621f2c0b not found: ID does not exist" Feb 02 14:08:04 crc kubenswrapper[4846]: I0202 14:08:04.999920 4846 scope.go:117] "RemoveContainer" containerID="4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209" Feb 02 14:08:05 crc kubenswrapper[4846]: E0202 14:08:05.000218 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209\": container with ID starting with 4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209 not found: ID does not exist" containerID="4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209" Feb 02 14:08:05 crc kubenswrapper[4846]: I0202 14:08:05.000265 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209"} err="failed to get container status \"4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209\": rpc error: code = NotFound desc = could not find container \"4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209\": container with ID starting with 4ed4e3cab924fc6366310f66bcd0de4f8dfa143d224ebb3a46cd861d85fbd209 not found: ID does not exist" Feb 02 14:08:05 crc kubenswrapper[4846]: I0202 14:08:05.435371 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d100bd2b-ca02-4dd2-abec-394722bee8a1" path="/var/lib/kubelet/pods/d100bd2b-ca02-4dd2-abec-394722bee8a1/volumes" Feb 02 14:08:06 crc kubenswrapper[4846]: I0202 14:08:06.425686 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:08:06 crc kubenswrapper[4846]: E0202 14:08:06.426097 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:08:06 crc kubenswrapper[4846]: I0202 14:08:06.904266 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e22c32b-e1f8-4bc4-89e1-e27acb1117ef" containerID="587b813fae819d8b8fde8198cc0cc36b691b3e3dabd949fa8959d2c5255ebe2a" exitCode=0 Feb 02 14:08:06 crc kubenswrapper[4846]: I0202 14:08:06.904392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" event={"ID":"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef","Type":"ContainerDied","Data":"587b813fae819d8b8fde8198cc0cc36b691b3e3dabd949fa8959d2c5255ebe2a"} Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.399064 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.503887 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-tripleo-cleanup-combined-ca-bundle\") pod \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.503967 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-ssh-key-openstack-cell1\") pod \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.504088 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzrdb\" (UniqueName: \"kubernetes.io/projected/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-kube-api-access-rzrdb\") pod \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.504195 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-inventory\") pod \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\" (UID: \"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef\") " Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.510780 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-kube-api-access-rzrdb" (OuterVolumeSpecName: "kube-api-access-rzrdb") pod "5e22c32b-e1f8-4bc4-89e1-e27acb1117ef" (UID: "5e22c32b-e1f8-4bc4-89e1-e27acb1117ef"). InnerVolumeSpecName "kube-api-access-rzrdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.510923 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "5e22c32b-e1f8-4bc4-89e1-e27acb1117ef" (UID: "5e22c32b-e1f8-4bc4-89e1-e27acb1117ef"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.535901 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-inventory" (OuterVolumeSpecName: "inventory") pod "5e22c32b-e1f8-4bc4-89e1-e27acb1117ef" (UID: "5e22c32b-e1f8-4bc4-89e1-e27acb1117ef"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.541838 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "5e22c32b-e1f8-4bc4-89e1-e27acb1117ef" (UID: "5e22c32b-e1f8-4bc4-89e1-e27acb1117ef"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.606996 4846 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.607342 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.607360 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzrdb\" (UniqueName: \"kubernetes.io/projected/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-kube-api-access-rzrdb\") on node \"crc\" DevicePath \"\"" Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.607371 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e22c32b-e1f8-4bc4-89e1-e27acb1117ef-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.924518 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" event={"ID":"5e22c32b-e1f8-4bc4-89e1-e27acb1117ef","Type":"ContainerDied","Data":"92dc655a7274c2ac58b73d6f6fd85c228f18caeaf11d53e9781be1e922bc3b0c"} Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.924568 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92dc655a7274c2ac58b73d6f6fd85c228f18caeaf11d53e9781be1e922bc3b0c" Feb 02 14:08:08 crc kubenswrapper[4846]: I0202 14:08:08.924643 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.421805 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-92q2f"] Feb 02 14:08:16 crc kubenswrapper[4846]: E0202 14:08:16.425054 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerName="registry-server" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.425074 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerName="registry-server" Feb 02 14:08:16 crc kubenswrapper[4846]: E0202 14:08:16.425100 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e22c32b-e1f8-4bc4-89e1-e27acb1117ef" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.425107 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e22c32b-e1f8-4bc4-89e1-e27acb1117ef" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Feb 02 14:08:16 crc kubenswrapper[4846]: E0202 14:08:16.425125 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerName="extract-content" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.425131 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerName="extract-content" Feb 02 14:08:16 crc kubenswrapper[4846]: E0202 14:08:16.425144 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerName="registry-server" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.425150 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerName="registry-server" Feb 02 14:08:16 crc kubenswrapper[4846]: E0202 14:08:16.425168 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerName="extract-utilities" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.425173 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerName="extract-utilities" Feb 02 14:08:16 crc kubenswrapper[4846]: E0202 14:08:16.425182 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerName="extract-utilities" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.425188 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerName="extract-utilities" Feb 02 14:08:16 crc kubenswrapper[4846]: E0202 14:08:16.425197 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerName="extract-content" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.425202 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerName="extract-content" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.425386 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e2e1a9-487b-45aa-abad-dfac6028b1e9" containerName="registry-server" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.425401 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e22c32b-e1f8-4bc4-89e1-e27acb1117ef" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.425428 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d100bd2b-ca02-4dd2-abec-394722bee8a1" containerName="registry-server" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.426490 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.430248 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.430296 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.430493 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.435452 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-92q2f"] Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.440908 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.591016 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.591095 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.591240 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-inventory\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.591318 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-467bl\" (UniqueName: \"kubernetes.io/projected/b0606b34-07f7-4eab-96e1-4601339b9bfa-kube-api-access-467bl\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.693834 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-inventory\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.693937 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-467bl\" (UniqueName: \"kubernetes.io/projected/b0606b34-07f7-4eab-96e1-4601339b9bfa-kube-api-access-467bl\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.694044 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.694076 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.708454 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.708901 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.710259 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-inventory\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.712414 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-467bl\" (UniqueName: \"kubernetes.io/projected/b0606b34-07f7-4eab-96e1-4601339b9bfa-kube-api-access-467bl\") pod \"bootstrap-openstack-openstack-cell1-92q2f\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:16 crc kubenswrapper[4846]: I0202 14:08:16.753693 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:08:17 crc kubenswrapper[4846]: I0202 14:08:17.310771 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-92q2f"] Feb 02 14:08:18 crc kubenswrapper[4846]: I0202 14:08:18.011176 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" event={"ID":"b0606b34-07f7-4eab-96e1-4601339b9bfa","Type":"ContainerStarted","Data":"17f07d6c04cc38ec0a2abb54240c547383e4224af3d8599b85530cc390fbb4bb"} Feb 02 14:08:19 crc kubenswrapper[4846]: I0202 14:08:19.022438 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" event={"ID":"b0606b34-07f7-4eab-96e1-4601339b9bfa","Type":"ContainerStarted","Data":"e9fc42db4393fbe5e5310f6cefc10970e2e482a0d47cd7ee7eef4da97039a2d2"} Feb 02 14:08:19 crc kubenswrapper[4846]: I0202 14:08:19.041416 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" podStartSLOduration=2.480370355 podStartE2EDuration="3.0413981s" podCreationTimestamp="2026-02-02 14:08:16 +0000 UTC" firstStartedPulling="2026-02-02 14:08:17.31236876 +0000 UTC m=+7128.540955623" lastFinishedPulling="2026-02-02 14:08:17.873396505 +0000 UTC m=+7129.101983368" observedRunningTime="2026-02-02 14:08:19.039887763 +0000 UTC m=+7130.268474646" watchObservedRunningTime="2026-02-02 14:08:19.0413981 +0000 UTC m=+7130.269984963" Feb 02 14:08:21 crc kubenswrapper[4846]: I0202 14:08:21.424749 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:08:21 crc kubenswrapper[4846]: E0202 14:08:21.425410 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:08:34 crc kubenswrapper[4846]: I0202 14:08:34.423049 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:08:34 crc kubenswrapper[4846]: E0202 14:08:34.423828 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:08:49 crc kubenswrapper[4846]: I0202 14:08:49.433393 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:08:49 crc kubenswrapper[4846]: E0202 14:08:49.434262 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:09:01 crc kubenswrapper[4846]: I0202 14:09:01.423451 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:09:01 crc kubenswrapper[4846]: E0202 14:09:01.424323 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:09:12 crc kubenswrapper[4846]: I0202 14:09:12.424074 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:09:12 crc kubenswrapper[4846]: E0202 14:09:12.424981 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:09:25 crc kubenswrapper[4846]: I0202 14:09:25.424321 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:09:25 crc kubenswrapper[4846]: E0202 14:09:25.426848 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:09:39 crc kubenswrapper[4846]: I0202 14:09:39.438334 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:09:39 crc kubenswrapper[4846]: E0202 14:09:39.439337 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:09:51 crc kubenswrapper[4846]: I0202 14:09:51.423761 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:09:51 crc kubenswrapper[4846]: E0202 14:09:51.424604 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:10:03 crc kubenswrapper[4846]: I0202 14:10:03.423822 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:10:03 crc kubenswrapper[4846]: E0202 14:10:03.424734 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:10:17 crc kubenswrapper[4846]: I0202 14:10:17.424123 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:10:17 crc kubenswrapper[4846]: E0202 14:10:17.425063 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:10:31 crc kubenswrapper[4846]: I0202 14:10:31.423722 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:10:31 crc kubenswrapper[4846]: E0202 14:10:31.424708 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:10:45 crc kubenswrapper[4846]: I0202 14:10:45.424222 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:10:45 crc kubenswrapper[4846]: E0202 14:10:45.425206 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:10:56 crc kubenswrapper[4846]: I0202 14:10:56.424048 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:10:56 crc kubenswrapper[4846]: E0202 14:10:56.424924 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:11:09 crc kubenswrapper[4846]: I0202 14:11:09.433649 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:11:09 crc kubenswrapper[4846]: E0202 14:11:09.434745 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:11:20 crc kubenswrapper[4846]: I0202 14:11:20.424433 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:11:20 crc kubenswrapper[4846]: E0202 14:11:20.425586 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:11:26 crc kubenswrapper[4846]: I0202 14:11:26.860183 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0606b34-07f7-4eab-96e1-4601339b9bfa" containerID="e9fc42db4393fbe5e5310f6cefc10970e2e482a0d47cd7ee7eef4da97039a2d2" exitCode=0 Feb 02 14:11:26 crc kubenswrapper[4846]: I0202 14:11:26.860261 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" event={"ID":"b0606b34-07f7-4eab-96e1-4601339b9bfa","Type":"ContainerDied","Data":"e9fc42db4393fbe5e5310f6cefc10970e2e482a0d47cd7ee7eef4da97039a2d2"} Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.385542 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.446189 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-467bl\" (UniqueName: \"kubernetes.io/projected/b0606b34-07f7-4eab-96e1-4601339b9bfa-kube-api-access-467bl\") pod \"b0606b34-07f7-4eab-96e1-4601339b9bfa\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.446259 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-inventory\") pod \"b0606b34-07f7-4eab-96e1-4601339b9bfa\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.446385 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-ssh-key-openstack-cell1\") pod \"b0606b34-07f7-4eab-96e1-4601339b9bfa\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.446487 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-bootstrap-combined-ca-bundle\") pod \"b0606b34-07f7-4eab-96e1-4601339b9bfa\" (UID: \"b0606b34-07f7-4eab-96e1-4601339b9bfa\") " Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.465434 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b0606b34-07f7-4eab-96e1-4601339b9bfa" (UID: "b0606b34-07f7-4eab-96e1-4601339b9bfa"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.470850 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0606b34-07f7-4eab-96e1-4601339b9bfa-kube-api-access-467bl" (OuterVolumeSpecName: "kube-api-access-467bl") pod "b0606b34-07f7-4eab-96e1-4601339b9bfa" (UID: "b0606b34-07f7-4eab-96e1-4601339b9bfa"). InnerVolumeSpecName "kube-api-access-467bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.517806 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-inventory" (OuterVolumeSpecName: "inventory") pod "b0606b34-07f7-4eab-96e1-4601339b9bfa" (UID: "b0606b34-07f7-4eab-96e1-4601339b9bfa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.525735 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "b0606b34-07f7-4eab-96e1-4601339b9bfa" (UID: "b0606b34-07f7-4eab-96e1-4601339b9bfa"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.553184 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-467bl\" (UniqueName: \"kubernetes.io/projected/b0606b34-07f7-4eab-96e1-4601339b9bfa-kube-api-access-467bl\") on node \"crc\" DevicePath \"\"" Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.553257 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.553275 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.553289 4846 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0606b34-07f7-4eab-96e1-4601339b9bfa-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.881925 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" event={"ID":"b0606b34-07f7-4eab-96e1-4601339b9bfa","Type":"ContainerDied","Data":"17f07d6c04cc38ec0a2abb54240c547383e4224af3d8599b85530cc390fbb4bb"} Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.882246 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17f07d6c04cc38ec0a2abb54240c547383e4224af3d8599b85530cc390fbb4bb" Feb 02 14:11:28 crc kubenswrapper[4846]: I0202 14:11:28.882098 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-92q2f" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.001725 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-hlgvr"] Feb 02 14:11:29 crc kubenswrapper[4846]: E0202 14:11:29.002368 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0606b34-07f7-4eab-96e1-4601339b9bfa" containerName="bootstrap-openstack-openstack-cell1" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.002394 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0606b34-07f7-4eab-96e1-4601339b9bfa" containerName="bootstrap-openstack-openstack-cell1" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.002687 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0606b34-07f7-4eab-96e1-4601339b9bfa" containerName="bootstrap-openstack-openstack-cell1" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.003849 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.008358 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.009234 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.009405 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.009589 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.019893 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-hlgvr"] Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.073399 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-inventory\") pod \"download-cache-openstack-openstack-cell1-hlgvr\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.073567 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nppj\" (UniqueName: \"kubernetes.io/projected/dac40561-426a-470e-8af9-05e0aea97e41-kube-api-access-8nppj\") pod \"download-cache-openstack-openstack-cell1-hlgvr\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.073713 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-hlgvr\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.165516 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tm9t2"] Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.168133 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.175424 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-inventory\") pod \"download-cache-openstack-openstack-cell1-hlgvr\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.175844 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nppj\" (UniqueName: \"kubernetes.io/projected/dac40561-426a-470e-8af9-05e0aea97e41-kube-api-access-8nppj\") pod \"download-cache-openstack-openstack-cell1-hlgvr\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.176075 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-hlgvr\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.182543 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-inventory\") pod \"download-cache-openstack-openstack-cell1-hlgvr\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.186212 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm9t2"] Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.186865 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-hlgvr\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.207152 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nppj\" (UniqueName: \"kubernetes.io/projected/dac40561-426a-470e-8af9-05e0aea97e41-kube-api-access-8nppj\") pod \"download-cache-openstack-openstack-cell1-hlgvr\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.289897 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-catalog-content\") pod \"redhat-marketplace-tm9t2\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.290275 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-utilities\") pod \"redhat-marketplace-tm9t2\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.290427 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nhzk\" (UniqueName: \"kubernetes.io/projected/bb288c36-18ec-4c18-bf8a-1d802170c514-kube-api-access-5nhzk\") pod \"redhat-marketplace-tm9t2\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.337863 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.392411 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-catalog-content\") pod \"redhat-marketplace-tm9t2\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.393987 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-utilities\") pod \"redhat-marketplace-tm9t2\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.394207 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nhzk\" (UniqueName: \"kubernetes.io/projected/bb288c36-18ec-4c18-bf8a-1d802170c514-kube-api-access-5nhzk\") pod \"redhat-marketplace-tm9t2\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.394714 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-catalog-content\") pod \"redhat-marketplace-tm9t2\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.395012 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-utilities\") pod \"redhat-marketplace-tm9t2\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.414331 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nhzk\" (UniqueName: \"kubernetes.io/projected/bb288c36-18ec-4c18-bf8a-1d802170c514-kube-api-access-5nhzk\") pod \"redhat-marketplace-tm9t2\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:29 crc kubenswrapper[4846]: I0202 14:11:29.501386 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:30 crc kubenswrapper[4846]: I0202 14:11:30.026466 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-hlgvr"] Feb 02 14:11:30 crc kubenswrapper[4846]: I0202 14:11:30.107695 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm9t2"] Feb 02 14:11:30 crc kubenswrapper[4846]: W0202 14:11:30.110670 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb288c36_18ec_4c18_bf8a_1d802170c514.slice/crio-a71d6b3f9c9dcc084621117323052848442c20e246fbac49f9809f6520f77410 WatchSource:0}: Error finding container a71d6b3f9c9dcc084621117323052848442c20e246fbac49f9809f6520f77410: Status 404 returned error can't find the container with id a71d6b3f9c9dcc084621117323052848442c20e246fbac49f9809f6520f77410 Feb 02 14:11:30 crc kubenswrapper[4846]: I0202 14:11:30.746648 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:11:30 crc kubenswrapper[4846]: I0202 14:11:30.903215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" event={"ID":"dac40561-426a-470e-8af9-05e0aea97e41","Type":"ContainerStarted","Data":"42c402261658b7ece4d756c7c71b3294dec2114b97352392951b1c073dfea7c7"} Feb 02 14:11:30 crc kubenswrapper[4846]: I0202 14:11:30.906446 4846 generic.go:334] "Generic (PLEG): container finished" podID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerID="0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006" exitCode=0 Feb 02 14:11:30 crc kubenswrapper[4846]: I0202 14:11:30.906503 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm9t2" event={"ID":"bb288c36-18ec-4c18-bf8a-1d802170c514","Type":"ContainerDied","Data":"0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006"} Feb 02 14:11:30 crc kubenswrapper[4846]: I0202 14:11:30.906538 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm9t2" event={"ID":"bb288c36-18ec-4c18-bf8a-1d802170c514","Type":"ContainerStarted","Data":"a71d6b3f9c9dcc084621117323052848442c20e246fbac49f9809f6520f77410"} Feb 02 14:11:31 crc kubenswrapper[4846]: I0202 14:11:31.424607 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:11:31 crc kubenswrapper[4846]: E0202 14:11:31.426047 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:11:31 crc kubenswrapper[4846]: I0202 14:11:31.922082 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" event={"ID":"dac40561-426a-470e-8af9-05e0aea97e41","Type":"ContainerStarted","Data":"e0d69b96d5ff12c8c28b42cecb13911a40c2b0d59c2492ed7c8ac70d8408b1ef"} Feb 02 14:11:31 crc kubenswrapper[4846]: I0202 14:11:31.971066 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" podStartSLOduration=3.26256121 podStartE2EDuration="3.971036996s" podCreationTimestamp="2026-02-02 14:11:28 +0000 UTC" firstStartedPulling="2026-02-02 14:11:30.034888106 +0000 UTC m=+7321.263474969" lastFinishedPulling="2026-02-02 14:11:30.743363892 +0000 UTC m=+7321.971950755" observedRunningTime="2026-02-02 14:11:31.94443242 +0000 UTC m=+7323.173019303" watchObservedRunningTime="2026-02-02 14:11:31.971036996 +0000 UTC m=+7323.199623859" Feb 02 14:11:32 crc kubenswrapper[4846]: I0202 14:11:32.939024 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm9t2" event={"ID":"bb288c36-18ec-4c18-bf8a-1d802170c514","Type":"ContainerStarted","Data":"8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e"} Feb 02 14:11:33 crc kubenswrapper[4846]: I0202 14:11:33.954536 4846 generic.go:334] "Generic (PLEG): container finished" podID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerID="8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e" exitCode=0 Feb 02 14:11:33 crc kubenswrapper[4846]: I0202 14:11:33.954656 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm9t2" event={"ID":"bb288c36-18ec-4c18-bf8a-1d802170c514","Type":"ContainerDied","Data":"8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e"} Feb 02 14:11:34 crc kubenswrapper[4846]: I0202 14:11:34.969962 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm9t2" event={"ID":"bb288c36-18ec-4c18-bf8a-1d802170c514","Type":"ContainerStarted","Data":"cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d"} Feb 02 14:11:35 crc kubenswrapper[4846]: I0202 14:11:35.001968 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tm9t2" podStartSLOduration=2.261974366 podStartE2EDuration="6.001944903s" podCreationTimestamp="2026-02-02 14:11:29 +0000 UTC" firstStartedPulling="2026-02-02 14:11:30.909189928 +0000 UTC m=+7322.137776791" lastFinishedPulling="2026-02-02 14:11:34.649160465 +0000 UTC m=+7325.877747328" observedRunningTime="2026-02-02 14:11:34.99772546 +0000 UTC m=+7326.226312343" watchObservedRunningTime="2026-02-02 14:11:35.001944903 +0000 UTC m=+7326.230531766" Feb 02 14:11:39 crc kubenswrapper[4846]: I0202 14:11:39.503038 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:39 crc kubenswrapper[4846]: I0202 14:11:39.503752 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:39 crc kubenswrapper[4846]: I0202 14:11:39.560181 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:40 crc kubenswrapper[4846]: I0202 14:11:40.063760 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:40 crc kubenswrapper[4846]: I0202 14:11:40.118641 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm9t2"] Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.030751 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tm9t2" podUID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerName="registry-server" containerID="cri-o://cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d" gracePeriod=2 Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.640650 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.724457 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-catalog-content\") pod \"bb288c36-18ec-4c18-bf8a-1d802170c514\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.725075 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-utilities\") pod \"bb288c36-18ec-4c18-bf8a-1d802170c514\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.725811 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nhzk\" (UniqueName: \"kubernetes.io/projected/bb288c36-18ec-4c18-bf8a-1d802170c514-kube-api-access-5nhzk\") pod \"bb288c36-18ec-4c18-bf8a-1d802170c514\" (UID: \"bb288c36-18ec-4c18-bf8a-1d802170c514\") " Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.726070 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-utilities" (OuterVolumeSpecName: "utilities") pod "bb288c36-18ec-4c18-bf8a-1d802170c514" (UID: "bb288c36-18ec-4c18-bf8a-1d802170c514"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.726445 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.731337 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb288c36-18ec-4c18-bf8a-1d802170c514-kube-api-access-5nhzk" (OuterVolumeSpecName: "kube-api-access-5nhzk") pod "bb288c36-18ec-4c18-bf8a-1d802170c514" (UID: "bb288c36-18ec-4c18-bf8a-1d802170c514"). InnerVolumeSpecName "kube-api-access-5nhzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.746578 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb288c36-18ec-4c18-bf8a-1d802170c514" (UID: "bb288c36-18ec-4c18-bf8a-1d802170c514"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.828351 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nhzk\" (UniqueName: \"kubernetes.io/projected/bb288c36-18ec-4c18-bf8a-1d802170c514-kube-api-access-5nhzk\") on node \"crc\" DevicePath \"\"" Feb 02 14:11:42 crc kubenswrapper[4846]: I0202 14:11:42.828395 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb288c36-18ec-4c18-bf8a-1d802170c514-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.044563 4846 generic.go:334] "Generic (PLEG): container finished" podID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerID="cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d" exitCode=0 Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.044609 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm9t2" event={"ID":"bb288c36-18ec-4c18-bf8a-1d802170c514","Type":"ContainerDied","Data":"cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d"} Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.044668 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tm9t2" event={"ID":"bb288c36-18ec-4c18-bf8a-1d802170c514","Type":"ContainerDied","Data":"a71d6b3f9c9dcc084621117323052848442c20e246fbac49f9809f6520f77410"} Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.044690 4846 scope.go:117] "RemoveContainer" containerID="cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.044703 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tm9t2" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.085408 4846 scope.go:117] "RemoveContainer" containerID="8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.095928 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm9t2"] Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.121699 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tm9t2"] Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.122381 4846 scope.go:117] "RemoveContainer" containerID="0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.160716 4846 scope.go:117] "RemoveContainer" containerID="cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d" Feb 02 14:11:43 crc kubenswrapper[4846]: E0202 14:11:43.161488 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d\": container with ID starting with cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d not found: ID does not exist" containerID="cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.161521 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d"} err="failed to get container status \"cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d\": rpc error: code = NotFound desc = could not find container \"cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d\": container with ID starting with cdd871c76d667433d57f107f9cabdca0da74fceeafb25327e81bb56e0e6b778d not found: ID does not exist" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.161542 4846 scope.go:117] "RemoveContainer" containerID="8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e" Feb 02 14:11:43 crc kubenswrapper[4846]: E0202 14:11:43.162206 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e\": container with ID starting with 8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e not found: ID does not exist" containerID="8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.162240 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e"} err="failed to get container status \"8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e\": rpc error: code = NotFound desc = could not find container \"8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e\": container with ID starting with 8a9705ae034893fc83fc0c9ceec6014a072a09e0d19188f4d4d1c046c608412e not found: ID does not exist" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.162262 4846 scope.go:117] "RemoveContainer" containerID="0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006" Feb 02 14:11:43 crc kubenswrapper[4846]: E0202 14:11:43.162507 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006\": container with ID starting with 0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006 not found: ID does not exist" containerID="0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.162540 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006"} err="failed to get container status \"0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006\": rpc error: code = NotFound desc = could not find container \"0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006\": container with ID starting with 0ec5ed104d9b501393f367d3c3eb16b8a095da76a9ac67da7e4b7f6844b97006 not found: ID does not exist" Feb 02 14:11:43 crc kubenswrapper[4846]: I0202 14:11:43.436190 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb288c36-18ec-4c18-bf8a-1d802170c514" path="/var/lib/kubelet/pods/bb288c36-18ec-4c18-bf8a-1d802170c514/volumes" Feb 02 14:11:44 crc kubenswrapper[4846]: I0202 14:11:44.423175 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:11:44 crc kubenswrapper[4846]: E0202 14:11:44.423874 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:11:56 crc kubenswrapper[4846]: I0202 14:11:56.423930 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:11:56 crc kubenswrapper[4846]: E0202 14:11:56.424755 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:12:11 crc kubenswrapper[4846]: I0202 14:12:11.424309 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:12:12 crc kubenswrapper[4846]: I0202 14:12:12.336361 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"2eb1e85e549a6d9036406934409325d3a7ed0792302a62a04d481a0163f45cb9"} Feb 02 14:13:02 crc kubenswrapper[4846]: I0202 14:13:02.904471 4846 generic.go:334] "Generic (PLEG): container finished" podID="dac40561-426a-470e-8af9-05e0aea97e41" containerID="e0d69b96d5ff12c8c28b42cecb13911a40c2b0d59c2492ed7c8ac70d8408b1ef" exitCode=0 Feb 02 14:13:02 crc kubenswrapper[4846]: I0202 14:13:02.904558 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" event={"ID":"dac40561-426a-470e-8af9-05e0aea97e41","Type":"ContainerDied","Data":"e0d69b96d5ff12c8c28b42cecb13911a40c2b0d59c2492ed7c8ac70d8408b1ef"} Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.434047 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.578902 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nppj\" (UniqueName: \"kubernetes.io/projected/dac40561-426a-470e-8af9-05e0aea97e41-kube-api-access-8nppj\") pod \"dac40561-426a-470e-8af9-05e0aea97e41\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.579094 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-inventory\") pod \"dac40561-426a-470e-8af9-05e0aea97e41\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.579185 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-ssh-key-openstack-cell1\") pod \"dac40561-426a-470e-8af9-05e0aea97e41\" (UID: \"dac40561-426a-470e-8af9-05e0aea97e41\") " Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.586547 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dac40561-426a-470e-8af9-05e0aea97e41-kube-api-access-8nppj" (OuterVolumeSpecName: "kube-api-access-8nppj") pod "dac40561-426a-470e-8af9-05e0aea97e41" (UID: "dac40561-426a-470e-8af9-05e0aea97e41"). InnerVolumeSpecName "kube-api-access-8nppj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.622972 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "dac40561-426a-470e-8af9-05e0aea97e41" (UID: "dac40561-426a-470e-8af9-05e0aea97e41"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.626714 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-inventory" (OuterVolumeSpecName: "inventory") pod "dac40561-426a-470e-8af9-05e0aea97e41" (UID: "dac40561-426a-470e-8af9-05e0aea97e41"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.681774 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.681812 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nppj\" (UniqueName: \"kubernetes.io/projected/dac40561-426a-470e-8af9-05e0aea97e41-kube-api-access-8nppj\") on node \"crc\" DevicePath \"\"" Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.681824 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dac40561-426a-470e-8af9-05e0aea97e41-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.925860 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" event={"ID":"dac40561-426a-470e-8af9-05e0aea97e41","Type":"ContainerDied","Data":"42c402261658b7ece4d756c7c71b3294dec2114b97352392951b1c073dfea7c7"} Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.926157 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42c402261658b7ece4d756c7c71b3294dec2114b97352392951b1c073dfea7c7" Feb 02 14:13:04 crc kubenswrapper[4846]: I0202 14:13:04.926213 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hlgvr" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.023959 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-ktqhx"] Feb 02 14:13:05 crc kubenswrapper[4846]: E0202 14:13:05.024573 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerName="extract-utilities" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.024592 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerName="extract-utilities" Feb 02 14:13:05 crc kubenswrapper[4846]: E0202 14:13:05.024641 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dac40561-426a-470e-8af9-05e0aea97e41" containerName="download-cache-openstack-openstack-cell1" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.024651 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac40561-426a-470e-8af9-05e0aea97e41" containerName="download-cache-openstack-openstack-cell1" Feb 02 14:13:05 crc kubenswrapper[4846]: E0202 14:13:05.024665 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerName="extract-content" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.024674 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerName="extract-content" Feb 02 14:13:05 crc kubenswrapper[4846]: E0202 14:13:05.024726 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerName="registry-server" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.024735 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerName="registry-server" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.024976 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="dac40561-426a-470e-8af9-05e0aea97e41" containerName="download-cache-openstack-openstack-cell1" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.025009 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb288c36-18ec-4c18-bf8a-1d802170c514" containerName="registry-server" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.026045 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.029159 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.029840 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.030243 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.055696 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-ktqhx"] Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.035575 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.196942 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-inventory\") pod \"configure-network-openstack-openstack-cell1-ktqhx\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.197055 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7svr6\" (UniqueName: \"kubernetes.io/projected/8284a56e-0573-4a1e-85a0-58e62178133d-kube-api-access-7svr6\") pod \"configure-network-openstack-openstack-cell1-ktqhx\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.197106 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-ktqhx\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.299560 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7svr6\" (UniqueName: \"kubernetes.io/projected/8284a56e-0573-4a1e-85a0-58e62178133d-kube-api-access-7svr6\") pod \"configure-network-openstack-openstack-cell1-ktqhx\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.299780 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-ktqhx\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.300148 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-inventory\") pod \"configure-network-openstack-openstack-cell1-ktqhx\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.308072 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-inventory\") pod \"configure-network-openstack-openstack-cell1-ktqhx\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.308072 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-ktqhx\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.323309 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7svr6\" (UniqueName: \"kubernetes.io/projected/8284a56e-0573-4a1e-85a0-58e62178133d-kube-api-access-7svr6\") pod \"configure-network-openstack-openstack-cell1-ktqhx\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.398447 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.978885 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-ktqhx"] Feb 02 14:13:05 crc kubenswrapper[4846]: W0202 14:13:05.984770 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8284a56e_0573_4a1e_85a0_58e62178133d.slice/crio-3366b432fa5a6152aca1d73c6e047389fd33bfbe78e45dc3d5551c77e4045adb WatchSource:0}: Error finding container 3366b432fa5a6152aca1d73c6e047389fd33bfbe78e45dc3d5551c77e4045adb: Status 404 returned error can't find the container with id 3366b432fa5a6152aca1d73c6e047389fd33bfbe78e45dc3d5551c77e4045adb Feb 02 14:13:05 crc kubenswrapper[4846]: I0202 14:13:05.992650 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 14:13:06 crc kubenswrapper[4846]: I0202 14:13:06.943838 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" event={"ID":"8284a56e-0573-4a1e-85a0-58e62178133d","Type":"ContainerStarted","Data":"72eaae8b8a5bc858cb75e63eab59a05efdfdc96d9f0cfc8ac412ccfebeabbbde"} Feb 02 14:13:06 crc kubenswrapper[4846]: I0202 14:13:06.944150 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" event={"ID":"8284a56e-0573-4a1e-85a0-58e62178133d","Type":"ContainerStarted","Data":"3366b432fa5a6152aca1d73c6e047389fd33bfbe78e45dc3d5551c77e4045adb"} Feb 02 14:13:06 crc kubenswrapper[4846]: I0202 14:13:06.970718 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" podStartSLOduration=2.542947347 podStartE2EDuration="2.970695535s" podCreationTimestamp="2026-02-02 14:13:04 +0000 UTC" firstStartedPulling="2026-02-02 14:13:05.992395797 +0000 UTC m=+7417.220982660" lastFinishedPulling="2026-02-02 14:13:06.420143985 +0000 UTC m=+7417.648730848" observedRunningTime="2026-02-02 14:13:06.963673625 +0000 UTC m=+7418.192260498" watchObservedRunningTime="2026-02-02 14:13:06.970695535 +0000 UTC m=+7418.199282418" Feb 02 14:14:27 crc kubenswrapper[4846]: I0202 14:14:27.730893 4846 generic.go:334] "Generic (PLEG): container finished" podID="8284a56e-0573-4a1e-85a0-58e62178133d" containerID="72eaae8b8a5bc858cb75e63eab59a05efdfdc96d9f0cfc8ac412ccfebeabbbde" exitCode=0 Feb 02 14:14:27 crc kubenswrapper[4846]: I0202 14:14:27.730983 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" event={"ID":"8284a56e-0573-4a1e-85a0-58e62178133d","Type":"ContainerDied","Data":"72eaae8b8a5bc858cb75e63eab59a05efdfdc96d9f0cfc8ac412ccfebeabbbde"} Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.237528 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.309657 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7svr6\" (UniqueName: \"kubernetes.io/projected/8284a56e-0573-4a1e-85a0-58e62178133d-kube-api-access-7svr6\") pod \"8284a56e-0573-4a1e-85a0-58e62178133d\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.309938 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-inventory\") pod \"8284a56e-0573-4a1e-85a0-58e62178133d\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.309983 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-ssh-key-openstack-cell1\") pod \"8284a56e-0573-4a1e-85a0-58e62178133d\" (UID: \"8284a56e-0573-4a1e-85a0-58e62178133d\") " Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.322923 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8284a56e-0573-4a1e-85a0-58e62178133d-kube-api-access-7svr6" (OuterVolumeSpecName: "kube-api-access-7svr6") pod "8284a56e-0573-4a1e-85a0-58e62178133d" (UID: "8284a56e-0573-4a1e-85a0-58e62178133d"). InnerVolumeSpecName "kube-api-access-7svr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.342073 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "8284a56e-0573-4a1e-85a0-58e62178133d" (UID: "8284a56e-0573-4a1e-85a0-58e62178133d"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.354366 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-inventory" (OuterVolumeSpecName: "inventory") pod "8284a56e-0573-4a1e-85a0-58e62178133d" (UID: "8284a56e-0573-4a1e-85a0-58e62178133d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.413444 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.413479 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/8284a56e-0573-4a1e-85a0-58e62178133d-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.413496 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7svr6\" (UniqueName: \"kubernetes.io/projected/8284a56e-0573-4a1e-85a0-58e62178133d-kube-api-access-7svr6\") on node \"crc\" DevicePath \"\"" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.753580 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" event={"ID":"8284a56e-0573-4a1e-85a0-58e62178133d","Type":"ContainerDied","Data":"3366b432fa5a6152aca1d73c6e047389fd33bfbe78e45dc3d5551c77e4045adb"} Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.754915 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3366b432fa5a6152aca1d73c6e047389fd33bfbe78e45dc3d5551c77e4045adb" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.754138 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-ktqhx" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.851587 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-8hgxz"] Feb 02 14:14:29 crc kubenswrapper[4846]: E0202 14:14:29.852066 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8284a56e-0573-4a1e-85a0-58e62178133d" containerName="configure-network-openstack-openstack-cell1" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.852086 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8284a56e-0573-4a1e-85a0-58e62178133d" containerName="configure-network-openstack-openstack-cell1" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.852279 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8284a56e-0573-4a1e-85a0-58e62178133d" containerName="configure-network-openstack-openstack-cell1" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.857243 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.860367 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.860665 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.860836 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.860979 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.870917 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-8hgxz"] Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.923894 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-inventory\") pod \"validate-network-openstack-openstack-cell1-8hgxz\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.924083 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-8hgxz\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:29 crc kubenswrapper[4846]: I0202 14:14:29.924111 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmwvp\" (UniqueName: \"kubernetes.io/projected/55b40167-7f63-4a89-8043-2e83365fbc1b-kube-api-access-hmwvp\") pod \"validate-network-openstack-openstack-cell1-8hgxz\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.026364 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-inventory\") pod \"validate-network-openstack-openstack-cell1-8hgxz\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.026568 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-8hgxz\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.026598 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmwvp\" (UniqueName: \"kubernetes.io/projected/55b40167-7f63-4a89-8043-2e83365fbc1b-kube-api-access-hmwvp\") pod \"validate-network-openstack-openstack-cell1-8hgxz\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.031146 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-inventory\") pod \"validate-network-openstack-openstack-cell1-8hgxz\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.038198 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-8hgxz\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.061684 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmwvp\" (UniqueName: \"kubernetes.io/projected/55b40167-7f63-4a89-8043-2e83365fbc1b-kube-api-access-hmwvp\") pod \"validate-network-openstack-openstack-cell1-8hgxz\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.172106 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.478819 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.479130 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.719554 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-8hgxz"] Feb 02 14:14:30 crc kubenswrapper[4846]: I0202 14:14:30.764482 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" event={"ID":"55b40167-7f63-4a89-8043-2e83365fbc1b","Type":"ContainerStarted","Data":"926b3f005d107ab191548f0da12dbfe988d3d8bf094dd3fa7e915cf59a25e817"} Feb 02 14:14:31 crc kubenswrapper[4846]: I0202 14:14:31.776719 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" event={"ID":"55b40167-7f63-4a89-8043-2e83365fbc1b","Type":"ContainerStarted","Data":"08f53b7c1b6b554397e45ee021646ac8a51cba1803c9dca002d42f701994dd1b"} Feb 02 14:14:31 crc kubenswrapper[4846]: I0202 14:14:31.795609 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" podStartSLOduration=2.252374525 podStartE2EDuration="2.795591897s" podCreationTimestamp="2026-02-02 14:14:29 +0000 UTC" firstStartedPulling="2026-02-02 14:14:30.730415519 +0000 UTC m=+7501.959002382" lastFinishedPulling="2026-02-02 14:14:31.273632891 +0000 UTC m=+7502.502219754" observedRunningTime="2026-02-02 14:14:31.79360322 +0000 UTC m=+7503.022190083" watchObservedRunningTime="2026-02-02 14:14:31.795591897 +0000 UTC m=+7503.024178760" Feb 02 14:14:36 crc kubenswrapper[4846]: I0202 14:14:36.834909 4846 generic.go:334] "Generic (PLEG): container finished" podID="55b40167-7f63-4a89-8043-2e83365fbc1b" containerID="08f53b7c1b6b554397e45ee021646ac8a51cba1803c9dca002d42f701994dd1b" exitCode=0 Feb 02 14:14:36 crc kubenswrapper[4846]: I0202 14:14:36.835019 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" event={"ID":"55b40167-7f63-4a89-8043-2e83365fbc1b","Type":"ContainerDied","Data":"08f53b7c1b6b554397e45ee021646ac8a51cba1803c9dca002d42f701994dd1b"} Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.415253 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.541108 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-ssh-key-openstack-cell1\") pod \"55b40167-7f63-4a89-8043-2e83365fbc1b\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.541280 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmwvp\" (UniqueName: \"kubernetes.io/projected/55b40167-7f63-4a89-8043-2e83365fbc1b-kube-api-access-hmwvp\") pod \"55b40167-7f63-4a89-8043-2e83365fbc1b\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.541362 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-inventory\") pod \"55b40167-7f63-4a89-8043-2e83365fbc1b\" (UID: \"55b40167-7f63-4a89-8043-2e83365fbc1b\") " Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.547845 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55b40167-7f63-4a89-8043-2e83365fbc1b-kube-api-access-hmwvp" (OuterVolumeSpecName: "kube-api-access-hmwvp") pod "55b40167-7f63-4a89-8043-2e83365fbc1b" (UID: "55b40167-7f63-4a89-8043-2e83365fbc1b"). InnerVolumeSpecName "kube-api-access-hmwvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.574207 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-inventory" (OuterVolumeSpecName: "inventory") pod "55b40167-7f63-4a89-8043-2e83365fbc1b" (UID: "55b40167-7f63-4a89-8043-2e83365fbc1b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.575073 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "55b40167-7f63-4a89-8043-2e83365fbc1b" (UID: "55b40167-7f63-4a89-8043-2e83365fbc1b"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.643973 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.644003 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/55b40167-7f63-4a89-8043-2e83365fbc1b-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.644013 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmwvp\" (UniqueName: \"kubernetes.io/projected/55b40167-7f63-4a89-8043-2e83365fbc1b-kube-api-access-hmwvp\") on node \"crc\" DevicePath \"\"" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.855974 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" event={"ID":"55b40167-7f63-4a89-8043-2e83365fbc1b","Type":"ContainerDied","Data":"926b3f005d107ab191548f0da12dbfe988d3d8bf094dd3fa7e915cf59a25e817"} Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.856021 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="926b3f005d107ab191548f0da12dbfe988d3d8bf094dd3fa7e915cf59a25e817" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.856380 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-8hgxz" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.943196 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-ds2qr"] Feb 02 14:14:38 crc kubenswrapper[4846]: E0202 14:14:38.944122 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b40167-7f63-4a89-8043-2e83365fbc1b" containerName="validate-network-openstack-openstack-cell1" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.944141 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b40167-7f63-4a89-8043-2e83365fbc1b" containerName="validate-network-openstack-openstack-cell1" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.944341 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="55b40167-7f63-4a89-8043-2e83365fbc1b" containerName="validate-network-openstack-openstack-cell1" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.944979 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.949857 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.950118 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.950258 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.950514 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:14:38 crc kubenswrapper[4846]: I0202 14:14:38.966174 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-ds2qr"] Feb 02 14:14:39 crc kubenswrapper[4846]: E0202 14:14:39.025121 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55b40167_7f63_4a89_8043_2e83365fbc1b.slice/crio-926b3f005d107ab191548f0da12dbfe988d3d8bf094dd3fa7e915cf59a25e817\": RecentStats: unable to find data in memory cache]" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.052650 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-ds2qr\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.052884 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w67kv\" (UniqueName: \"kubernetes.io/projected/1dc40137-f83c-4dd8-9a91-cda128768635-kube-api-access-w67kv\") pod \"install-os-openstack-openstack-cell1-ds2qr\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.052975 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-inventory\") pod \"install-os-openstack-openstack-cell1-ds2qr\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.155881 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-ds2qr\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.155979 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w67kv\" (UniqueName: \"kubernetes.io/projected/1dc40137-f83c-4dd8-9a91-cda128768635-kube-api-access-w67kv\") pod \"install-os-openstack-openstack-cell1-ds2qr\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.156018 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-inventory\") pod \"install-os-openstack-openstack-cell1-ds2qr\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.161515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-inventory\") pod \"install-os-openstack-openstack-cell1-ds2qr\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.162014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-ds2qr\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.175434 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w67kv\" (UniqueName: \"kubernetes.io/projected/1dc40137-f83c-4dd8-9a91-cda128768635-kube-api-access-w67kv\") pod \"install-os-openstack-openstack-cell1-ds2qr\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.267257 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.815977 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-ds2qr"] Feb 02 14:14:39 crc kubenswrapper[4846]: I0202 14:14:39.865377 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-ds2qr" event={"ID":"1dc40137-f83c-4dd8-9a91-cda128768635","Type":"ContainerStarted","Data":"f70c48eab287c9f08c28c13d4b33b26400264242770e206f4190293e9721d3ea"} Feb 02 14:14:40 crc kubenswrapper[4846]: I0202 14:14:40.875668 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-ds2qr" event={"ID":"1dc40137-f83c-4dd8-9a91-cda128768635","Type":"ContainerStarted","Data":"0bc50639773f3b038aff63ba4de349724d111da68bfa95959340425fcb6ece12"} Feb 02 14:14:40 crc kubenswrapper[4846]: I0202 14:14:40.903134 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-ds2qr" podStartSLOduration=2.2483071040000002 podStartE2EDuration="2.903114357s" podCreationTimestamp="2026-02-02 14:14:38 +0000 UTC" firstStartedPulling="2026-02-02 14:14:39.820067024 +0000 UTC m=+7511.048653887" lastFinishedPulling="2026-02-02 14:14:40.474874277 +0000 UTC m=+7511.703461140" observedRunningTime="2026-02-02 14:14:40.897294365 +0000 UTC m=+7512.125881258" watchObservedRunningTime="2026-02-02 14:14:40.903114357 +0000 UTC m=+7512.131701220" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.200094 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv"] Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.202868 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.205254 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.206385 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.209738 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv"] Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.296981 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4jzr\" (UniqueName: \"kubernetes.io/projected/06045b74-efe8-45b2-b491-0c754b34c5d9-kube-api-access-g4jzr\") pod \"collect-profiles-29500695-9jrzv\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.297036 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06045b74-efe8-45b2-b491-0c754b34c5d9-secret-volume\") pod \"collect-profiles-29500695-9jrzv\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.297118 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06045b74-efe8-45b2-b491-0c754b34c5d9-config-volume\") pod \"collect-profiles-29500695-9jrzv\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.400354 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4jzr\" (UniqueName: \"kubernetes.io/projected/06045b74-efe8-45b2-b491-0c754b34c5d9-kube-api-access-g4jzr\") pod \"collect-profiles-29500695-9jrzv\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.400466 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06045b74-efe8-45b2-b491-0c754b34c5d9-secret-volume\") pod \"collect-profiles-29500695-9jrzv\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.400549 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06045b74-efe8-45b2-b491-0c754b34c5d9-config-volume\") pod \"collect-profiles-29500695-9jrzv\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.401678 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06045b74-efe8-45b2-b491-0c754b34c5d9-config-volume\") pod \"collect-profiles-29500695-9jrzv\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.412494 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06045b74-efe8-45b2-b491-0c754b34c5d9-secret-volume\") pod \"collect-profiles-29500695-9jrzv\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.423480 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4jzr\" (UniqueName: \"kubernetes.io/projected/06045b74-efe8-45b2-b491-0c754b34c5d9-kube-api-access-g4jzr\") pod \"collect-profiles-29500695-9jrzv\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.479385 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.479842 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:15:00 crc kubenswrapper[4846]: I0202 14:15:00.530352 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:01 crc kubenswrapper[4846]: I0202 14:15:01.065873 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv"] Feb 02 14:15:01 crc kubenswrapper[4846]: W0202 14:15:01.072934 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06045b74_efe8_45b2_b491_0c754b34c5d9.slice/crio-1963fa43337c7968525f9a9d0ebaa01d1c1fae796dfbb8f4734c0fca0bb1a28d WatchSource:0}: Error finding container 1963fa43337c7968525f9a9d0ebaa01d1c1fae796dfbb8f4734c0fca0bb1a28d: Status 404 returned error can't find the container with id 1963fa43337c7968525f9a9d0ebaa01d1c1fae796dfbb8f4734c0fca0bb1a28d Feb 02 14:15:01 crc kubenswrapper[4846]: I0202 14:15:01.120270 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" event={"ID":"06045b74-efe8-45b2-b491-0c754b34c5d9","Type":"ContainerStarted","Data":"1963fa43337c7968525f9a9d0ebaa01d1c1fae796dfbb8f4734c0fca0bb1a28d"} Feb 02 14:15:02 crc kubenswrapper[4846]: I0202 14:15:02.140731 4846 generic.go:334] "Generic (PLEG): container finished" podID="06045b74-efe8-45b2-b491-0c754b34c5d9" containerID="8028f3cf13f779c508e09233f7d2d6ca366fecfa97e65d3f0d41a378fdcc7d2f" exitCode=0 Feb 02 14:15:02 crc kubenswrapper[4846]: I0202 14:15:02.141172 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" event={"ID":"06045b74-efe8-45b2-b491-0c754b34c5d9","Type":"ContainerDied","Data":"8028f3cf13f779c508e09233f7d2d6ca366fecfa97e65d3f0d41a378fdcc7d2f"} Feb 02 14:15:03 crc kubenswrapper[4846]: I0202 14:15:03.575754 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:03 crc kubenswrapper[4846]: I0202 14:15:03.703008 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06045b74-efe8-45b2-b491-0c754b34c5d9-config-volume\") pod \"06045b74-efe8-45b2-b491-0c754b34c5d9\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " Feb 02 14:15:03 crc kubenswrapper[4846]: I0202 14:15:03.703295 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06045b74-efe8-45b2-b491-0c754b34c5d9-secret-volume\") pod \"06045b74-efe8-45b2-b491-0c754b34c5d9\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " Feb 02 14:15:03 crc kubenswrapper[4846]: I0202 14:15:03.703445 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4jzr\" (UniqueName: \"kubernetes.io/projected/06045b74-efe8-45b2-b491-0c754b34c5d9-kube-api-access-g4jzr\") pod \"06045b74-efe8-45b2-b491-0c754b34c5d9\" (UID: \"06045b74-efe8-45b2-b491-0c754b34c5d9\") " Feb 02 14:15:03 crc kubenswrapper[4846]: I0202 14:15:03.704080 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06045b74-efe8-45b2-b491-0c754b34c5d9-config-volume" (OuterVolumeSpecName: "config-volume") pod "06045b74-efe8-45b2-b491-0c754b34c5d9" (UID: "06045b74-efe8-45b2-b491-0c754b34c5d9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 14:15:03 crc kubenswrapper[4846]: I0202 14:15:03.709406 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06045b74-efe8-45b2-b491-0c754b34c5d9-kube-api-access-g4jzr" (OuterVolumeSpecName: "kube-api-access-g4jzr") pod "06045b74-efe8-45b2-b491-0c754b34c5d9" (UID: "06045b74-efe8-45b2-b491-0c754b34c5d9"). InnerVolumeSpecName "kube-api-access-g4jzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:15:03 crc kubenswrapper[4846]: I0202 14:15:03.715106 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06045b74-efe8-45b2-b491-0c754b34c5d9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "06045b74-efe8-45b2-b491-0c754b34c5d9" (UID: "06045b74-efe8-45b2-b491-0c754b34c5d9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:15:03 crc kubenswrapper[4846]: I0202 14:15:03.806736 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06045b74-efe8-45b2-b491-0c754b34c5d9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 14:15:03 crc kubenswrapper[4846]: I0202 14:15:03.806798 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4jzr\" (UniqueName: \"kubernetes.io/projected/06045b74-efe8-45b2-b491-0c754b34c5d9-kube-api-access-g4jzr\") on node \"crc\" DevicePath \"\"" Feb 02 14:15:03 crc kubenswrapper[4846]: I0202 14:15:03.806823 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06045b74-efe8-45b2-b491-0c754b34c5d9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 14:15:04 crc kubenswrapper[4846]: I0202 14:15:04.164869 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" event={"ID":"06045b74-efe8-45b2-b491-0c754b34c5d9","Type":"ContainerDied","Data":"1963fa43337c7968525f9a9d0ebaa01d1c1fae796dfbb8f4734c0fca0bb1a28d"} Feb 02 14:15:04 crc kubenswrapper[4846]: I0202 14:15:04.164929 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1963fa43337c7968525f9a9d0ebaa01d1c1fae796dfbb8f4734c0fca0bb1a28d" Feb 02 14:15:04 crc kubenswrapper[4846]: I0202 14:15:04.164938 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500695-9jrzv" Feb 02 14:15:04 crc kubenswrapper[4846]: I0202 14:15:04.662501 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25"] Feb 02 14:15:04 crc kubenswrapper[4846]: I0202 14:15:04.674935 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500650-2hs25"] Feb 02 14:15:05 crc kubenswrapper[4846]: I0202 14:15:05.447600 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8022e42-1a6e-477b-aefc-7d5c76c26fd2" path="/var/lib/kubelet/pods/c8022e42-1a6e-477b-aefc-7d5c76c26fd2/volumes" Feb 02 14:15:26 crc kubenswrapper[4846]: I0202 14:15:26.412163 4846 generic.go:334] "Generic (PLEG): container finished" podID="1dc40137-f83c-4dd8-9a91-cda128768635" containerID="0bc50639773f3b038aff63ba4de349724d111da68bfa95959340425fcb6ece12" exitCode=0 Feb 02 14:15:26 crc kubenswrapper[4846]: I0202 14:15:26.412247 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-ds2qr" event={"ID":"1dc40137-f83c-4dd8-9a91-cda128768635","Type":"ContainerDied","Data":"0bc50639773f3b038aff63ba4de349724d111da68bfa95959340425fcb6ece12"} Feb 02 14:15:27 crc kubenswrapper[4846]: I0202 14:15:27.929845 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.023957 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-inventory\") pod \"1dc40137-f83c-4dd8-9a91-cda128768635\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.024197 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w67kv\" (UniqueName: \"kubernetes.io/projected/1dc40137-f83c-4dd8-9a91-cda128768635-kube-api-access-w67kv\") pod \"1dc40137-f83c-4dd8-9a91-cda128768635\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.024238 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-ssh-key-openstack-cell1\") pod \"1dc40137-f83c-4dd8-9a91-cda128768635\" (UID: \"1dc40137-f83c-4dd8-9a91-cda128768635\") " Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.029997 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dc40137-f83c-4dd8-9a91-cda128768635-kube-api-access-w67kv" (OuterVolumeSpecName: "kube-api-access-w67kv") pod "1dc40137-f83c-4dd8-9a91-cda128768635" (UID: "1dc40137-f83c-4dd8-9a91-cda128768635"). InnerVolumeSpecName "kube-api-access-w67kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.057437 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-inventory" (OuterVolumeSpecName: "inventory") pod "1dc40137-f83c-4dd8-9a91-cda128768635" (UID: "1dc40137-f83c-4dd8-9a91-cda128768635"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.061202 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "1dc40137-f83c-4dd8-9a91-cda128768635" (UID: "1dc40137-f83c-4dd8-9a91-cda128768635"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.126763 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w67kv\" (UniqueName: \"kubernetes.io/projected/1dc40137-f83c-4dd8-9a91-cda128768635-kube-api-access-w67kv\") on node \"crc\" DevicePath \"\"" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.127008 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.127066 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc40137-f83c-4dd8-9a91-cda128768635-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.444494 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-ds2qr" event={"ID":"1dc40137-f83c-4dd8-9a91-cda128768635","Type":"ContainerDied","Data":"f70c48eab287c9f08c28c13d4b33b26400264242770e206f4190293e9721d3ea"} Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.444540 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f70c48eab287c9f08c28c13d4b33b26400264242770e206f4190293e9721d3ea" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.444986 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-ds2qr" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.518957 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-txvs7"] Feb 02 14:15:28 crc kubenswrapper[4846]: E0202 14:15:28.519371 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06045b74-efe8-45b2-b491-0c754b34c5d9" containerName="collect-profiles" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.519390 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="06045b74-efe8-45b2-b491-0c754b34c5d9" containerName="collect-profiles" Feb 02 14:15:28 crc kubenswrapper[4846]: E0202 14:15:28.519423 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dc40137-f83c-4dd8-9a91-cda128768635" containerName="install-os-openstack-openstack-cell1" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.519429 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc40137-f83c-4dd8-9a91-cda128768635" containerName="install-os-openstack-openstack-cell1" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.519614 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dc40137-f83c-4dd8-9a91-cda128768635" containerName="install-os-openstack-openstack-cell1" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.519653 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="06045b74-efe8-45b2-b491-0c754b34c5d9" containerName="collect-profiles" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.520381 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.522470 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.522952 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.523223 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.525461 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.537847 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-txvs7"] Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.644289 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-txvs7\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.644418 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gzq9\" (UniqueName: \"kubernetes.io/projected/e2556352-fb48-4800-a719-ece237df78f5-kube-api-access-6gzq9\") pod \"configure-os-openstack-openstack-cell1-txvs7\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.644488 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-inventory\") pod \"configure-os-openstack-openstack-cell1-txvs7\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.747345 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-inventory\") pod \"configure-os-openstack-openstack-cell1-txvs7\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.747474 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-txvs7\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.747579 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gzq9\" (UniqueName: \"kubernetes.io/projected/e2556352-fb48-4800-a719-ece237df78f5-kube-api-access-6gzq9\") pod \"configure-os-openstack-openstack-cell1-txvs7\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.751549 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-txvs7\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.755188 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-inventory\") pod \"configure-os-openstack-openstack-cell1-txvs7\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.765695 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gzq9\" (UniqueName: \"kubernetes.io/projected/e2556352-fb48-4800-a719-ece237df78f5-kube-api-access-6gzq9\") pod \"configure-os-openstack-openstack-cell1-txvs7\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:28 crc kubenswrapper[4846]: I0202 14:15:28.838299 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:15:29 crc kubenswrapper[4846]: I0202 14:15:29.444700 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-txvs7"] Feb 02 14:15:30 crc kubenswrapper[4846]: I0202 14:15:30.092244 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:15:30 crc kubenswrapper[4846]: I0202 14:15:30.475856 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-txvs7" event={"ID":"e2556352-fb48-4800-a719-ece237df78f5","Type":"ContainerStarted","Data":"9bf44fd4833245d90ff457ca2627fc2de5a4f8da07e27e5343779571b042d633"} Feb 02 14:15:30 crc kubenswrapper[4846]: I0202 14:15:30.476290 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-txvs7" event={"ID":"e2556352-fb48-4800-a719-ece237df78f5","Type":"ContainerStarted","Data":"991d2babc4386caa17b19cfd1a80dd4be7dd35d1e9cecfc3ad3bb8cb8262c740"} Feb 02 14:15:30 crc kubenswrapper[4846]: I0202 14:15:30.478933 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:15:30 crc kubenswrapper[4846]: I0202 14:15:30.478997 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:15:30 crc kubenswrapper[4846]: I0202 14:15:30.479064 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 14:15:30 crc kubenswrapper[4846]: I0202 14:15:30.480094 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2eb1e85e549a6d9036406934409325d3a7ed0792302a62a04d481a0163f45cb9"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 14:15:30 crc kubenswrapper[4846]: I0202 14:15:30.480161 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://2eb1e85e549a6d9036406934409325d3a7ed0792302a62a04d481a0163f45cb9" gracePeriod=600 Feb 02 14:15:31 crc kubenswrapper[4846]: I0202 14:15:31.488465 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="2eb1e85e549a6d9036406934409325d3a7ed0792302a62a04d481a0163f45cb9" exitCode=0 Feb 02 14:15:31 crc kubenswrapper[4846]: I0202 14:15:31.488522 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"2eb1e85e549a6d9036406934409325d3a7ed0792302a62a04d481a0163f45cb9"} Feb 02 14:15:31 crc kubenswrapper[4846]: I0202 14:15:31.489149 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719"} Feb 02 14:15:31 crc kubenswrapper[4846]: I0202 14:15:31.489185 4846 scope.go:117] "RemoveContainer" containerID="03d6e2b339598c72f2e08c5a53841a271699c26bdccea2cdb163402251731e17" Feb 02 14:15:31 crc kubenswrapper[4846]: I0202 14:15:31.512303 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-txvs7" podStartSLOduration=2.879495306 podStartE2EDuration="3.512282183s" podCreationTimestamp="2026-02-02 14:15:28 +0000 UTC" firstStartedPulling="2026-02-02 14:15:29.456156929 +0000 UTC m=+7560.684743792" lastFinishedPulling="2026-02-02 14:15:30.088943806 +0000 UTC m=+7561.317530669" observedRunningTime="2026-02-02 14:15:31.509952076 +0000 UTC m=+7562.738538959" watchObservedRunningTime="2026-02-02 14:15:31.512282183 +0000 UTC m=+7562.740869056" Feb 02 14:15:43 crc kubenswrapper[4846]: I0202 14:15:43.911375 4846 scope.go:117] "RemoveContainer" containerID="a53234168abee55a3e09d40a916442af33d348439dd91dfbdf9162fc3b4d3dc7" Feb 02 14:16:17 crc kubenswrapper[4846]: I0202 14:16:17.024942 4846 generic.go:334] "Generic (PLEG): container finished" podID="e2556352-fb48-4800-a719-ece237df78f5" containerID="9bf44fd4833245d90ff457ca2627fc2de5a4f8da07e27e5343779571b042d633" exitCode=0 Feb 02 14:16:17 crc kubenswrapper[4846]: I0202 14:16:17.025406 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-txvs7" event={"ID":"e2556352-fb48-4800-a719-ece237df78f5","Type":"ContainerDied","Data":"9bf44fd4833245d90ff457ca2627fc2de5a4f8da07e27e5343779571b042d633"} Feb 02 14:16:18 crc kubenswrapper[4846]: I0202 14:16:18.501697 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:16:18 crc kubenswrapper[4846]: I0202 14:16:18.676049 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-inventory\") pod \"e2556352-fb48-4800-a719-ece237df78f5\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " Feb 02 14:16:18 crc kubenswrapper[4846]: I0202 14:16:18.676223 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gzq9\" (UniqueName: \"kubernetes.io/projected/e2556352-fb48-4800-a719-ece237df78f5-kube-api-access-6gzq9\") pod \"e2556352-fb48-4800-a719-ece237df78f5\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " Feb 02 14:16:18 crc kubenswrapper[4846]: I0202 14:16:18.676454 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-ssh-key-openstack-cell1\") pod \"e2556352-fb48-4800-a719-ece237df78f5\" (UID: \"e2556352-fb48-4800-a719-ece237df78f5\") " Feb 02 14:16:18 crc kubenswrapper[4846]: I0202 14:16:18.685985 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2556352-fb48-4800-a719-ece237df78f5-kube-api-access-6gzq9" (OuterVolumeSpecName: "kube-api-access-6gzq9") pod "e2556352-fb48-4800-a719-ece237df78f5" (UID: "e2556352-fb48-4800-a719-ece237df78f5"). InnerVolumeSpecName "kube-api-access-6gzq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:16:18 crc kubenswrapper[4846]: I0202 14:16:18.708158 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "e2556352-fb48-4800-a719-ece237df78f5" (UID: "e2556352-fb48-4800-a719-ece237df78f5"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:16:18 crc kubenswrapper[4846]: I0202 14:16:18.708274 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-inventory" (OuterVolumeSpecName: "inventory") pod "e2556352-fb48-4800-a719-ece237df78f5" (UID: "e2556352-fb48-4800-a719-ece237df78f5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:16:18 crc kubenswrapper[4846]: I0202 14:16:18.778910 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gzq9\" (UniqueName: \"kubernetes.io/projected/e2556352-fb48-4800-a719-ece237df78f5-kube-api-access-6gzq9\") on node \"crc\" DevicePath \"\"" Feb 02 14:16:18 crc kubenswrapper[4846]: I0202 14:16:18.778960 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:16:18 crc kubenswrapper[4846]: I0202 14:16:18.778971 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2556352-fb48-4800-a719-ece237df78f5-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.044507 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-txvs7" event={"ID":"e2556352-fb48-4800-a719-ece237df78f5","Type":"ContainerDied","Data":"991d2babc4386caa17b19cfd1a80dd4be7dd35d1e9cecfc3ad3bb8cb8262c740"} Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.044550 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="991d2babc4386caa17b19cfd1a80dd4be7dd35d1e9cecfc3ad3bb8cb8262c740" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.044608 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-txvs7" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.125119 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-qd5tq"] Feb 02 14:16:19 crc kubenswrapper[4846]: E0202 14:16:19.125874 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2556352-fb48-4800-a719-ece237df78f5" containerName="configure-os-openstack-openstack-cell1" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.125897 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2556352-fb48-4800-a719-ece237df78f5" containerName="configure-os-openstack-openstack-cell1" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.126077 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2556352-fb48-4800-a719-ece237df78f5" containerName="configure-os-openstack-openstack-cell1" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.126827 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.130053 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.130291 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.130471 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.130699 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.139127 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-qd5tq"] Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.288652 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-inventory-0\") pod \"ssh-known-hosts-openstack-qd5tq\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.288776 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd6kk\" (UniqueName: \"kubernetes.io/projected/408627ad-8ac6-484a-822c-f8b1aec54895-kube-api-access-kd6kk\") pod \"ssh-known-hosts-openstack-qd5tq\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.288876 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-qd5tq\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.391324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-inventory-0\") pod \"ssh-known-hosts-openstack-qd5tq\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.391504 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd6kk\" (UniqueName: \"kubernetes.io/projected/408627ad-8ac6-484a-822c-f8b1aec54895-kube-api-access-kd6kk\") pod \"ssh-known-hosts-openstack-qd5tq\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.391678 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-qd5tq\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.396702 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-inventory-0\") pod \"ssh-known-hosts-openstack-qd5tq\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.396878 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-qd5tq\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.409823 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd6kk\" (UniqueName: \"kubernetes.io/projected/408627ad-8ac6-484a-822c-f8b1aec54895-kube-api-access-kd6kk\") pod \"ssh-known-hosts-openstack-qd5tq\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:19 crc kubenswrapper[4846]: I0202 14:16:19.482442 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:20 crc kubenswrapper[4846]: I0202 14:16:20.058766 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-qd5tq"] Feb 02 14:16:21 crc kubenswrapper[4846]: I0202 14:16:21.068420 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-qd5tq" event={"ID":"408627ad-8ac6-484a-822c-f8b1aec54895","Type":"ContainerStarted","Data":"71ce4f8aa31ed22223902f5ab9116d85e5e8bad023cf6c7b9d10b354252e56ca"} Feb 02 14:16:21 crc kubenswrapper[4846]: I0202 14:16:21.068934 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-qd5tq" event={"ID":"408627ad-8ac6-484a-822c-f8b1aec54895","Type":"ContainerStarted","Data":"499d5699c17769c29071318c51af29f86bce04c8659402ec6b6a09ae2dbea364"} Feb 02 14:16:21 crc kubenswrapper[4846]: I0202 14:16:21.095030 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-qd5tq" podStartSLOduration=1.6612877959999999 podStartE2EDuration="2.095009409s" podCreationTimestamp="2026-02-02 14:16:19 +0000 UTC" firstStartedPulling="2026-02-02 14:16:20.070803966 +0000 UTC m=+7611.299390829" lastFinishedPulling="2026-02-02 14:16:20.504525579 +0000 UTC m=+7611.733112442" observedRunningTime="2026-02-02 14:16:21.086744658 +0000 UTC m=+7612.315331541" watchObservedRunningTime="2026-02-02 14:16:21.095009409 +0000 UTC m=+7612.323596272" Feb 02 14:16:30 crc kubenswrapper[4846]: I0202 14:16:30.162678 4846 generic.go:334] "Generic (PLEG): container finished" podID="408627ad-8ac6-484a-822c-f8b1aec54895" containerID="71ce4f8aa31ed22223902f5ab9116d85e5e8bad023cf6c7b9d10b354252e56ca" exitCode=0 Feb 02 14:16:30 crc kubenswrapper[4846]: I0202 14:16:30.162746 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-qd5tq" event={"ID":"408627ad-8ac6-484a-822c-f8b1aec54895","Type":"ContainerDied","Data":"71ce4f8aa31ed22223902f5ab9116d85e5e8bad023cf6c7b9d10b354252e56ca"} Feb 02 14:16:31 crc kubenswrapper[4846]: I0202 14:16:31.619777 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:31 crc kubenswrapper[4846]: I0202 14:16:31.806506 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-ssh-key-openstack-cell1\") pod \"408627ad-8ac6-484a-822c-f8b1aec54895\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " Feb 02 14:16:31 crc kubenswrapper[4846]: I0202 14:16:31.806761 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kd6kk\" (UniqueName: \"kubernetes.io/projected/408627ad-8ac6-484a-822c-f8b1aec54895-kube-api-access-kd6kk\") pod \"408627ad-8ac6-484a-822c-f8b1aec54895\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " Feb 02 14:16:31 crc kubenswrapper[4846]: I0202 14:16:31.806993 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-inventory-0\") pod \"408627ad-8ac6-484a-822c-f8b1aec54895\" (UID: \"408627ad-8ac6-484a-822c-f8b1aec54895\") " Feb 02 14:16:31 crc kubenswrapper[4846]: I0202 14:16:31.815944 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/408627ad-8ac6-484a-822c-f8b1aec54895-kube-api-access-kd6kk" (OuterVolumeSpecName: "kube-api-access-kd6kk") pod "408627ad-8ac6-484a-822c-f8b1aec54895" (UID: "408627ad-8ac6-484a-822c-f8b1aec54895"). InnerVolumeSpecName "kube-api-access-kd6kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:16:31 crc kubenswrapper[4846]: I0202 14:16:31.837169 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "408627ad-8ac6-484a-822c-f8b1aec54895" (UID: "408627ad-8ac6-484a-822c-f8b1aec54895"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:16:31 crc kubenswrapper[4846]: I0202 14:16:31.844178 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "408627ad-8ac6-484a-822c-f8b1aec54895" (UID: "408627ad-8ac6-484a-822c-f8b1aec54895"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:16:31 crc kubenswrapper[4846]: I0202 14:16:31.910832 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kd6kk\" (UniqueName: \"kubernetes.io/projected/408627ad-8ac6-484a-822c-f8b1aec54895-kube-api-access-kd6kk\") on node \"crc\" DevicePath \"\"" Feb 02 14:16:31 crc kubenswrapper[4846]: I0202 14:16:31.910877 4846 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:16:31 crc kubenswrapper[4846]: I0202 14:16:31.910913 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/408627ad-8ac6-484a-822c-f8b1aec54895-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.181951 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-qd5tq" event={"ID":"408627ad-8ac6-484a-822c-f8b1aec54895","Type":"ContainerDied","Data":"499d5699c17769c29071318c51af29f86bce04c8659402ec6b6a09ae2dbea364"} Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.182136 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="499d5699c17769c29071318c51af29f86bce04c8659402ec6b6a09ae2dbea364" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.182007 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-qd5tq" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.262002 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-j9gn9"] Feb 02 14:16:32 crc kubenswrapper[4846]: E0202 14:16:32.262431 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408627ad-8ac6-484a-822c-f8b1aec54895" containerName="ssh-known-hosts-openstack" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.262447 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="408627ad-8ac6-484a-822c-f8b1aec54895" containerName="ssh-known-hosts-openstack" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.262658 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="408627ad-8ac6-484a-822c-f8b1aec54895" containerName="ssh-known-hosts-openstack" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.263349 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.269685 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.269762 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.269692 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.269902 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.274963 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-j9gn9"] Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.421649 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-j9gn9\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.421729 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-inventory\") pod \"run-os-openstack-openstack-cell1-j9gn9\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.421824 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv88t\" (UniqueName: \"kubernetes.io/projected/acd9db54-1e35-46b2-8e31-be31c11303a8-kube-api-access-kv88t\") pod \"run-os-openstack-openstack-cell1-j9gn9\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.523754 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-inventory\") pod \"run-os-openstack-openstack-cell1-j9gn9\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.524840 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv88t\" (UniqueName: \"kubernetes.io/projected/acd9db54-1e35-46b2-8e31-be31c11303a8-kube-api-access-kv88t\") pod \"run-os-openstack-openstack-cell1-j9gn9\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.525221 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-j9gn9\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.539316 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-inventory\") pod \"run-os-openstack-openstack-cell1-j9gn9\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.539567 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-j9gn9\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.543581 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv88t\" (UniqueName: \"kubernetes.io/projected/acd9db54-1e35-46b2-8e31-be31c11303a8-kube-api-access-kv88t\") pod \"run-os-openstack-openstack-cell1-j9gn9\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:32 crc kubenswrapper[4846]: I0202 14:16:32.581050 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:33 crc kubenswrapper[4846]: I0202 14:16:33.148407 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-j9gn9"] Feb 02 14:16:33 crc kubenswrapper[4846]: I0202 14:16:33.195328 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-j9gn9" event={"ID":"acd9db54-1e35-46b2-8e31-be31c11303a8","Type":"ContainerStarted","Data":"4ac7da05d027785913f6ff9347c4bf6bc428846331b8ff20355e13e6173d1b7c"} Feb 02 14:16:34 crc kubenswrapper[4846]: I0202 14:16:34.209219 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-j9gn9" event={"ID":"acd9db54-1e35-46b2-8e31-be31c11303a8","Type":"ContainerStarted","Data":"936b04be05b0bea6f3bdcc767fed070efc6438710cf82c066b28e71c80fb7c1c"} Feb 02 14:16:34 crc kubenswrapper[4846]: I0202 14:16:34.237950 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-j9gn9" podStartSLOduration=1.8037924539999999 podStartE2EDuration="2.237928777s" podCreationTimestamp="2026-02-02 14:16:32 +0000 UTC" firstStartedPulling="2026-02-02 14:16:33.155238334 +0000 UTC m=+7624.383825187" lastFinishedPulling="2026-02-02 14:16:33.589374647 +0000 UTC m=+7624.817961510" observedRunningTime="2026-02-02 14:16:34.226978861 +0000 UTC m=+7625.455565734" watchObservedRunningTime="2026-02-02 14:16:34.237928777 +0000 UTC m=+7625.466515640" Feb 02 14:16:42 crc kubenswrapper[4846]: I0202 14:16:42.296407 4846 generic.go:334] "Generic (PLEG): container finished" podID="acd9db54-1e35-46b2-8e31-be31c11303a8" containerID="936b04be05b0bea6f3bdcc767fed070efc6438710cf82c066b28e71c80fb7c1c" exitCode=0 Feb 02 14:16:42 crc kubenswrapper[4846]: I0202 14:16:42.296475 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-j9gn9" event={"ID":"acd9db54-1e35-46b2-8e31-be31c11303a8","Type":"ContainerDied","Data":"936b04be05b0bea6f3bdcc767fed070efc6438710cf82c066b28e71c80fb7c1c"} Feb 02 14:16:43 crc kubenswrapper[4846]: I0202 14:16:43.752129 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:43 crc kubenswrapper[4846]: I0202 14:16:43.828114 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv88t\" (UniqueName: \"kubernetes.io/projected/acd9db54-1e35-46b2-8e31-be31c11303a8-kube-api-access-kv88t\") pod \"acd9db54-1e35-46b2-8e31-be31c11303a8\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " Feb 02 14:16:43 crc kubenswrapper[4846]: I0202 14:16:43.828218 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-ssh-key-openstack-cell1\") pod \"acd9db54-1e35-46b2-8e31-be31c11303a8\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " Feb 02 14:16:43 crc kubenswrapper[4846]: I0202 14:16:43.828338 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-inventory\") pod \"acd9db54-1e35-46b2-8e31-be31c11303a8\" (UID: \"acd9db54-1e35-46b2-8e31-be31c11303a8\") " Feb 02 14:16:43 crc kubenswrapper[4846]: I0202 14:16:43.837270 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd9db54-1e35-46b2-8e31-be31c11303a8-kube-api-access-kv88t" (OuterVolumeSpecName: "kube-api-access-kv88t") pod "acd9db54-1e35-46b2-8e31-be31c11303a8" (UID: "acd9db54-1e35-46b2-8e31-be31c11303a8"). InnerVolumeSpecName "kube-api-access-kv88t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:16:43 crc kubenswrapper[4846]: I0202 14:16:43.864353 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "acd9db54-1e35-46b2-8e31-be31c11303a8" (UID: "acd9db54-1e35-46b2-8e31-be31c11303a8"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:16:43 crc kubenswrapper[4846]: I0202 14:16:43.870468 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-inventory" (OuterVolumeSpecName: "inventory") pod "acd9db54-1e35-46b2-8e31-be31c11303a8" (UID: "acd9db54-1e35-46b2-8e31-be31c11303a8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:16:43 crc kubenswrapper[4846]: I0202 14:16:43.932155 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kv88t\" (UniqueName: \"kubernetes.io/projected/acd9db54-1e35-46b2-8e31-be31c11303a8-kube-api-access-kv88t\") on node \"crc\" DevicePath \"\"" Feb 02 14:16:43 crc kubenswrapper[4846]: I0202 14:16:43.932196 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:16:43 crc kubenswrapper[4846]: I0202 14:16:43.932210 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acd9db54-1e35-46b2-8e31-be31c11303a8-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.317605 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-j9gn9" event={"ID":"acd9db54-1e35-46b2-8e31-be31c11303a8","Type":"ContainerDied","Data":"4ac7da05d027785913f6ff9347c4bf6bc428846331b8ff20355e13e6173d1b7c"} Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.318060 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ac7da05d027785913f6ff9347c4bf6bc428846331b8ff20355e13e6173d1b7c" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.319790 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-j9gn9" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.402483 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-dmq2m"] Feb 02 14:16:44 crc kubenswrapper[4846]: E0202 14:16:44.403137 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd9db54-1e35-46b2-8e31-be31c11303a8" containerName="run-os-openstack-openstack-cell1" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.403156 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd9db54-1e35-46b2-8e31-be31c11303a8" containerName="run-os-openstack-openstack-cell1" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.403421 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="acd9db54-1e35-46b2-8e31-be31c11303a8" containerName="run-os-openstack-openstack-cell1" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.404435 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.408246 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.408469 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.409735 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.409880 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.421031 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-dmq2m"] Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.547325 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-inventory\") pod \"reboot-os-openstack-openstack-cell1-dmq2m\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.547477 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq8nm\" (UniqueName: \"kubernetes.io/projected/27ffd62f-5922-47de-8a2a-8374a1540888-kube-api-access-fq8nm\") pod \"reboot-os-openstack-openstack-cell1-dmq2m\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.547704 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-dmq2m\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.649988 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-dmq2m\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.650116 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-inventory\") pod \"reboot-os-openstack-openstack-cell1-dmq2m\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.650193 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq8nm\" (UniqueName: \"kubernetes.io/projected/27ffd62f-5922-47de-8a2a-8374a1540888-kube-api-access-fq8nm\") pod \"reboot-os-openstack-openstack-cell1-dmq2m\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.654726 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-inventory\") pod \"reboot-os-openstack-openstack-cell1-dmq2m\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.655588 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-dmq2m\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.673408 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq8nm\" (UniqueName: \"kubernetes.io/projected/27ffd62f-5922-47de-8a2a-8374a1540888-kube-api-access-fq8nm\") pod \"reboot-os-openstack-openstack-cell1-dmq2m\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:44 crc kubenswrapper[4846]: I0202 14:16:44.733367 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:16:45 crc kubenswrapper[4846]: I0202 14:16:45.345431 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-dmq2m"] Feb 02 14:16:45 crc kubenswrapper[4846]: W0202 14:16:45.362864 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27ffd62f_5922_47de_8a2a_8374a1540888.slice/crio-ca1bdbb46a154c469fb6f8117d0a5ae7c4b2a25bc6b3e61a0014c41ce9760728 WatchSource:0}: Error finding container ca1bdbb46a154c469fb6f8117d0a5ae7c4b2a25bc6b3e61a0014c41ce9760728: Status 404 returned error can't find the container with id ca1bdbb46a154c469fb6f8117d0a5ae7c4b2a25bc6b3e61a0014c41ce9760728 Feb 02 14:16:46 crc kubenswrapper[4846]: I0202 14:16:46.341216 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" event={"ID":"27ffd62f-5922-47de-8a2a-8374a1540888","Type":"ContainerStarted","Data":"d127f2919e1a5d8b649d434ac60597cf8842afca99fad7a102c2852d0702ee29"} Feb 02 14:16:46 crc kubenswrapper[4846]: I0202 14:16:46.342303 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" event={"ID":"27ffd62f-5922-47de-8a2a-8374a1540888","Type":"ContainerStarted","Data":"ca1bdbb46a154c469fb6f8117d0a5ae7c4b2a25bc6b3e61a0014c41ce9760728"} Feb 02 14:16:46 crc kubenswrapper[4846]: I0202 14:16:46.363108 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" podStartSLOduration=1.889950701 podStartE2EDuration="2.363085811s" podCreationTimestamp="2026-02-02 14:16:44 +0000 UTC" firstStartedPulling="2026-02-02 14:16:45.365690179 +0000 UTC m=+7636.594277042" lastFinishedPulling="2026-02-02 14:16:45.838825289 +0000 UTC m=+7637.067412152" observedRunningTime="2026-02-02 14:16:46.361134314 +0000 UTC m=+7637.589721257" watchObservedRunningTime="2026-02-02 14:16:46.363085811 +0000 UTC m=+7637.591672684" Feb 02 14:17:02 crc kubenswrapper[4846]: I0202 14:17:02.528233 4846 generic.go:334] "Generic (PLEG): container finished" podID="27ffd62f-5922-47de-8a2a-8374a1540888" containerID="d127f2919e1a5d8b649d434ac60597cf8842afca99fad7a102c2852d0702ee29" exitCode=0 Feb 02 14:17:02 crc kubenswrapper[4846]: I0202 14:17:02.529120 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" event={"ID":"27ffd62f-5922-47de-8a2a-8374a1540888","Type":"ContainerDied","Data":"d127f2919e1a5d8b649d434ac60597cf8842afca99fad7a102c2852d0702ee29"} Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.009971 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.145470 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq8nm\" (UniqueName: \"kubernetes.io/projected/27ffd62f-5922-47de-8a2a-8374a1540888-kube-api-access-fq8nm\") pod \"27ffd62f-5922-47de-8a2a-8374a1540888\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.145541 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-inventory\") pod \"27ffd62f-5922-47de-8a2a-8374a1540888\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.145816 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-ssh-key-openstack-cell1\") pod \"27ffd62f-5922-47de-8a2a-8374a1540888\" (UID: \"27ffd62f-5922-47de-8a2a-8374a1540888\") " Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.151558 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27ffd62f-5922-47de-8a2a-8374a1540888-kube-api-access-fq8nm" (OuterVolumeSpecName: "kube-api-access-fq8nm") pod "27ffd62f-5922-47de-8a2a-8374a1540888" (UID: "27ffd62f-5922-47de-8a2a-8374a1540888"). InnerVolumeSpecName "kube-api-access-fq8nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.175762 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-inventory" (OuterVolumeSpecName: "inventory") pod "27ffd62f-5922-47de-8a2a-8374a1540888" (UID: "27ffd62f-5922-47de-8a2a-8374a1540888"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.178555 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "27ffd62f-5922-47de-8a2a-8374a1540888" (UID: "27ffd62f-5922-47de-8a2a-8374a1540888"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.248704 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq8nm\" (UniqueName: \"kubernetes.io/projected/27ffd62f-5922-47de-8a2a-8374a1540888-kube-api-access-fq8nm\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.248753 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.248766 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/27ffd62f-5922-47de-8a2a-8374a1540888-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.549732 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" event={"ID":"27ffd62f-5922-47de-8a2a-8374a1540888","Type":"ContainerDied","Data":"ca1bdbb46a154c469fb6f8117d0a5ae7c4b2a25bc6b3e61a0014c41ce9760728"} Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.549780 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca1bdbb46a154c469fb6f8117d0a5ae7c4b2a25bc6b3e61a0014c41ce9760728" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.549885 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-dmq2m" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.677399 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-swnxb"] Feb 02 14:17:04 crc kubenswrapper[4846]: E0202 14:17:04.678150 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ffd62f-5922-47de-8a2a-8374a1540888" containerName="reboot-os-openstack-openstack-cell1" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.678256 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ffd62f-5922-47de-8a2a-8374a1540888" containerName="reboot-os-openstack-openstack-cell1" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.683023 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="27ffd62f-5922-47de-8a2a-8374a1540888" containerName="reboot-os-openstack-openstack-cell1" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.684537 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.688706 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-libvirt-default-certs-0" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.689130 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-telemetry-default-certs-0" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.689224 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.689340 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-ovn-default-certs-0" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.689414 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-swnxb"] Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.689441 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.689580 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.689477 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-neutron-metadata-default-certs-0" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.689543 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.759787 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.759852 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.759896 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.759941 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.759982 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.760004 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.760029 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.760062 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.760127 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-inventory\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.760151 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.760180 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.760336 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.760425 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.760467 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tk9h\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-kube-api-access-4tk9h\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.760514 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862103 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862162 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862192 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862230 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862291 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-inventory\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862391 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862413 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862462 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tk9h\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-kube-api-access-4tk9h\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862522 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862551 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862573 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862596 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.862643 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.869585 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.869823 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.869884 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.869940 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.870420 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.872503 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.873097 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.873602 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.873731 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.873962 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-inventory\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.874609 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.874644 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.874737 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.877726 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:04 crc kubenswrapper[4846]: I0202 14:17:04.882510 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tk9h\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-kube-api-access-4tk9h\") pod \"install-certs-openstack-openstack-cell1-swnxb\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:05 crc kubenswrapper[4846]: I0202 14:17:05.015603 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:05 crc kubenswrapper[4846]: I0202 14:17:05.571426 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-swnxb"] Feb 02 14:17:06 crc kubenswrapper[4846]: I0202 14:17:06.576942 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-swnxb" event={"ID":"392df862-6b24-4249-9a20-52d721c57a73","Type":"ContainerStarted","Data":"924f4b6b57616b89285e0c51688e0c2778df8ed3967887dec464acbb4839a9d8"} Feb 02 14:17:06 crc kubenswrapper[4846]: I0202 14:17:06.577711 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-swnxb" event={"ID":"392df862-6b24-4249-9a20-52d721c57a73","Type":"ContainerStarted","Data":"caaf3f6c2cf24bd57c7a6d33c08d987d3ed5093e89ab105ee4d195d049652c20"} Feb 02 14:17:06 crc kubenswrapper[4846]: I0202 14:17:06.614075 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-swnxb" podStartSLOduration=1.894085806 podStartE2EDuration="2.614053561s" podCreationTimestamp="2026-02-02 14:17:04 +0000 UTC" firstStartedPulling="2026-02-02 14:17:05.575031417 +0000 UTC m=+7656.803618280" lastFinishedPulling="2026-02-02 14:17:06.294999122 +0000 UTC m=+7657.523586035" observedRunningTime="2026-02-02 14:17:06.601506736 +0000 UTC m=+7657.830093689" watchObservedRunningTime="2026-02-02 14:17:06.614053561 +0000 UTC m=+7657.842640424" Feb 02 14:17:30 crc kubenswrapper[4846]: I0202 14:17:30.478980 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:17:30 crc kubenswrapper[4846]: I0202 14:17:30.479505 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:17:42 crc kubenswrapper[4846]: I0202 14:17:42.992408 4846 generic.go:334] "Generic (PLEG): container finished" podID="392df862-6b24-4249-9a20-52d721c57a73" containerID="924f4b6b57616b89285e0c51688e0c2778df8ed3967887dec464acbb4839a9d8" exitCode=0 Feb 02 14:17:42 crc kubenswrapper[4846]: I0202 14:17:42.992515 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-swnxb" event={"ID":"392df862-6b24-4249-9a20-52d721c57a73","Type":"ContainerDied","Data":"924f4b6b57616b89285e0c51688e0c2778df8ed3967887dec464acbb4839a9d8"} Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.578957 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675210 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-metadata-combined-ca-bundle\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675305 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-libvirt-default-certs-0\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675410 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ssh-key-openstack-cell1\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675435 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-inventory\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675474 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-dhcp-combined-ca-bundle\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675514 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-neutron-metadata-default-certs-0\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675540 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-libvirt-combined-ca-bundle\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675562 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-nova-combined-ca-bundle\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675590 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-telemetry-combined-ca-bundle\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675723 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-bootstrap-combined-ca-bundle\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675838 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ovn-combined-ca-bundle\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675871 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-ovn-default-certs-0\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675906 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-sriov-combined-ca-bundle\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675953 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-telemetry-default-certs-0\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.675996 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tk9h\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-kube-api-access-4tk9h\") pod \"392df862-6b24-4249-9a20-52d721c57a73\" (UID: \"392df862-6b24-4249-9a20-52d721c57a73\") " Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.684969 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-kube-api-access-4tk9h" (OuterVolumeSpecName: "kube-api-access-4tk9h") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "kube-api-access-4tk9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.688155 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.688359 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.690461 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.691659 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.691609 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-libvirt-default-certs-0") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "openstack-cell1-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.692492 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-telemetry-default-certs-0") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "openstack-cell1-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.692585 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.694325 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.694385 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.696014 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.696125 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-neutron-metadata-default-certs-0") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "openstack-cell1-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.696175 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-ovn-default-certs-0") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "openstack-cell1-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.727049 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-inventory" (OuterVolumeSpecName: "inventory") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.731556 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "392df862-6b24-4249-9a20-52d721c57a73" (UID: "392df862-6b24-4249-9a20-52d721c57a73"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778734 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778765 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778775 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778785 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778795 4846 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778806 4846 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778813 4846 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778821 4846 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778830 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778840 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778850 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778862 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778871 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tk9h\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-kube-api-access-4tk9h\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778879 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/392df862-6b24-4249-9a20-52d721c57a73-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:44 crc kubenswrapper[4846]: I0202 14:17:44.778888 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/392df862-6b24-4249-9a20-52d721c57a73-openstack-cell1-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.016101 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-swnxb" event={"ID":"392df862-6b24-4249-9a20-52d721c57a73","Type":"ContainerDied","Data":"caaf3f6c2cf24bd57c7a6d33c08d987d3ed5093e89ab105ee4d195d049652c20"} Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.016153 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caaf3f6c2cf24bd57c7a6d33c08d987d3ed5093e89ab105ee4d195d049652c20" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.016206 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-swnxb" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.161173 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-vqz97"] Feb 02 14:17:45 crc kubenswrapper[4846]: E0202 14:17:45.161791 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392df862-6b24-4249-9a20-52d721c57a73" containerName="install-certs-openstack-openstack-cell1" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.161812 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="392df862-6b24-4249-9a20-52d721c57a73" containerName="install-certs-openstack-openstack-cell1" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.162052 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="392df862-6b24-4249-9a20-52d721c57a73" containerName="install-certs-openstack-openstack-cell1" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.162855 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.170463 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.170509 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.170514 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.170510 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.171891 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.175300 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-vqz97"] Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.303753 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.304165 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-inventory\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.304241 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfk46\" (UniqueName: \"kubernetes.io/projected/650a9e6b-b0ff-4129-b83c-b547a072b2af-kube-api-access-vfk46\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.304432 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.304565 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.406756 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.406839 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.406865 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-inventory\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.406940 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfk46\" (UniqueName: \"kubernetes.io/projected/650a9e6b-b0ff-4129-b83c-b547a072b2af-kube-api-access-vfk46\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.407237 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.407713 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.411079 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.424833 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.426990 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-inventory\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.430385 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfk46\" (UniqueName: \"kubernetes.io/projected/650a9e6b-b0ff-4129-b83c-b547a072b2af-kube-api-access-vfk46\") pod \"ovn-openstack-openstack-cell1-vqz97\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:45 crc kubenswrapper[4846]: I0202 14:17:45.482232 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:17:46 crc kubenswrapper[4846]: I0202 14:17:46.087474 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-vqz97"] Feb 02 14:17:47 crc kubenswrapper[4846]: I0202 14:17:47.036757 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-vqz97" event={"ID":"650a9e6b-b0ff-4129-b83c-b547a072b2af","Type":"ContainerStarted","Data":"259fbfaa67d76603e5e69b892e95a118a781c4a9cecd94ad960b72e88d7fa541"} Feb 02 14:17:48 crc kubenswrapper[4846]: I0202 14:17:48.049474 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-vqz97" event={"ID":"650a9e6b-b0ff-4129-b83c-b547a072b2af","Type":"ContainerStarted","Data":"28b8b0541bc113550483cad8b7d23bb11753cf1adc65d9669ceaf5491bb8c611"} Feb 02 14:17:59 crc kubenswrapper[4846]: I0202 14:17:59.636291 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-vqz97" podStartSLOduration=13.901800074 podStartE2EDuration="14.63626866s" podCreationTimestamp="2026-02-02 14:17:45 +0000 UTC" firstStartedPulling="2026-02-02 14:17:46.085916757 +0000 UTC m=+7697.314503620" lastFinishedPulling="2026-02-02 14:17:46.820385343 +0000 UTC m=+7698.048972206" observedRunningTime="2026-02-02 14:17:48.082269339 +0000 UTC m=+7699.310856212" watchObservedRunningTime="2026-02-02 14:17:59.63626866 +0000 UTC m=+7710.864855523" Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.629608 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.051774562s: [/var/lib/containers/storage/overlay/88a4d734ad1efd5afeda48a454aacd13f0b02dda143f0fa4a2971751bf02a0e4/diff ]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: E0202 14:18:01.630804 4846 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.206s" Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.631051 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.112328042s: [/var/lib/containers/storage/overlay/ba5bd3654c58d1b182ef802c452e31141fccf7f71257b3c39b971bf5a5b4c1dd/diff /var/log/pods/openstack-operators_watcher-operator-controller-manager-564965969-8gkkw_8ee9f8c5-8078-4ad1-b7c8-4a8e2dfd4e41/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.631364 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.112633961s: [/var/lib/containers/storage/overlay/f24d76f57282311e94f1be58a4f3c9bce84009aeb49329c80fa3a013fb18e465/diff /var/log/pods/openstack-operators_keystone-operator-controller-manager-84f48565d4-hfvt6_a70f0094-0775-49ea-9002-9f3da3ff87d1/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.638236 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.638498 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.638935 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.120254756s: [/var/lib/containers/storage/overlay/68da2623f1a5379f34fcafb2c5027f859142a0b7da7561c9bc8e6128cf13e99c/diff /var/log/pods/openstack-operators_neutron-operator-controller-manager-585dbc889-hrnrk_f8ebd280-c5fa-4cf3-890b-93a018bcb2d2/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.639079 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.120388579s: [/var/lib/containers/storage/overlay/4878db6dbed29330518fb470538b556d3b40b7dce58fc46499aab83205d0018b/diff /var/log/pods/openstack-operators_ovn-operator-controller-manager-788c46999f-tw44b_2497b300-a910-41c8-b70d-1c9acd871a90/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.639196 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.120493192s: [/var/lib/containers/storage/overlay/39fe7cba942be580d09684458c6a452312aa5b9c5e91be8d78e3adea5252c581/diff /var/log/pods/openstack-operators_barbican-operator-controller-manager-7b6c4d8c5f-8lrz8_c0f1cd1a-da1b-4292-8622-87119d40ec03/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.639295 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.120550673s: [/var/lib/containers/storage/overlay/d15394cf621e70a1e11c5432deac3fd235129751d93104abac4ce5e685c19c4f/diff /var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf948998-nvnpj_2e0bfdb6-f6e5-41c4-ab53-f13863d687b8/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.642200 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.084349413s: [/var/lib/containers/storage/overlay/20f3b7c0059d216d632dc7787130876fa7f0b576284f0014ec17d81969be72d3/diff /var/log/pods/openstack-operators_nova-operator-controller-manager-55bff696bd-89dd7_f4ec43b3-d763-4478-8ef9-0453a2fb730a/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.642346 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.084502336s: [/var/lib/containers/storage/overlay/a9a2f0774a23c6cc5c0ea01d0153acdd1768513d3efe0df6396a3f8cfe84cf52/diff /var/log/pods/openstack-operators_heat-operator-controller-manager-69d6db494d-jscz6_478eaaaa-5884-4b31-a87e-8655bedb96f1/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.642486 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.078246825s: [/var/lib/containers/storage/overlay/c8e0970621774ebc664d8831fcbba6517d4ec9068b7a05f01c00fecaa59a71d5/diff /var/log/pods/openstack-operators_octavia-operator-controller-manager-6687f8d877-vfmzp_8d9a7d64-a757-48b8-bdb9-4ec3e2921321/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.642660 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.078408179s: [/var/lib/containers/storage/overlay/fa8d988aedeab116829ac494ed1bda3e7c5f7a7852db3a45f1ad82536d004fe3/diff /var/log/pods/openstack-operators_ironic-operator-controller-manager-5f4b8bd54d-7lr2n_69f3343d-ca8a-4ff2-bc0c-14121a6938a8/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.642885 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.065096376s: [/var/lib/containers/storage/overlay/ac87c47695e0b3b93e0e32fa9a1b09bc1ff67a761c5fc294c4624e4db14698d7/diff /var/log/pods/openstack-operators_horizon-operator-controller-manager-5fb775575f-cz76f_46d6dfa1-87b3-4e53-bcfd-27c709ad7911/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.643016 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.065203309s: [/var/lib/containers/storage/overlay/54c5e1689610bc761088931bfacf8e65a2902e0588db7a909f96aa3372532206/diff /var/log/pods/openstack-operators_glance-operator-controller-manager-8886f4c47-bklmw_b9c7230d-0e38-4540-9074-09a47500dd40/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.643145 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.065320292s: [/var/lib/containers/storage/overlay/e2645f731f1a9c74ffb1c5c88807aa06d59c0d4516aae3441fd797c2721c9f0e/diff /var/log/pods/openstack-operators_manila-operator-controller-manager-7dd968899f-5vzvg_d6f0f90b-675e-40e9-b541-7ff04333a885/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.643977 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.066136922s: [/var/lib/containers/storage/overlay/9e194efb50fe743670ddf3acec80638bfe2ba8083fc6ed62e20840e1f6a4f8ab/diff /var/log/pods/openstack-operators_cinder-operator-controller-manager-8d874c8fc-hc6td_6ba95849-98ff-4d50-8c41-605aa99c3b2f/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.645022 4846 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.052609042s: [/var/lib/containers/storage/overlay/44b9114e43de547269a764ed3736ee30b7e6a3bc3e3c7ff81f308763eecc7434/diff /var/log/pods/openstack-operators_designate-operator-controller-manager-6d9697b7f4-88lrz_fced078c-1f83-492e-9f21-e9e82c1f9275/manager/0.log]; will not log again for this container unless duration exceeds 2s Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.736110 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rxfrh"] Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.751180 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rxfrh"] Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.751345 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.943694 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-944bq\" (UniqueName: \"kubernetes.io/projected/3581ff65-3d23-4262-bd5e-590545b3047c-kube-api-access-944bq\") pod \"certified-operators-rxfrh\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.943771 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-utilities\") pod \"certified-operators-rxfrh\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:01 crc kubenswrapper[4846]: I0202 14:18:01.943956 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-catalog-content\") pod \"certified-operators-rxfrh\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:02 crc kubenswrapper[4846]: I0202 14:18:02.045753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-catalog-content\") pod \"certified-operators-rxfrh\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:02 crc kubenswrapper[4846]: I0202 14:18:02.045885 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-944bq\" (UniqueName: \"kubernetes.io/projected/3581ff65-3d23-4262-bd5e-590545b3047c-kube-api-access-944bq\") pod \"certified-operators-rxfrh\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:02 crc kubenswrapper[4846]: I0202 14:18:02.045924 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-utilities\") pod \"certified-operators-rxfrh\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:02 crc kubenswrapper[4846]: I0202 14:18:02.046476 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-catalog-content\") pod \"certified-operators-rxfrh\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:02 crc kubenswrapper[4846]: I0202 14:18:02.046516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-utilities\") pod \"certified-operators-rxfrh\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:02 crc kubenswrapper[4846]: I0202 14:18:02.077115 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-944bq\" (UniqueName: \"kubernetes.io/projected/3581ff65-3d23-4262-bd5e-590545b3047c-kube-api-access-944bq\") pod \"certified-operators-rxfrh\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:02 crc kubenswrapper[4846]: I0202 14:18:02.100281 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:02 crc kubenswrapper[4846]: I0202 14:18:02.630358 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rxfrh"] Feb 02 14:18:02 crc kubenswrapper[4846]: I0202 14:18:02.734392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxfrh" event={"ID":"3581ff65-3d23-4262-bd5e-590545b3047c","Type":"ContainerStarted","Data":"fb2812e5b0f3e36ac7cebca443fc3e2f424eb652b87351f2be92b8f4e106e38b"} Feb 02 14:18:03 crc kubenswrapper[4846]: I0202 14:18:03.749432 4846 generic.go:334] "Generic (PLEG): container finished" podID="3581ff65-3d23-4262-bd5e-590545b3047c" containerID="a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0" exitCode=0 Feb 02 14:18:03 crc kubenswrapper[4846]: I0202 14:18:03.749509 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxfrh" event={"ID":"3581ff65-3d23-4262-bd5e-590545b3047c","Type":"ContainerDied","Data":"a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0"} Feb 02 14:18:05 crc kubenswrapper[4846]: I0202 14:18:05.780977 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxfrh" event={"ID":"3581ff65-3d23-4262-bd5e-590545b3047c","Type":"ContainerStarted","Data":"a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75"} Feb 02 14:18:07 crc kubenswrapper[4846]: I0202 14:18:07.801113 4846 generic.go:334] "Generic (PLEG): container finished" podID="3581ff65-3d23-4262-bd5e-590545b3047c" containerID="a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75" exitCode=0 Feb 02 14:18:07 crc kubenswrapper[4846]: I0202 14:18:07.801186 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxfrh" event={"ID":"3581ff65-3d23-4262-bd5e-590545b3047c","Type":"ContainerDied","Data":"a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75"} Feb 02 14:18:07 crc kubenswrapper[4846]: I0202 14:18:07.805572 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 14:18:08 crc kubenswrapper[4846]: I0202 14:18:08.815821 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxfrh" event={"ID":"3581ff65-3d23-4262-bd5e-590545b3047c","Type":"ContainerStarted","Data":"19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8"} Feb 02 14:18:08 crc kubenswrapper[4846]: I0202 14:18:08.863020 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rxfrh" podStartSLOduration=5.221842223 podStartE2EDuration="9.862999055s" podCreationTimestamp="2026-02-02 14:17:59 +0000 UTC" firstStartedPulling="2026-02-02 14:18:03.753209041 +0000 UTC m=+7714.981795904" lastFinishedPulling="2026-02-02 14:18:08.394365863 +0000 UTC m=+7719.622952736" observedRunningTime="2026-02-02 14:18:08.846308429 +0000 UTC m=+7720.074895302" watchObservedRunningTime="2026-02-02 14:18:08.862999055 +0000 UTC m=+7720.091585918" Feb 02 14:18:12 crc kubenswrapper[4846]: I0202 14:18:12.100566 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:12 crc kubenswrapper[4846]: I0202 14:18:12.101221 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:13 crc kubenswrapper[4846]: I0202 14:18:13.152311 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rxfrh" podUID="3581ff65-3d23-4262-bd5e-590545b3047c" containerName="registry-server" probeResult="failure" output=< Feb 02 14:18:13 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:18:13 crc kubenswrapper[4846]: > Feb 02 14:18:22 crc kubenswrapper[4846]: I0202 14:18:22.179874 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:22 crc kubenswrapper[4846]: I0202 14:18:22.245522 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:22 crc kubenswrapper[4846]: I0202 14:18:22.419914 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rxfrh"] Feb 02 14:18:23 crc kubenswrapper[4846]: I0202 14:18:23.992650 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rxfrh" podUID="3581ff65-3d23-4262-bd5e-590545b3047c" containerName="registry-server" containerID="cri-o://19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8" gracePeriod=2 Feb 02 14:18:24 crc kubenswrapper[4846]: I0202 14:18:24.539789 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:24 crc kubenswrapper[4846]: I0202 14:18:24.633431 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-944bq\" (UniqueName: \"kubernetes.io/projected/3581ff65-3d23-4262-bd5e-590545b3047c-kube-api-access-944bq\") pod \"3581ff65-3d23-4262-bd5e-590545b3047c\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " Feb 02 14:18:24 crc kubenswrapper[4846]: I0202 14:18:24.633511 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-catalog-content\") pod \"3581ff65-3d23-4262-bd5e-590545b3047c\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " Feb 02 14:18:24 crc kubenswrapper[4846]: I0202 14:18:24.633604 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-utilities\") pod \"3581ff65-3d23-4262-bd5e-590545b3047c\" (UID: \"3581ff65-3d23-4262-bd5e-590545b3047c\") " Feb 02 14:18:24 crc kubenswrapper[4846]: I0202 14:18:24.634888 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-utilities" (OuterVolumeSpecName: "utilities") pod "3581ff65-3d23-4262-bd5e-590545b3047c" (UID: "3581ff65-3d23-4262-bd5e-590545b3047c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:18:24 crc kubenswrapper[4846]: I0202 14:18:24.660843 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3581ff65-3d23-4262-bd5e-590545b3047c-kube-api-access-944bq" (OuterVolumeSpecName: "kube-api-access-944bq") pod "3581ff65-3d23-4262-bd5e-590545b3047c" (UID: "3581ff65-3d23-4262-bd5e-590545b3047c"). InnerVolumeSpecName "kube-api-access-944bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:18:24 crc kubenswrapper[4846]: I0202 14:18:24.738469 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-944bq\" (UniqueName: \"kubernetes.io/projected/3581ff65-3d23-4262-bd5e-590545b3047c-kube-api-access-944bq\") on node \"crc\" DevicePath \"\"" Feb 02 14:18:24 crc kubenswrapper[4846]: I0202 14:18:24.738508 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:18:24 crc kubenswrapper[4846]: I0202 14:18:24.754004 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3581ff65-3d23-4262-bd5e-590545b3047c" (UID: "3581ff65-3d23-4262-bd5e-590545b3047c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:18:24 crc kubenswrapper[4846]: I0202 14:18:24.840143 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3581ff65-3d23-4262-bd5e-590545b3047c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.005762 4846 generic.go:334] "Generic (PLEG): container finished" podID="3581ff65-3d23-4262-bd5e-590545b3047c" containerID="19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8" exitCode=0 Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.005827 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxfrh" event={"ID":"3581ff65-3d23-4262-bd5e-590545b3047c","Type":"ContainerDied","Data":"19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8"} Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.005856 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxfrh" event={"ID":"3581ff65-3d23-4262-bd5e-590545b3047c","Type":"ContainerDied","Data":"fb2812e5b0f3e36ac7cebca443fc3e2f424eb652b87351f2be92b8f4e106e38b"} Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.005875 4846 scope.go:117] "RemoveContainer" containerID="19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.006055 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxfrh" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.049375 4846 scope.go:117] "RemoveContainer" containerID="a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.062740 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rxfrh"] Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.087061 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rxfrh"] Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.091506 4846 scope.go:117] "RemoveContainer" containerID="a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.141136 4846 scope.go:117] "RemoveContainer" containerID="19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8" Feb 02 14:18:25 crc kubenswrapper[4846]: E0202 14:18:25.141828 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8\": container with ID starting with 19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8 not found: ID does not exist" containerID="19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.142029 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8"} err="failed to get container status \"19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8\": rpc error: code = NotFound desc = could not find container \"19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8\": container with ID starting with 19acbcdc32eebc07d974812b5f3baeab6c915db989a9f234485f812e738275d8 not found: ID does not exist" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.142061 4846 scope.go:117] "RemoveContainer" containerID="a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75" Feb 02 14:18:25 crc kubenswrapper[4846]: E0202 14:18:25.142449 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75\": container with ID starting with a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75 not found: ID does not exist" containerID="a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.142506 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75"} err="failed to get container status \"a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75\": rpc error: code = NotFound desc = could not find container \"a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75\": container with ID starting with a0998476641eb9c3108cebfd512311d2aa5a3b01566b338708b26153cec07c75 not found: ID does not exist" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.142534 4846 scope.go:117] "RemoveContainer" containerID="a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0" Feb 02 14:18:25 crc kubenswrapper[4846]: E0202 14:18:25.143063 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0\": container with ID starting with a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0 not found: ID does not exist" containerID="a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.143119 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0"} err="failed to get container status \"a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0\": rpc error: code = NotFound desc = could not find container \"a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0\": container with ID starting with a5fc0ced9c5721c9a9a76c28d6d7f6cf535b0f5e16f6bc6318131c7c607fcbb0 not found: ID does not exist" Feb 02 14:18:25 crc kubenswrapper[4846]: I0202 14:18:25.444769 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3581ff65-3d23-4262-bd5e-590545b3047c" path="/var/lib/kubelet/pods/3581ff65-3d23-4262-bd5e-590545b3047c/volumes" Feb 02 14:18:30 crc kubenswrapper[4846]: I0202 14:18:30.479453 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:18:30 crc kubenswrapper[4846]: I0202 14:18:30.479950 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:18:30 crc kubenswrapper[4846]: I0202 14:18:30.479992 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 14:18:30 crc kubenswrapper[4846]: I0202 14:18:30.480574 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 14:18:30 crc kubenswrapper[4846]: I0202 14:18:30.480645 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" gracePeriod=600 Feb 02 14:18:30 crc kubenswrapper[4846]: E0202 14:18:30.607822 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:18:31 crc kubenswrapper[4846]: I0202 14:18:31.072989 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" exitCode=0 Feb 02 14:18:31 crc kubenswrapper[4846]: I0202 14:18:31.073039 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719"} Feb 02 14:18:31 crc kubenswrapper[4846]: I0202 14:18:31.073077 4846 scope.go:117] "RemoveContainer" containerID="2eb1e85e549a6d9036406934409325d3a7ed0792302a62a04d481a0163f45cb9" Feb 02 14:18:31 crc kubenswrapper[4846]: I0202 14:18:31.073867 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:18:31 crc kubenswrapper[4846]: E0202 14:18:31.074215 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:18:44 crc kubenswrapper[4846]: I0202 14:18:44.424091 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:18:44 crc kubenswrapper[4846]: E0202 14:18:44.424900 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:18:51 crc kubenswrapper[4846]: I0202 14:18:51.315062 4846 generic.go:334] "Generic (PLEG): container finished" podID="650a9e6b-b0ff-4129-b83c-b547a072b2af" containerID="28b8b0541bc113550483cad8b7d23bb11753cf1adc65d9669ceaf5491bb8c611" exitCode=0 Feb 02 14:18:51 crc kubenswrapper[4846]: I0202 14:18:51.315166 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-vqz97" event={"ID":"650a9e6b-b0ff-4129-b83c-b547a072b2af","Type":"ContainerDied","Data":"28b8b0541bc113550483cad8b7d23bb11753cf1adc65d9669ceaf5491bb8c611"} Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.826079 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.934968 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-inventory\") pod \"650a9e6b-b0ff-4129-b83c-b547a072b2af\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.935040 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ssh-key-openstack-cell1\") pod \"650a9e6b-b0ff-4129-b83c-b547a072b2af\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.935202 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovn-combined-ca-bundle\") pod \"650a9e6b-b0ff-4129-b83c-b547a072b2af\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.935345 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfk46\" (UniqueName: \"kubernetes.io/projected/650a9e6b-b0ff-4129-b83c-b547a072b2af-kube-api-access-vfk46\") pod \"650a9e6b-b0ff-4129-b83c-b547a072b2af\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.935391 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovncontroller-config-0\") pod \"650a9e6b-b0ff-4129-b83c-b547a072b2af\" (UID: \"650a9e6b-b0ff-4129-b83c-b547a072b2af\") " Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.954799 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "650a9e6b-b0ff-4129-b83c-b547a072b2af" (UID: "650a9e6b-b0ff-4129-b83c-b547a072b2af"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.956166 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/650a9e6b-b0ff-4129-b83c-b547a072b2af-kube-api-access-vfk46" (OuterVolumeSpecName: "kube-api-access-vfk46") pod "650a9e6b-b0ff-4129-b83c-b547a072b2af" (UID: "650a9e6b-b0ff-4129-b83c-b547a072b2af"). InnerVolumeSpecName "kube-api-access-vfk46". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.966257 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "650a9e6b-b0ff-4129-b83c-b547a072b2af" (UID: "650a9e6b-b0ff-4129-b83c-b547a072b2af"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.974012 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-inventory" (OuterVolumeSpecName: "inventory") pod "650a9e6b-b0ff-4129-b83c-b547a072b2af" (UID: "650a9e6b-b0ff-4129-b83c-b547a072b2af"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:18:52 crc kubenswrapper[4846]: I0202 14:18:52.981293 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "650a9e6b-b0ff-4129-b83c-b547a072b2af" (UID: "650a9e6b-b0ff-4129-b83c-b547a072b2af"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.037952 4846 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.037989 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.038005 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.038018 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650a9e6b-b0ff-4129-b83c-b547a072b2af-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.038028 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfk46\" (UniqueName: \"kubernetes.io/projected/650a9e6b-b0ff-4129-b83c-b547a072b2af-kube-api-access-vfk46\") on node \"crc\" DevicePath \"\"" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.339670 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-vqz97" event={"ID":"650a9e6b-b0ff-4129-b83c-b547a072b2af","Type":"ContainerDied","Data":"259fbfaa67d76603e5e69b892e95a118a781c4a9cecd94ad960b72e88d7fa541"} Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.340013 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="259fbfaa67d76603e5e69b892e95a118a781c4a9cecd94ad960b72e88d7fa541" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.339704 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-vqz97" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.554715 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-t7g6s"] Feb 02 14:18:53 crc kubenswrapper[4846]: E0202 14:18:53.555152 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3581ff65-3d23-4262-bd5e-590545b3047c" containerName="registry-server" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.555170 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3581ff65-3d23-4262-bd5e-590545b3047c" containerName="registry-server" Feb 02 14:18:53 crc kubenswrapper[4846]: E0202 14:18:53.555199 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="650a9e6b-b0ff-4129-b83c-b547a072b2af" containerName="ovn-openstack-openstack-cell1" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.555206 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="650a9e6b-b0ff-4129-b83c-b547a072b2af" containerName="ovn-openstack-openstack-cell1" Feb 02 14:18:53 crc kubenswrapper[4846]: E0202 14:18:53.555233 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3581ff65-3d23-4262-bd5e-590545b3047c" containerName="extract-content" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.555239 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3581ff65-3d23-4262-bd5e-590545b3047c" containerName="extract-content" Feb 02 14:18:53 crc kubenswrapper[4846]: E0202 14:18:53.555252 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3581ff65-3d23-4262-bd5e-590545b3047c" containerName="extract-utilities" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.555258 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3581ff65-3d23-4262-bd5e-590545b3047c" containerName="extract-utilities" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.555442 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3581ff65-3d23-4262-bd5e-590545b3047c" containerName="registry-server" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.555464 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="650a9e6b-b0ff-4129-b83c-b547a072b2af" containerName="ovn-openstack-openstack-cell1" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.556171 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.560451 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.560903 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.561075 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.561238 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.561379 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.563180 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.609446 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-t7g6s"] Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.679193 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.679258 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.679334 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.679361 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.679586 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9xxn\" (UniqueName: \"kubernetes.io/projected/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-kube-api-access-r9xxn\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.679656 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.781490 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9xxn\" (UniqueName: \"kubernetes.io/projected/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-kube-api-access-r9xxn\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.781550 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.781577 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.781600 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.781693 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.781726 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.786788 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.787010 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.787311 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.793572 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.803475 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.809833 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9xxn\" (UniqueName: \"kubernetes.io/projected/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-kube-api-access-r9xxn\") pod \"neutron-metadata-openstack-openstack-cell1-t7g6s\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:53 crc kubenswrapper[4846]: I0202 14:18:53.942143 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:18:54 crc kubenswrapper[4846]: I0202 14:18:54.615779 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-t7g6s"] Feb 02 14:18:55 crc kubenswrapper[4846]: I0202 14:18:55.362741 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" event={"ID":"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9","Type":"ContainerStarted","Data":"bc4cace910050d506d2aff498e679cd8bd2c60bbfa2fe1240107170bed65b795"} Feb 02 14:18:56 crc kubenswrapper[4846]: I0202 14:18:56.375426 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" event={"ID":"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9","Type":"ContainerStarted","Data":"e279d1d6508a820ffeb28e5936bd970b78c23169a453bcd3326390cbd70a3370"} Feb 02 14:18:56 crc kubenswrapper[4846]: I0202 14:18:56.412792 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" podStartSLOduration=2.687027176 podStartE2EDuration="3.412766431s" podCreationTimestamp="2026-02-02 14:18:53 +0000 UTC" firstStartedPulling="2026-02-02 14:18:54.636396412 +0000 UTC m=+7765.864983275" lastFinishedPulling="2026-02-02 14:18:55.362135667 +0000 UTC m=+7766.590722530" observedRunningTime="2026-02-02 14:18:56.396575539 +0000 UTC m=+7767.625162422" watchObservedRunningTime="2026-02-02 14:18:56.412766431 +0000 UTC m=+7767.641353334" Feb 02 14:18:56 crc kubenswrapper[4846]: I0202 14:18:56.424591 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:18:56 crc kubenswrapper[4846]: E0202 14:18:56.424832 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:19:09 crc kubenswrapper[4846]: I0202 14:19:09.435894 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:19:09 crc kubenswrapper[4846]: E0202 14:19:09.436860 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:19:23 crc kubenswrapper[4846]: I0202 14:19:23.423989 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:19:23 crc kubenswrapper[4846]: E0202 14:19:23.425000 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:19:36 crc kubenswrapper[4846]: I0202 14:19:36.424467 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:19:36 crc kubenswrapper[4846]: E0202 14:19:36.425402 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:19:45 crc kubenswrapper[4846]: I0202 14:19:45.960373 4846 generic.go:334] "Generic (PLEG): container finished" podID="fac8ff80-70d1-4d0c-974f-011c0ccb6bf9" containerID="e279d1d6508a820ffeb28e5936bd970b78c23169a453bcd3326390cbd70a3370" exitCode=0 Feb 02 14:19:45 crc kubenswrapper[4846]: I0202 14:19:45.960464 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" event={"ID":"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9","Type":"ContainerDied","Data":"e279d1d6508a820ffeb28e5936bd970b78c23169a453bcd3326390cbd70a3370"} Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.427189 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:19:47 crc kubenswrapper[4846]: E0202 14:19:47.427987 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.490551 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.569313 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9xxn\" (UniqueName: \"kubernetes.io/projected/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-kube-api-access-r9xxn\") pod \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.569380 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-inventory\") pod \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.569421 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.569561 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-nova-metadata-neutron-config-0\") pod \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.569655 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-metadata-combined-ca-bundle\") pod \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.569688 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-ssh-key-openstack-cell1\") pod \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\" (UID: \"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9\") " Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.576255 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-kube-api-access-r9xxn" (OuterVolumeSpecName: "kube-api-access-r9xxn") pod "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9" (UID: "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9"). InnerVolumeSpecName "kube-api-access-r9xxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.577897 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9" (UID: "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.599527 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-inventory" (OuterVolumeSpecName: "inventory") pod "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9" (UID: "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.614613 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9" (UID: "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.619586 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9" (UID: "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.635395 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9" (UID: "fac8ff80-70d1-4d0c-974f-011c0ccb6bf9"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.672642 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.672685 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.672701 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.672716 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9xxn\" (UniqueName: \"kubernetes.io/projected/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-kube-api-access-r9xxn\") on node \"crc\" DevicePath \"\"" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.672729 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.672742 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/fac8ff80-70d1-4d0c-974f-011c0ccb6bf9-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.995916 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" event={"ID":"fac8ff80-70d1-4d0c-974f-011c0ccb6bf9","Type":"ContainerDied","Data":"bc4cace910050d506d2aff498e679cd8bd2c60bbfa2fe1240107170bed65b795"} Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.995996 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-t7g6s" Feb 02 14:19:47 crc kubenswrapper[4846]: I0202 14:19:47.996027 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc4cace910050d506d2aff498e679cd8bd2c60bbfa2fe1240107170bed65b795" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.211985 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-qqkcn"] Feb 02 14:19:48 crc kubenswrapper[4846]: E0202 14:19:48.213968 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac8ff80-70d1-4d0c-974f-011c0ccb6bf9" containerName="neutron-metadata-openstack-openstack-cell1" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.214116 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac8ff80-70d1-4d0c-974f-011c0ccb6bf9" containerName="neutron-metadata-openstack-openstack-cell1" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.214743 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac8ff80-70d1-4d0c-974f-011c0ccb6bf9" containerName="neutron-metadata-openstack-openstack-cell1" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.216302 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.220872 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.221154 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.221875 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.222095 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.225166 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.249598 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-qqkcn"] Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.294509 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.295172 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bws4f\" (UniqueName: \"kubernetes.io/projected/8042645e-7a96-4a10-ab69-ba61f5280a61-kube-api-access-bws4f\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.295535 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.295677 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-inventory\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.295826 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.397719 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.397999 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bws4f\" (UniqueName: \"kubernetes.io/projected/8042645e-7a96-4a10-ab69-ba61f5280a61-kube-api-access-bws4f\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.398137 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.398228 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-inventory\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.398335 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.407369 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-inventory\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.407366 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.407441 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.411302 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.420667 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bws4f\" (UniqueName: \"kubernetes.io/projected/8042645e-7a96-4a10-ab69-ba61f5280a61-kube-api-access-bws4f\") pod \"libvirt-openstack-openstack-cell1-qqkcn\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.553090 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:19:48 crc kubenswrapper[4846]: I0202 14:19:48.968153 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-qqkcn"] Feb 02 14:19:48 crc kubenswrapper[4846]: W0202 14:19:48.970415 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8042645e_7a96_4a10_ab69_ba61f5280a61.slice/crio-a5145a2e272f74454efcb0e8dfa45da3e2f64aa2419972a4d516511c38e3b67e WatchSource:0}: Error finding container a5145a2e272f74454efcb0e8dfa45da3e2f64aa2419972a4d516511c38e3b67e: Status 404 returned error can't find the container with id a5145a2e272f74454efcb0e8dfa45da3e2f64aa2419972a4d516511c38e3b67e Feb 02 14:19:49 crc kubenswrapper[4846]: I0202 14:19:49.008454 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" event={"ID":"8042645e-7a96-4a10-ab69-ba61f5280a61","Type":"ContainerStarted","Data":"a5145a2e272f74454efcb0e8dfa45da3e2f64aa2419972a4d516511c38e3b67e"} Feb 02 14:19:50 crc kubenswrapper[4846]: I0202 14:19:50.020703 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" event={"ID":"8042645e-7a96-4a10-ab69-ba61f5280a61","Type":"ContainerStarted","Data":"92a5c4243ceced6961c60ebe767bbac32f31618c3427e4250d9f728096f55b03"} Feb 02 14:19:50 crc kubenswrapper[4846]: I0202 14:19:50.046243 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" podStartSLOduration=1.462565154 podStartE2EDuration="2.046218937s" podCreationTimestamp="2026-02-02 14:19:48 +0000 UTC" firstStartedPulling="2026-02-02 14:19:48.974510991 +0000 UTC m=+7820.203097874" lastFinishedPulling="2026-02-02 14:19:49.558164794 +0000 UTC m=+7820.786751657" observedRunningTime="2026-02-02 14:19:50.045372066 +0000 UTC m=+7821.273958969" watchObservedRunningTime="2026-02-02 14:19:50.046218937 +0000 UTC m=+7821.274805800" Feb 02 14:19:59 crc kubenswrapper[4846]: I0202 14:19:59.440500 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:19:59 crc kubenswrapper[4846]: E0202 14:19:59.441705 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.083539 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5tjzb"] Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.089669 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.137035 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5tjzb"] Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.176340 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-catalog-content\") pod \"redhat-operators-5tjzb\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.176431 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsk7m\" (UniqueName: \"kubernetes.io/projected/49bef830-de5d-4257-88b7-b4f0f3adaa39-kube-api-access-nsk7m\") pod \"redhat-operators-5tjzb\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.176532 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-utilities\") pod \"redhat-operators-5tjzb\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.278204 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-catalog-content\") pod \"redhat-operators-5tjzb\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.278262 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsk7m\" (UniqueName: \"kubernetes.io/projected/49bef830-de5d-4257-88b7-b4f0f3adaa39-kube-api-access-nsk7m\") pod \"redhat-operators-5tjzb\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.278325 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-utilities\") pod \"redhat-operators-5tjzb\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.278865 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-utilities\") pod \"redhat-operators-5tjzb\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.279110 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-catalog-content\") pod \"redhat-operators-5tjzb\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.307472 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsk7m\" (UniqueName: \"kubernetes.io/projected/49bef830-de5d-4257-88b7-b4f0f3adaa39-kube-api-access-nsk7m\") pod \"redhat-operators-5tjzb\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.450644 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:05 crc kubenswrapper[4846]: I0202 14:20:05.947399 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5tjzb"] Feb 02 14:20:06 crc kubenswrapper[4846]: I0202 14:20:06.225054 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tjzb" event={"ID":"49bef830-de5d-4257-88b7-b4f0f3adaa39","Type":"ContainerStarted","Data":"ea262b4a04692eab2a8402637823a946a8d386359c5435aade43176289378ffd"} Feb 02 14:20:06 crc kubenswrapper[4846]: I0202 14:20:06.225103 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tjzb" event={"ID":"49bef830-de5d-4257-88b7-b4f0f3adaa39","Type":"ContainerStarted","Data":"90712881616cb2ebef188659280815a1e97e163c26b606994264af3f7dd34c83"} Feb 02 14:20:07 crc kubenswrapper[4846]: I0202 14:20:07.243490 4846 generic.go:334] "Generic (PLEG): container finished" podID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerID="ea262b4a04692eab2a8402637823a946a8d386359c5435aade43176289378ffd" exitCode=0 Feb 02 14:20:07 crc kubenswrapper[4846]: I0202 14:20:07.243601 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tjzb" event={"ID":"49bef830-de5d-4257-88b7-b4f0f3adaa39","Type":"ContainerDied","Data":"ea262b4a04692eab2a8402637823a946a8d386359c5435aade43176289378ffd"} Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.088983 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gvd8q"] Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.091611 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.110448 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gvd8q"] Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.254303 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-catalog-content\") pod \"community-operators-gvd8q\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.254878 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpfzw\" (UniqueName: \"kubernetes.io/projected/4f7d3631-938a-4757-b83b-b9d80f3037e3-kube-api-access-hpfzw\") pod \"community-operators-gvd8q\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.254960 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-utilities\") pod \"community-operators-gvd8q\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.372029 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpfzw\" (UniqueName: \"kubernetes.io/projected/4f7d3631-938a-4757-b83b-b9d80f3037e3-kube-api-access-hpfzw\") pod \"community-operators-gvd8q\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.372119 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-utilities\") pod \"community-operators-gvd8q\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.372187 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-catalog-content\") pod \"community-operators-gvd8q\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.372871 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-catalog-content\") pod \"community-operators-gvd8q\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.374056 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-utilities\") pod \"community-operators-gvd8q\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.397562 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpfzw\" (UniqueName: \"kubernetes.io/projected/4f7d3631-938a-4757-b83b-b9d80f3037e3-kube-api-access-hpfzw\") pod \"community-operators-gvd8q\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:08 crc kubenswrapper[4846]: I0202 14:20:08.431582 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:09 crc kubenswrapper[4846]: I0202 14:20:09.270879 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tjzb" event={"ID":"49bef830-de5d-4257-88b7-b4f0f3adaa39","Type":"ContainerStarted","Data":"366ad8bedd54d9b6694d2f8913b6c4ef2e88d65c85ce1437050dab2dc2374937"} Feb 02 14:20:09 crc kubenswrapper[4846]: I0202 14:20:09.604997 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gvd8q"] Feb 02 14:20:10 crc kubenswrapper[4846]: I0202 14:20:10.283837 4846 generic.go:334] "Generic (PLEG): container finished" podID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerID="5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862" exitCode=0 Feb 02 14:20:10 crc kubenswrapper[4846]: I0202 14:20:10.283905 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvd8q" event={"ID":"4f7d3631-938a-4757-b83b-b9d80f3037e3","Type":"ContainerDied","Data":"5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862"} Feb 02 14:20:10 crc kubenswrapper[4846]: I0202 14:20:10.284474 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvd8q" event={"ID":"4f7d3631-938a-4757-b83b-b9d80f3037e3","Type":"ContainerStarted","Data":"5065f6d89f747d0b689bc4c8a64ce591bebac4a9dd57315d6fb5ad41ae8ca7d4"} Feb 02 14:20:12 crc kubenswrapper[4846]: I0202 14:20:12.312572 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvd8q" event={"ID":"4f7d3631-938a-4757-b83b-b9d80f3037e3","Type":"ContainerStarted","Data":"312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e"} Feb 02 14:20:13 crc kubenswrapper[4846]: I0202 14:20:13.424832 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:20:13 crc kubenswrapper[4846]: E0202 14:20:13.426261 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:20:14 crc kubenswrapper[4846]: I0202 14:20:14.365197 4846 generic.go:334] "Generic (PLEG): container finished" podID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerID="366ad8bedd54d9b6694d2f8913b6c4ef2e88d65c85ce1437050dab2dc2374937" exitCode=0 Feb 02 14:20:14 crc kubenswrapper[4846]: I0202 14:20:14.365297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tjzb" event={"ID":"49bef830-de5d-4257-88b7-b4f0f3adaa39","Type":"ContainerDied","Data":"366ad8bedd54d9b6694d2f8913b6c4ef2e88d65c85ce1437050dab2dc2374937"} Feb 02 14:20:15 crc kubenswrapper[4846]: I0202 14:20:15.381119 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tjzb" event={"ID":"49bef830-de5d-4257-88b7-b4f0f3adaa39","Type":"ContainerStarted","Data":"44df8824c18d04591792085a0544f96ac39be468f776ebf35d0821bc6242aa15"} Feb 02 14:20:15 crc kubenswrapper[4846]: I0202 14:20:15.383774 4846 generic.go:334] "Generic (PLEG): container finished" podID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerID="312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e" exitCode=0 Feb 02 14:20:15 crc kubenswrapper[4846]: I0202 14:20:15.383882 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvd8q" event={"ID":"4f7d3631-938a-4757-b83b-b9d80f3037e3","Type":"ContainerDied","Data":"312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e"} Feb 02 14:20:15 crc kubenswrapper[4846]: I0202 14:20:15.407711 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5tjzb" podStartSLOduration=2.86378893 podStartE2EDuration="10.407681656s" podCreationTimestamp="2026-02-02 14:20:05 +0000 UTC" firstStartedPulling="2026-02-02 14:20:07.246612951 +0000 UTC m=+7838.475199824" lastFinishedPulling="2026-02-02 14:20:14.790505647 +0000 UTC m=+7846.019092550" observedRunningTime="2026-02-02 14:20:15.399772343 +0000 UTC m=+7846.628359246" watchObservedRunningTime="2026-02-02 14:20:15.407681656 +0000 UTC m=+7846.636268519" Feb 02 14:20:15 crc kubenswrapper[4846]: I0202 14:20:15.451378 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:15 crc kubenswrapper[4846]: I0202 14:20:15.451436 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:16 crc kubenswrapper[4846]: I0202 14:20:16.396824 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvd8q" event={"ID":"4f7d3631-938a-4757-b83b-b9d80f3037e3","Type":"ContainerStarted","Data":"80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7"} Feb 02 14:20:16 crc kubenswrapper[4846]: I0202 14:20:16.423529 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gvd8q" podStartSLOduration=2.882873399 podStartE2EDuration="8.423509055s" podCreationTimestamp="2026-02-02 14:20:08 +0000 UTC" firstStartedPulling="2026-02-02 14:20:10.286496555 +0000 UTC m=+7841.515083438" lastFinishedPulling="2026-02-02 14:20:15.827132211 +0000 UTC m=+7847.055719094" observedRunningTime="2026-02-02 14:20:16.413540513 +0000 UTC m=+7847.642127386" watchObservedRunningTime="2026-02-02 14:20:16.423509055 +0000 UTC m=+7847.652095918" Feb 02 14:20:16 crc kubenswrapper[4846]: I0202 14:20:16.506952 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5tjzb" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="registry-server" probeResult="failure" output=< Feb 02 14:20:16 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:20:16 crc kubenswrapper[4846]: > Feb 02 14:20:18 crc kubenswrapper[4846]: I0202 14:20:18.431822 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:18 crc kubenswrapper[4846]: I0202 14:20:18.432182 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:19 crc kubenswrapper[4846]: I0202 14:20:19.486950 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gvd8q" podUID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerName="registry-server" probeResult="failure" output=< Feb 02 14:20:19 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:20:19 crc kubenswrapper[4846]: > Feb 02 14:20:24 crc kubenswrapper[4846]: I0202 14:20:24.424235 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:20:24 crc kubenswrapper[4846]: E0202 14:20:24.425207 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:20:26 crc kubenswrapper[4846]: I0202 14:20:26.505102 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5tjzb" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="registry-server" probeResult="failure" output=< Feb 02 14:20:26 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:20:26 crc kubenswrapper[4846]: > Feb 02 14:20:28 crc kubenswrapper[4846]: I0202 14:20:28.486854 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:28 crc kubenswrapper[4846]: I0202 14:20:28.567357 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:28 crc kubenswrapper[4846]: I0202 14:20:28.747603 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gvd8q"] Feb 02 14:20:29 crc kubenswrapper[4846]: I0202 14:20:29.549947 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gvd8q" podUID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerName="registry-server" containerID="cri-o://80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7" gracePeriod=2 Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.073854 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.161498 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpfzw\" (UniqueName: \"kubernetes.io/projected/4f7d3631-938a-4757-b83b-b9d80f3037e3-kube-api-access-hpfzw\") pod \"4f7d3631-938a-4757-b83b-b9d80f3037e3\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.161572 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-catalog-content\") pod \"4f7d3631-938a-4757-b83b-b9d80f3037e3\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.161757 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-utilities\") pod \"4f7d3631-938a-4757-b83b-b9d80f3037e3\" (UID: \"4f7d3631-938a-4757-b83b-b9d80f3037e3\") " Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.163074 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-utilities" (OuterVolumeSpecName: "utilities") pod "4f7d3631-938a-4757-b83b-b9d80f3037e3" (UID: "4f7d3631-938a-4757-b83b-b9d80f3037e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.170921 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7d3631-938a-4757-b83b-b9d80f3037e3-kube-api-access-hpfzw" (OuterVolumeSpecName: "kube-api-access-hpfzw") pod "4f7d3631-938a-4757-b83b-b9d80f3037e3" (UID: "4f7d3631-938a-4757-b83b-b9d80f3037e3"). InnerVolumeSpecName "kube-api-access-hpfzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.223249 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f7d3631-938a-4757-b83b-b9d80f3037e3" (UID: "4f7d3631-938a-4757-b83b-b9d80f3037e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.265509 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpfzw\" (UniqueName: \"kubernetes.io/projected/4f7d3631-938a-4757-b83b-b9d80f3037e3-kube-api-access-hpfzw\") on node \"crc\" DevicePath \"\"" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.265552 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.265565 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f7d3631-938a-4757-b83b-b9d80f3037e3-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.561272 4846 generic.go:334] "Generic (PLEG): container finished" podID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerID="80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7" exitCode=0 Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.561321 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvd8q" event={"ID":"4f7d3631-938a-4757-b83b-b9d80f3037e3","Type":"ContainerDied","Data":"80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7"} Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.561350 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gvd8q" event={"ID":"4f7d3631-938a-4757-b83b-b9d80f3037e3","Type":"ContainerDied","Data":"5065f6d89f747d0b689bc4c8a64ce591bebac4a9dd57315d6fb5ad41ae8ca7d4"} Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.561370 4846 scope.go:117] "RemoveContainer" containerID="80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.561525 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gvd8q" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.585727 4846 scope.go:117] "RemoveContainer" containerID="312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.602332 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gvd8q"] Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.611535 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gvd8q"] Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.628168 4846 scope.go:117] "RemoveContainer" containerID="5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.686234 4846 scope.go:117] "RemoveContainer" containerID="80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7" Feb 02 14:20:30 crc kubenswrapper[4846]: E0202 14:20:30.686695 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7\": container with ID starting with 80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7 not found: ID does not exist" containerID="80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.686747 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7"} err="failed to get container status \"80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7\": rpc error: code = NotFound desc = could not find container \"80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7\": container with ID starting with 80e0f8968ee84ff63b6a12e3cc0b23108b2cad2ff63eff95bdbb0dfd0f3c98f7 not found: ID does not exist" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.686780 4846 scope.go:117] "RemoveContainer" containerID="312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e" Feb 02 14:20:30 crc kubenswrapper[4846]: E0202 14:20:30.687105 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e\": container with ID starting with 312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e not found: ID does not exist" containerID="312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.687150 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e"} err="failed to get container status \"312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e\": rpc error: code = NotFound desc = could not find container \"312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e\": container with ID starting with 312d360fd792b311704f7b5037eafdc85f272862873e0c7313ccb048b51b462e not found: ID does not exist" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.687180 4846 scope.go:117] "RemoveContainer" containerID="5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862" Feb 02 14:20:30 crc kubenswrapper[4846]: E0202 14:20:30.687562 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862\": container with ID starting with 5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862 not found: ID does not exist" containerID="5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862" Feb 02 14:20:30 crc kubenswrapper[4846]: I0202 14:20:30.687612 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862"} err="failed to get container status \"5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862\": rpc error: code = NotFound desc = could not find container \"5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862\": container with ID starting with 5980effbf751e9cd0723ce13bc8b8efe64e7b537bb8a21d25a11975eca72b862 not found: ID does not exist" Feb 02 14:20:31 crc kubenswrapper[4846]: I0202 14:20:31.441025 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f7d3631-938a-4757-b83b-b9d80f3037e3" path="/var/lib/kubelet/pods/4f7d3631-938a-4757-b83b-b9d80f3037e3/volumes" Feb 02 14:20:36 crc kubenswrapper[4846]: I0202 14:20:36.500756 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5tjzb" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="registry-server" probeResult="failure" output=< Feb 02 14:20:36 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:20:36 crc kubenswrapper[4846]: > Feb 02 14:20:39 crc kubenswrapper[4846]: I0202 14:20:39.442055 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:20:39 crc kubenswrapper[4846]: E0202 14:20:39.442522 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:20:45 crc kubenswrapper[4846]: I0202 14:20:45.517991 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:45 crc kubenswrapper[4846]: I0202 14:20:45.575933 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:45 crc kubenswrapper[4846]: I0202 14:20:45.759996 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5tjzb"] Feb 02 14:20:46 crc kubenswrapper[4846]: I0202 14:20:46.785375 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5tjzb" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="registry-server" containerID="cri-o://44df8824c18d04591792085a0544f96ac39be468f776ebf35d0821bc6242aa15" gracePeriod=2 Feb 02 14:20:47 crc kubenswrapper[4846]: I0202 14:20:47.795846 4846 generic.go:334] "Generic (PLEG): container finished" podID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerID="44df8824c18d04591792085a0544f96ac39be468f776ebf35d0821bc6242aa15" exitCode=0 Feb 02 14:20:47 crc kubenswrapper[4846]: I0202 14:20:47.795914 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tjzb" event={"ID":"49bef830-de5d-4257-88b7-b4f0f3adaa39","Type":"ContainerDied","Data":"44df8824c18d04591792085a0544f96ac39be468f776ebf35d0821bc6242aa15"} Feb 02 14:20:47 crc kubenswrapper[4846]: I0202 14:20:47.796153 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tjzb" event={"ID":"49bef830-de5d-4257-88b7-b4f0f3adaa39","Type":"ContainerDied","Data":"90712881616cb2ebef188659280815a1e97e163c26b606994264af3f7dd34c83"} Feb 02 14:20:47 crc kubenswrapper[4846]: I0202 14:20:47.796167 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90712881616cb2ebef188659280815a1e97e163c26b606994264af3f7dd34c83" Feb 02 14:20:47 crc kubenswrapper[4846]: I0202 14:20:47.812032 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:47 crc kubenswrapper[4846]: I0202 14:20:47.944252 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-catalog-content\") pod \"49bef830-de5d-4257-88b7-b4f0f3adaa39\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " Feb 02 14:20:47 crc kubenswrapper[4846]: I0202 14:20:47.944409 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsk7m\" (UniqueName: \"kubernetes.io/projected/49bef830-de5d-4257-88b7-b4f0f3adaa39-kube-api-access-nsk7m\") pod \"49bef830-de5d-4257-88b7-b4f0f3adaa39\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " Feb 02 14:20:47 crc kubenswrapper[4846]: I0202 14:20:47.945646 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-utilities\") pod \"49bef830-de5d-4257-88b7-b4f0f3adaa39\" (UID: \"49bef830-de5d-4257-88b7-b4f0f3adaa39\") " Feb 02 14:20:47 crc kubenswrapper[4846]: I0202 14:20:47.946454 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-utilities" (OuterVolumeSpecName: "utilities") pod "49bef830-de5d-4257-88b7-b4f0f3adaa39" (UID: "49bef830-de5d-4257-88b7-b4f0f3adaa39"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:20:47 crc kubenswrapper[4846]: I0202 14:20:47.953897 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49bef830-de5d-4257-88b7-b4f0f3adaa39-kube-api-access-nsk7m" (OuterVolumeSpecName: "kube-api-access-nsk7m") pod "49bef830-de5d-4257-88b7-b4f0f3adaa39" (UID: "49bef830-de5d-4257-88b7-b4f0f3adaa39"). InnerVolumeSpecName "kube-api-access-nsk7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:20:48 crc kubenswrapper[4846]: I0202 14:20:48.048156 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:20:48 crc kubenswrapper[4846]: I0202 14:20:48.048197 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsk7m\" (UniqueName: \"kubernetes.io/projected/49bef830-de5d-4257-88b7-b4f0f3adaa39-kube-api-access-nsk7m\") on node \"crc\" DevicePath \"\"" Feb 02 14:20:48 crc kubenswrapper[4846]: I0202 14:20:48.082788 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49bef830-de5d-4257-88b7-b4f0f3adaa39" (UID: "49bef830-de5d-4257-88b7-b4f0f3adaa39"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:20:48 crc kubenswrapper[4846]: I0202 14:20:48.149926 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49bef830-de5d-4257-88b7-b4f0f3adaa39-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:20:48 crc kubenswrapper[4846]: I0202 14:20:48.805766 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tjzb" Feb 02 14:20:48 crc kubenswrapper[4846]: I0202 14:20:48.849381 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5tjzb"] Feb 02 14:20:48 crc kubenswrapper[4846]: I0202 14:20:48.861792 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5tjzb"] Feb 02 14:20:49 crc kubenswrapper[4846]: I0202 14:20:49.437394 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" path="/var/lib/kubelet/pods/49bef830-de5d-4257-88b7-b4f0f3adaa39/volumes" Feb 02 14:20:54 crc kubenswrapper[4846]: I0202 14:20:54.424342 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:20:54 crc kubenswrapper[4846]: E0202 14:20:54.425168 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:21:09 crc kubenswrapper[4846]: I0202 14:21:09.433312 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:21:09 crc kubenswrapper[4846]: E0202 14:21:09.434024 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:21:21 crc kubenswrapper[4846]: I0202 14:21:21.424492 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:21:21 crc kubenswrapper[4846]: E0202 14:21:21.428161 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:21:36 crc kubenswrapper[4846]: I0202 14:21:36.424157 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:21:36 crc kubenswrapper[4846]: E0202 14:21:36.425363 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.821642 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ftbgh"] Feb 02 14:21:39 crc kubenswrapper[4846]: E0202 14:21:39.823687 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="extract-utilities" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.823703 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="extract-utilities" Feb 02 14:21:39 crc kubenswrapper[4846]: E0202 14:21:39.823733 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerName="extract-utilities" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.823739 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerName="extract-utilities" Feb 02 14:21:39 crc kubenswrapper[4846]: E0202 14:21:39.823747 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerName="extract-content" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.823753 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerName="extract-content" Feb 02 14:21:39 crc kubenswrapper[4846]: E0202 14:21:39.823763 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerName="registry-server" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.823771 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerName="registry-server" Feb 02 14:21:39 crc kubenswrapper[4846]: E0202 14:21:39.823790 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="extract-content" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.823796 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="extract-content" Feb 02 14:21:39 crc kubenswrapper[4846]: E0202 14:21:39.823819 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="registry-server" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.823825 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="registry-server" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.824014 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="49bef830-de5d-4257-88b7-b4f0f3adaa39" containerName="registry-server" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.824030 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f7d3631-938a-4757-b83b-b9d80f3037e3" containerName="registry-server" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.825538 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.859537 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftbgh"] Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.928021 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-utilities\") pod \"redhat-marketplace-ftbgh\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.928536 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-catalog-content\") pod \"redhat-marketplace-ftbgh\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:39 crc kubenswrapper[4846]: I0202 14:21:39.929897 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7r7h\" (UniqueName: \"kubernetes.io/projected/30a9f1b7-a616-44a2-adb4-9c84712008ff-kube-api-access-b7r7h\") pod \"redhat-marketplace-ftbgh\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:40 crc kubenswrapper[4846]: I0202 14:21:40.032532 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7r7h\" (UniqueName: \"kubernetes.io/projected/30a9f1b7-a616-44a2-adb4-9c84712008ff-kube-api-access-b7r7h\") pod \"redhat-marketplace-ftbgh\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:40 crc kubenswrapper[4846]: I0202 14:21:40.032781 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-utilities\") pod \"redhat-marketplace-ftbgh\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:40 crc kubenswrapper[4846]: I0202 14:21:40.032879 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-catalog-content\") pod \"redhat-marketplace-ftbgh\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:40 crc kubenswrapper[4846]: I0202 14:21:40.033412 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-utilities\") pod \"redhat-marketplace-ftbgh\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:40 crc kubenswrapper[4846]: I0202 14:21:40.033444 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-catalog-content\") pod \"redhat-marketplace-ftbgh\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:40 crc kubenswrapper[4846]: I0202 14:21:40.055825 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7r7h\" (UniqueName: \"kubernetes.io/projected/30a9f1b7-a616-44a2-adb4-9c84712008ff-kube-api-access-b7r7h\") pod \"redhat-marketplace-ftbgh\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:40 crc kubenswrapper[4846]: I0202 14:21:40.153551 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:40 crc kubenswrapper[4846]: I0202 14:21:40.651421 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftbgh"] Feb 02 14:21:41 crc kubenswrapper[4846]: I0202 14:21:41.506408 4846 generic.go:334] "Generic (PLEG): container finished" podID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerID="1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e" exitCode=0 Feb 02 14:21:41 crc kubenswrapper[4846]: I0202 14:21:41.506502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftbgh" event={"ID":"30a9f1b7-a616-44a2-adb4-9c84712008ff","Type":"ContainerDied","Data":"1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e"} Feb 02 14:21:41 crc kubenswrapper[4846]: I0202 14:21:41.506803 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftbgh" event={"ID":"30a9f1b7-a616-44a2-adb4-9c84712008ff","Type":"ContainerStarted","Data":"f9be58cc9778d44a4c34bc9b7607be720ab149b6623ad3d2bc3f6f27d1d52871"} Feb 02 14:21:43 crc kubenswrapper[4846]: I0202 14:21:43.537136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftbgh" event={"ID":"30a9f1b7-a616-44a2-adb4-9c84712008ff","Type":"ContainerStarted","Data":"44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5"} Feb 02 14:21:44 crc kubenswrapper[4846]: I0202 14:21:44.548993 4846 generic.go:334] "Generic (PLEG): container finished" podID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerID="44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5" exitCode=0 Feb 02 14:21:44 crc kubenswrapper[4846]: I0202 14:21:44.549112 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftbgh" event={"ID":"30a9f1b7-a616-44a2-adb4-9c84712008ff","Type":"ContainerDied","Data":"44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5"} Feb 02 14:21:46 crc kubenswrapper[4846]: I0202 14:21:46.569557 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftbgh" event={"ID":"30a9f1b7-a616-44a2-adb4-9c84712008ff","Type":"ContainerStarted","Data":"0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140"} Feb 02 14:21:46 crc kubenswrapper[4846]: I0202 14:21:46.604518 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ftbgh" podStartSLOduration=3.662626171 podStartE2EDuration="7.60449811s" podCreationTimestamp="2026-02-02 14:21:39 +0000 UTC" firstStartedPulling="2026-02-02 14:21:41.511511745 +0000 UTC m=+7932.740098608" lastFinishedPulling="2026-02-02 14:21:45.453383674 +0000 UTC m=+7936.681970547" observedRunningTime="2026-02-02 14:21:46.59955032 +0000 UTC m=+7937.828137223" watchObservedRunningTime="2026-02-02 14:21:46.60449811 +0000 UTC m=+7937.833084973" Feb 02 14:21:50 crc kubenswrapper[4846]: I0202 14:21:50.154652 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:50 crc kubenswrapper[4846]: I0202 14:21:50.155268 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:50 crc kubenswrapper[4846]: I0202 14:21:50.207650 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:50 crc kubenswrapper[4846]: I0202 14:21:50.689821 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:50 crc kubenswrapper[4846]: I0202 14:21:50.754582 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftbgh"] Feb 02 14:21:51 crc kubenswrapper[4846]: I0202 14:21:51.424583 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:21:51 crc kubenswrapper[4846]: E0202 14:21:51.425056 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:21:52 crc kubenswrapper[4846]: I0202 14:21:52.645052 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ftbgh" podUID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerName="registry-server" containerID="cri-o://0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140" gracePeriod=2 Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.216089 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.247473 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-utilities\") pod \"30a9f1b7-a616-44a2-adb4-9c84712008ff\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.247525 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-catalog-content\") pod \"30a9f1b7-a616-44a2-adb4-9c84712008ff\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.247824 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7r7h\" (UniqueName: \"kubernetes.io/projected/30a9f1b7-a616-44a2-adb4-9c84712008ff-kube-api-access-b7r7h\") pod \"30a9f1b7-a616-44a2-adb4-9c84712008ff\" (UID: \"30a9f1b7-a616-44a2-adb4-9c84712008ff\") " Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.248740 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-utilities" (OuterVolumeSpecName: "utilities") pod "30a9f1b7-a616-44a2-adb4-9c84712008ff" (UID: "30a9f1b7-a616-44a2-adb4-9c84712008ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.255041 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a9f1b7-a616-44a2-adb4-9c84712008ff-kube-api-access-b7r7h" (OuterVolumeSpecName: "kube-api-access-b7r7h") pod "30a9f1b7-a616-44a2-adb4-9c84712008ff" (UID: "30a9f1b7-a616-44a2-adb4-9c84712008ff"). InnerVolumeSpecName "kube-api-access-b7r7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.319063 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30a9f1b7-a616-44a2-adb4-9c84712008ff" (UID: "30a9f1b7-a616-44a2-adb4-9c84712008ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.350335 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.350370 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30a9f1b7-a616-44a2-adb4-9c84712008ff-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.350384 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7r7h\" (UniqueName: \"kubernetes.io/projected/30a9f1b7-a616-44a2-adb4-9c84712008ff-kube-api-access-b7r7h\") on node \"crc\" DevicePath \"\"" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.655332 4846 generic.go:334] "Generic (PLEG): container finished" podID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerID="0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140" exitCode=0 Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.655399 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftbgh" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.655425 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftbgh" event={"ID":"30a9f1b7-a616-44a2-adb4-9c84712008ff","Type":"ContainerDied","Data":"0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140"} Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.656278 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftbgh" event={"ID":"30a9f1b7-a616-44a2-adb4-9c84712008ff","Type":"ContainerDied","Data":"f9be58cc9778d44a4c34bc9b7607be720ab149b6623ad3d2bc3f6f27d1d52871"} Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.656303 4846 scope.go:117] "RemoveContainer" containerID="0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.685178 4846 scope.go:117] "RemoveContainer" containerID="44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.685451 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftbgh"] Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.700414 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftbgh"] Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.711170 4846 scope.go:117] "RemoveContainer" containerID="1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.773396 4846 scope.go:117] "RemoveContainer" containerID="0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140" Feb 02 14:21:53 crc kubenswrapper[4846]: E0202 14:21:53.773781 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140\": container with ID starting with 0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140 not found: ID does not exist" containerID="0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.773819 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140"} err="failed to get container status \"0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140\": rpc error: code = NotFound desc = could not find container \"0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140\": container with ID starting with 0fa81c6552918733523e84dbd65ab228f9726b562a68f7fce5b801971aa74140 not found: ID does not exist" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.773846 4846 scope.go:117] "RemoveContainer" containerID="44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5" Feb 02 14:21:53 crc kubenswrapper[4846]: E0202 14:21:53.774131 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5\": container with ID starting with 44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5 not found: ID does not exist" containerID="44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.774181 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5"} err="failed to get container status \"44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5\": rpc error: code = NotFound desc = could not find container \"44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5\": container with ID starting with 44dd7d51cbcd7fcc0ae60054c7856779e9b12da1f13e855279ca1681e0b8e1f5 not found: ID does not exist" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.774212 4846 scope.go:117] "RemoveContainer" containerID="1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e" Feb 02 14:21:53 crc kubenswrapper[4846]: E0202 14:21:53.774476 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e\": container with ID starting with 1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e not found: ID does not exist" containerID="1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e" Feb 02 14:21:53 crc kubenswrapper[4846]: I0202 14:21:53.774502 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e"} err="failed to get container status \"1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e\": rpc error: code = NotFound desc = could not find container \"1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e\": container with ID starting with 1ac4387553aec53c6e5b7baefe4f91209fbd2ee3ce570f5e4017b75346ceb69e not found: ID does not exist" Feb 02 14:21:55 crc kubenswrapper[4846]: I0202 14:21:55.440671 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a9f1b7-a616-44a2-adb4-9c84712008ff" path="/var/lib/kubelet/pods/30a9f1b7-a616-44a2-adb4-9c84712008ff/volumes" Feb 02 14:22:04 crc kubenswrapper[4846]: I0202 14:22:04.423595 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:22:04 crc kubenswrapper[4846]: E0202 14:22:04.424288 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:22:17 crc kubenswrapper[4846]: I0202 14:22:17.424053 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:22:17 crc kubenswrapper[4846]: E0202 14:22:17.425416 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:22:30 crc kubenswrapper[4846]: I0202 14:22:30.426438 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:22:30 crc kubenswrapper[4846]: E0202 14:22:30.427746 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:22:44 crc kubenswrapper[4846]: I0202 14:22:44.425153 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:22:44 crc kubenswrapper[4846]: E0202 14:22:44.426176 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:22:58 crc kubenswrapper[4846]: I0202 14:22:58.423783 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:22:58 crc kubenswrapper[4846]: E0202 14:22:58.424559 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:23:13 crc kubenswrapper[4846]: I0202 14:23:13.424382 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:23:13 crc kubenswrapper[4846]: E0202 14:23:13.425614 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:23:26 crc kubenswrapper[4846]: I0202 14:23:26.424100 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:23:26 crc kubenswrapper[4846]: E0202 14:23:26.425555 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:23:38 crc kubenswrapper[4846]: I0202 14:23:38.423295 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:23:38 crc kubenswrapper[4846]: I0202 14:23:38.944476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"142167fe3c3a3e3800b91c942706509a1f1ffc663317249d41a0425b6ba59540"} Feb 02 14:24:16 crc kubenswrapper[4846]: I0202 14:24:16.392757 4846 generic.go:334] "Generic (PLEG): container finished" podID="8042645e-7a96-4a10-ab69-ba61f5280a61" containerID="92a5c4243ceced6961c60ebe767bbac32f31618c3427e4250d9f728096f55b03" exitCode=0 Feb 02 14:24:16 crc kubenswrapper[4846]: I0202 14:24:16.392861 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" event={"ID":"8042645e-7a96-4a10-ab69-ba61f5280a61","Type":"ContainerDied","Data":"92a5c4243ceced6961c60ebe767bbac32f31618c3427e4250d9f728096f55b03"} Feb 02 14:24:17 crc kubenswrapper[4846]: I0202 14:24:17.961374 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.108021 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-secret-0\") pod \"8042645e-7a96-4a10-ab69-ba61f5280a61\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.108206 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-inventory\") pod \"8042645e-7a96-4a10-ab69-ba61f5280a61\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.108301 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-ssh-key-openstack-cell1\") pod \"8042645e-7a96-4a10-ab69-ba61f5280a61\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.108342 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-combined-ca-bundle\") pod \"8042645e-7a96-4a10-ab69-ba61f5280a61\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.108458 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bws4f\" (UniqueName: \"kubernetes.io/projected/8042645e-7a96-4a10-ab69-ba61f5280a61-kube-api-access-bws4f\") pod \"8042645e-7a96-4a10-ab69-ba61f5280a61\" (UID: \"8042645e-7a96-4a10-ab69-ba61f5280a61\") " Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.128963 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "8042645e-7a96-4a10-ab69-ba61f5280a61" (UID: "8042645e-7a96-4a10-ab69-ba61f5280a61"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.129123 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8042645e-7a96-4a10-ab69-ba61f5280a61-kube-api-access-bws4f" (OuterVolumeSpecName: "kube-api-access-bws4f") pod "8042645e-7a96-4a10-ab69-ba61f5280a61" (UID: "8042645e-7a96-4a10-ab69-ba61f5280a61"). InnerVolumeSpecName "kube-api-access-bws4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.143714 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-inventory" (OuterVolumeSpecName: "inventory") pod "8042645e-7a96-4a10-ab69-ba61f5280a61" (UID: "8042645e-7a96-4a10-ab69-ba61f5280a61"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.149188 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "8042645e-7a96-4a10-ab69-ba61f5280a61" (UID: "8042645e-7a96-4a10-ab69-ba61f5280a61"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.160684 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "8042645e-7a96-4a10-ab69-ba61f5280a61" (UID: "8042645e-7a96-4a10-ab69-ba61f5280a61"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.210888 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.211220 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.211288 4846 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.211348 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bws4f\" (UniqueName: \"kubernetes.io/projected/8042645e-7a96-4a10-ab69-ba61f5280a61-kube-api-access-bws4f\") on node \"crc\" DevicePath \"\"" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.211402 4846 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/8042645e-7a96-4a10-ab69-ba61f5280a61-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.424893 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" event={"ID":"8042645e-7a96-4a10-ab69-ba61f5280a61","Type":"ContainerDied","Data":"a5145a2e272f74454efcb0e8dfa45da3e2f64aa2419972a4d516511c38e3b67e"} Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.425503 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5145a2e272f74454efcb0e8dfa45da3e2f64aa2419972a4d516511c38e3b67e" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.425729 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-qqkcn" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.545302 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-vzlgm"] Feb 02 14:24:18 crc kubenswrapper[4846]: E0202 14:24:18.545944 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerName="registry-server" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.545964 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerName="registry-server" Feb 02 14:24:18 crc kubenswrapper[4846]: E0202 14:24:18.545992 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerName="extract-content" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.546000 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerName="extract-content" Feb 02 14:24:18 crc kubenswrapper[4846]: E0202 14:24:18.546016 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerName="extract-utilities" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.546024 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerName="extract-utilities" Feb 02 14:24:18 crc kubenswrapper[4846]: E0202 14:24:18.546035 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8042645e-7a96-4a10-ab69-ba61f5280a61" containerName="libvirt-openstack-openstack-cell1" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.546044 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8042645e-7a96-4a10-ab69-ba61f5280a61" containerName="libvirt-openstack-openstack-cell1" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.546331 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8042645e-7a96-4a10-ab69-ba61f5280a61" containerName="libvirt-openstack-openstack-cell1" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.546346 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a9f1b7-a616-44a2-adb4-9c84712008ff" containerName="registry-server" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.547518 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.553114 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.553605 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.553777 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.553939 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.554103 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.554330 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.554682 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.559084 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-vzlgm"] Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.628249 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.628360 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.628660 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmvj9\" (UniqueName: \"kubernetes.io/projected/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-kube-api-access-tmvj9\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.628789 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.628861 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.628947 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-inventory\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.628994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.629180 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.629302 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.731981 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.732040 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-inventory\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.732081 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.732143 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.732190 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.732262 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.732395 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.732523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmvj9\" (UniqueName: \"kubernetes.io/projected/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-kube-api-access-tmvj9\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.732580 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.734677 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.740120 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-inventory\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.740684 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.742050 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.742802 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.743198 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.743831 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.754309 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.757835 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmvj9\" (UniqueName: \"kubernetes.io/projected/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-kube-api-access-tmvj9\") pod \"nova-cell1-openstack-openstack-cell1-vzlgm\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:18 crc kubenswrapper[4846]: I0202 14:24:18.873013 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:24:19 crc kubenswrapper[4846]: I0202 14:24:19.459233 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-vzlgm"] Feb 02 14:24:19 crc kubenswrapper[4846]: I0202 14:24:19.484023 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 14:24:20 crc kubenswrapper[4846]: I0202 14:24:20.457941 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" event={"ID":"a1cd8b9b-045d-422d-ad00-60bccdec0ca7","Type":"ContainerStarted","Data":"ee7836778e5e9cd6d535ac6dc1d8e28bb9eacf287933825ddfeb6a710b08a436"} Feb 02 14:24:20 crc kubenswrapper[4846]: I0202 14:24:20.458499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" event={"ID":"a1cd8b9b-045d-422d-ad00-60bccdec0ca7","Type":"ContainerStarted","Data":"8078d44e98eae0cefe2b019a69cd5bd88700669d35f41c1a2ce483fc467db619"} Feb 02 14:24:20 crc kubenswrapper[4846]: I0202 14:24:20.478704 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" podStartSLOduration=1.975064584 podStartE2EDuration="2.478683679s" podCreationTimestamp="2026-02-02 14:24:18 +0000 UTC" firstStartedPulling="2026-02-02 14:24:19.483672746 +0000 UTC m=+8090.712259619" lastFinishedPulling="2026-02-02 14:24:19.987291851 +0000 UTC m=+8091.215878714" observedRunningTime="2026-02-02 14:24:20.47261142 +0000 UTC m=+8091.701198283" watchObservedRunningTime="2026-02-02 14:24:20.478683679 +0000 UTC m=+8091.707270552" Feb 02 14:26:00 crc kubenswrapper[4846]: I0202 14:26:00.479337 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:26:00 crc kubenswrapper[4846]: I0202 14:26:00.480046 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:26:30 crc kubenswrapper[4846]: I0202 14:26:30.478873 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:26:30 crc kubenswrapper[4846]: I0202 14:26:30.479338 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:26:44 crc kubenswrapper[4846]: I0202 14:26:44.982028 4846 scope.go:117] "RemoveContainer" containerID="ea262b4a04692eab2a8402637823a946a8d386359c5435aade43176289378ffd" Feb 02 14:26:45 crc kubenswrapper[4846]: I0202 14:26:45.026447 4846 scope.go:117] "RemoveContainer" containerID="366ad8bedd54d9b6694d2f8913b6c4ef2e88d65c85ce1437050dab2dc2374937" Feb 02 14:26:45 crc kubenswrapper[4846]: I0202 14:26:45.091609 4846 scope.go:117] "RemoveContainer" containerID="44df8824c18d04591792085a0544f96ac39be468f776ebf35d0821bc6242aa15" Feb 02 14:27:00 crc kubenswrapper[4846]: I0202 14:27:00.479426 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:27:00 crc kubenswrapper[4846]: I0202 14:27:00.480464 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:27:00 crc kubenswrapper[4846]: I0202 14:27:00.480544 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 14:27:00 crc kubenswrapper[4846]: I0202 14:27:00.481666 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"142167fe3c3a3e3800b91c942706509a1f1ffc663317249d41a0425b6ba59540"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 14:27:00 crc kubenswrapper[4846]: I0202 14:27:00.481829 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://142167fe3c3a3e3800b91c942706509a1f1ffc663317249d41a0425b6ba59540" gracePeriod=600 Feb 02 14:27:01 crc kubenswrapper[4846]: I0202 14:27:01.229468 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="142167fe3c3a3e3800b91c942706509a1f1ffc663317249d41a0425b6ba59540" exitCode=0 Feb 02 14:27:01 crc kubenswrapper[4846]: I0202 14:27:01.229663 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"142167fe3c3a3e3800b91c942706509a1f1ffc663317249d41a0425b6ba59540"} Feb 02 14:27:01 crc kubenswrapper[4846]: I0202 14:27:01.229983 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03"} Feb 02 14:27:01 crc kubenswrapper[4846]: I0202 14:27:01.230010 4846 scope.go:117] "RemoveContainer" containerID="f45d71ec9ccefb7d7d45ce9830c37ae326c3710b56ae46f4e4afe133fdb4e719" Feb 02 14:27:02 crc kubenswrapper[4846]: I0202 14:27:02.242989 4846 generic.go:334] "Generic (PLEG): container finished" podID="a1cd8b9b-045d-422d-ad00-60bccdec0ca7" containerID="ee7836778e5e9cd6d535ac6dc1d8e28bb9eacf287933825ddfeb6a710b08a436" exitCode=0 Feb 02 14:27:02 crc kubenswrapper[4846]: I0202 14:27:02.243098 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" event={"ID":"a1cd8b9b-045d-422d-ad00-60bccdec0ca7","Type":"ContainerDied","Data":"ee7836778e5e9cd6d535ac6dc1d8e28bb9eacf287933825ddfeb6a710b08a436"} Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.788074 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.965034 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-combined-ca-bundle\") pod \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.965390 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-0\") pod \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.965479 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-1\") pod \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.965495 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cells-global-config-0\") pod \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.965517 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-1\") pod \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.965562 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-ssh-key-openstack-cell1\") pod \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.965600 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmvj9\" (UniqueName: \"kubernetes.io/projected/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-kube-api-access-tmvj9\") pod \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.965678 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-0\") pod \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.965742 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-inventory\") pod \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\" (UID: \"a1cd8b9b-045d-422d-ad00-60bccdec0ca7\") " Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.973440 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "a1cd8b9b-045d-422d-ad00-60bccdec0ca7" (UID: "a1cd8b9b-045d-422d-ad00-60bccdec0ca7"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.973539 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-kube-api-access-tmvj9" (OuterVolumeSpecName: "kube-api-access-tmvj9") pod "a1cd8b9b-045d-422d-ad00-60bccdec0ca7" (UID: "a1cd8b9b-045d-422d-ad00-60bccdec0ca7"). InnerVolumeSpecName "kube-api-access-tmvj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:27:03 crc kubenswrapper[4846]: I0202 14:27:03.996390 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "a1cd8b9b-045d-422d-ad00-60bccdec0ca7" (UID: "a1cd8b9b-045d-422d-ad00-60bccdec0ca7"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.003567 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "a1cd8b9b-045d-422d-ad00-60bccdec0ca7" (UID: "a1cd8b9b-045d-422d-ad00-60bccdec0ca7"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.003676 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "a1cd8b9b-045d-422d-ad00-60bccdec0ca7" (UID: "a1cd8b9b-045d-422d-ad00-60bccdec0ca7"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.004244 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-inventory" (OuterVolumeSpecName: "inventory") pod "a1cd8b9b-045d-422d-ad00-60bccdec0ca7" (UID: "a1cd8b9b-045d-422d-ad00-60bccdec0ca7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.004670 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "a1cd8b9b-045d-422d-ad00-60bccdec0ca7" (UID: "a1cd8b9b-045d-422d-ad00-60bccdec0ca7"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.005286 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "a1cd8b9b-045d-422d-ad00-60bccdec0ca7" (UID: "a1cd8b9b-045d-422d-ad00-60bccdec0ca7"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.010139 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "a1cd8b9b-045d-422d-ad00-60bccdec0ca7" (UID: "a1cd8b9b-045d-422d-ad00-60bccdec0ca7"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.068507 4846 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.068540 4846 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.068549 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.068557 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.068577 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.068586 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmvj9\" (UniqueName: \"kubernetes.io/projected/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-kube-api-access-tmvj9\") on node \"crc\" DevicePath \"\"" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.068594 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.068603 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.068612 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cd8b9b-045d-422d-ad00-60bccdec0ca7-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.297608 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" event={"ID":"a1cd8b9b-045d-422d-ad00-60bccdec0ca7","Type":"ContainerDied","Data":"8078d44e98eae0cefe2b019a69cd5bd88700669d35f41c1a2ce483fc467db619"} Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.297722 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8078d44e98eae0cefe2b019a69cd5bd88700669d35f41c1a2ce483fc467db619" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.297787 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-vzlgm" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.393473 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-b5j8z"] Feb 02 14:27:04 crc kubenswrapper[4846]: E0202 14:27:04.394019 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1cd8b9b-045d-422d-ad00-60bccdec0ca7" containerName="nova-cell1-openstack-openstack-cell1" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.394035 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1cd8b9b-045d-422d-ad00-60bccdec0ca7" containerName="nova-cell1-openstack-openstack-cell1" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.394824 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1cd8b9b-045d-422d-ad00-60bccdec0ca7" containerName="nova-cell1-openstack-openstack-cell1" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.395803 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.398675 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.398964 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.399280 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.399525 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.399737 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.403564 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-b5j8z"] Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.477246 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.477354 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-inventory\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.477384 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw8p4\" (UniqueName: \"kubernetes.io/projected/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-kube-api-access-zw8p4\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.477486 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.477528 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.477603 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.477780 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.579662 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.579755 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.579793 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.579857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.579977 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.580066 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-inventory\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.580092 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw8p4\" (UniqueName: \"kubernetes.io/projected/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-kube-api-access-zw8p4\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.583669 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.584004 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.584767 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.585298 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-inventory\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.589040 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.597478 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.602528 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw8p4\" (UniqueName: \"kubernetes.io/projected/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-kube-api-access-zw8p4\") pod \"telemetry-openstack-openstack-cell1-b5j8z\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:04 crc kubenswrapper[4846]: I0202 14:27:04.731165 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:27:05 crc kubenswrapper[4846]: I0202 14:27:05.313330 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-b5j8z"] Feb 02 14:27:06 crc kubenswrapper[4846]: I0202 14:27:06.344145 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" event={"ID":"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545","Type":"ContainerStarted","Data":"9d4e6dd631f34ed36c69a8f7fc079a01b587b772f408806d4a44b3bc863f7554"} Feb 02 14:27:06 crc kubenswrapper[4846]: I0202 14:27:06.344570 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" event={"ID":"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545","Type":"ContainerStarted","Data":"ffef1dc80878b236062833c989dfada7980c3332d9db1fb284dd9cf987eb6798"} Feb 02 14:27:06 crc kubenswrapper[4846]: I0202 14:27:06.368465 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" podStartSLOduration=1.869451456 podStartE2EDuration="2.368443818s" podCreationTimestamp="2026-02-02 14:27:04 +0000 UTC" firstStartedPulling="2026-02-02 14:27:05.318482001 +0000 UTC m=+8256.547068864" lastFinishedPulling="2026-02-02 14:27:05.817474363 +0000 UTC m=+8257.046061226" observedRunningTime="2026-02-02 14:27:06.363325854 +0000 UTC m=+8257.591912797" watchObservedRunningTime="2026-02-02 14:27:06.368443818 +0000 UTC m=+8257.597030681" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.474704 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5vtzq"] Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.479347 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.496062 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5vtzq"] Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.551057 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-utilities\") pod \"certified-operators-5vtzq\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.551185 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-catalog-content\") pod \"certified-operators-5vtzq\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.551374 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhlgw\" (UniqueName: \"kubernetes.io/projected/c1edc7e2-42b3-4279-a421-fc09eba93f63-kube-api-access-bhlgw\") pod \"certified-operators-5vtzq\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.654163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-catalog-content\") pod \"certified-operators-5vtzq\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.654325 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhlgw\" (UniqueName: \"kubernetes.io/projected/c1edc7e2-42b3-4279-a421-fc09eba93f63-kube-api-access-bhlgw\") pod \"certified-operators-5vtzq\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.654477 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-utilities\") pod \"certified-operators-5vtzq\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.654982 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-catalog-content\") pod \"certified-operators-5vtzq\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.655001 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-utilities\") pod \"certified-operators-5vtzq\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.680940 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhlgw\" (UniqueName: \"kubernetes.io/projected/c1edc7e2-42b3-4279-a421-fc09eba93f63-kube-api-access-bhlgw\") pod \"certified-operators-5vtzq\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:18 crc kubenswrapper[4846]: I0202 14:28:18.802128 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:19 crc kubenswrapper[4846]: I0202 14:28:19.419106 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5vtzq"] Feb 02 14:28:19 crc kubenswrapper[4846]: W0202 14:28:19.449594 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1edc7e2_42b3_4279_a421_fc09eba93f63.slice/crio-7a6196d96d5b72c4bbbdc175635ec9aca95439b6983d83cba170a80e343c606c WatchSource:0}: Error finding container 7a6196d96d5b72c4bbbdc175635ec9aca95439b6983d83cba170a80e343c606c: Status 404 returned error can't find the container with id 7a6196d96d5b72c4bbbdc175635ec9aca95439b6983d83cba170a80e343c606c Feb 02 14:28:20 crc kubenswrapper[4846]: I0202 14:28:20.279390 4846 generic.go:334] "Generic (PLEG): container finished" podID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerID="7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201" exitCode=0 Feb 02 14:28:20 crc kubenswrapper[4846]: I0202 14:28:20.279454 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5vtzq" event={"ID":"c1edc7e2-42b3-4279-a421-fc09eba93f63","Type":"ContainerDied","Data":"7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201"} Feb 02 14:28:20 crc kubenswrapper[4846]: I0202 14:28:20.279874 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5vtzq" event={"ID":"c1edc7e2-42b3-4279-a421-fc09eba93f63","Type":"ContainerStarted","Data":"7a6196d96d5b72c4bbbdc175635ec9aca95439b6983d83cba170a80e343c606c"} Feb 02 14:28:22 crc kubenswrapper[4846]: I0202 14:28:22.302898 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5vtzq" event={"ID":"c1edc7e2-42b3-4279-a421-fc09eba93f63","Type":"ContainerStarted","Data":"9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe"} Feb 02 14:28:26 crc kubenswrapper[4846]: I0202 14:28:26.363171 4846 generic.go:334] "Generic (PLEG): container finished" podID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerID="9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe" exitCode=0 Feb 02 14:28:26 crc kubenswrapper[4846]: I0202 14:28:26.363296 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5vtzq" event={"ID":"c1edc7e2-42b3-4279-a421-fc09eba93f63","Type":"ContainerDied","Data":"9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe"} Feb 02 14:28:27 crc kubenswrapper[4846]: I0202 14:28:27.376803 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5vtzq" event={"ID":"c1edc7e2-42b3-4279-a421-fc09eba93f63","Type":"ContainerStarted","Data":"91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22"} Feb 02 14:28:27 crc kubenswrapper[4846]: I0202 14:28:27.403808 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5vtzq" podStartSLOduration=2.791154777 podStartE2EDuration="9.403784081s" podCreationTimestamp="2026-02-02 14:28:18 +0000 UTC" firstStartedPulling="2026-02-02 14:28:20.283060874 +0000 UTC m=+8331.511647777" lastFinishedPulling="2026-02-02 14:28:26.895690218 +0000 UTC m=+8338.124277081" observedRunningTime="2026-02-02 14:28:27.395429528 +0000 UTC m=+8338.624016411" watchObservedRunningTime="2026-02-02 14:28:27.403784081 +0000 UTC m=+8338.632370954" Feb 02 14:28:28 crc kubenswrapper[4846]: I0202 14:28:28.802614 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:28 crc kubenswrapper[4846]: I0202 14:28:28.803012 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:29 crc kubenswrapper[4846]: I0202 14:28:29.882337 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-5vtzq" podUID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerName="registry-server" probeResult="failure" output=< Feb 02 14:28:29 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:28:29 crc kubenswrapper[4846]: > Feb 02 14:28:38 crc kubenswrapper[4846]: I0202 14:28:38.857153 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:38 crc kubenswrapper[4846]: I0202 14:28:38.919820 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:39 crc kubenswrapper[4846]: I0202 14:28:39.107658 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5vtzq"] Feb 02 14:28:40 crc kubenswrapper[4846]: I0202 14:28:40.515027 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5vtzq" podUID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerName="registry-server" containerID="cri-o://91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22" gracePeriod=2 Feb 02 14:28:40 crc kubenswrapper[4846]: I0202 14:28:40.999359 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.053942 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhlgw\" (UniqueName: \"kubernetes.io/projected/c1edc7e2-42b3-4279-a421-fc09eba93f63-kube-api-access-bhlgw\") pod \"c1edc7e2-42b3-4279-a421-fc09eba93f63\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.054651 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-utilities\") pod \"c1edc7e2-42b3-4279-a421-fc09eba93f63\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.054968 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-catalog-content\") pod \"c1edc7e2-42b3-4279-a421-fc09eba93f63\" (UID: \"c1edc7e2-42b3-4279-a421-fc09eba93f63\") " Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.055597 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-utilities" (OuterVolumeSpecName: "utilities") pod "c1edc7e2-42b3-4279-a421-fc09eba93f63" (UID: "c1edc7e2-42b3-4279-a421-fc09eba93f63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.062501 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.064268 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1edc7e2-42b3-4279-a421-fc09eba93f63-kube-api-access-bhlgw" (OuterVolumeSpecName: "kube-api-access-bhlgw") pod "c1edc7e2-42b3-4279-a421-fc09eba93f63" (UID: "c1edc7e2-42b3-4279-a421-fc09eba93f63"). InnerVolumeSpecName "kube-api-access-bhlgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.114432 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1edc7e2-42b3-4279-a421-fc09eba93f63" (UID: "c1edc7e2-42b3-4279-a421-fc09eba93f63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.164166 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhlgw\" (UniqueName: \"kubernetes.io/projected/c1edc7e2-42b3-4279-a421-fc09eba93f63-kube-api-access-bhlgw\") on node \"crc\" DevicePath \"\"" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.164236 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1edc7e2-42b3-4279-a421-fc09eba93f63-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.529509 4846 generic.go:334] "Generic (PLEG): container finished" podID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerID="91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22" exitCode=0 Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.529573 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5vtzq" event={"ID":"c1edc7e2-42b3-4279-a421-fc09eba93f63","Type":"ContainerDied","Data":"91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22"} Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.529600 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5vtzq" event={"ID":"c1edc7e2-42b3-4279-a421-fc09eba93f63","Type":"ContainerDied","Data":"7a6196d96d5b72c4bbbdc175635ec9aca95439b6983d83cba170a80e343c606c"} Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.529679 4846 scope.go:117] "RemoveContainer" containerID="91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.530928 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5vtzq" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.561412 4846 scope.go:117] "RemoveContainer" containerID="9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.564000 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5vtzq"] Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.588568 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5vtzq"] Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.599029 4846 scope.go:117] "RemoveContainer" containerID="7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.659302 4846 scope.go:117] "RemoveContainer" containerID="91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22" Feb 02 14:28:41 crc kubenswrapper[4846]: E0202 14:28:41.659997 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22\": container with ID starting with 91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22 not found: ID does not exist" containerID="91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.660059 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22"} err="failed to get container status \"91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22\": rpc error: code = NotFound desc = could not find container \"91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22\": container with ID starting with 91bc40bb8b9d73e8b276b6eedf2a345b2c255e03e0639cf2a935a0bd22486f22 not found: ID does not exist" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.660096 4846 scope.go:117] "RemoveContainer" containerID="9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe" Feb 02 14:28:41 crc kubenswrapper[4846]: E0202 14:28:41.660897 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe\": container with ID starting with 9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe not found: ID does not exist" containerID="9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.660940 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe"} err="failed to get container status \"9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe\": rpc error: code = NotFound desc = could not find container \"9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe\": container with ID starting with 9ed4be2a28e92c6e21e947e8580612705c2f749cb2476a2c0dccca64e195d4fe not found: ID does not exist" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.660966 4846 scope.go:117] "RemoveContainer" containerID="7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201" Feb 02 14:28:41 crc kubenswrapper[4846]: E0202 14:28:41.665322 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201\": container with ID starting with 7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201 not found: ID does not exist" containerID="7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201" Feb 02 14:28:41 crc kubenswrapper[4846]: I0202 14:28:41.665353 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201"} err="failed to get container status \"7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201\": rpc error: code = NotFound desc = could not find container \"7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201\": container with ID starting with 7f790ee98a9d503f7ca86c77a87c69659d2fd00ab6447b8ff393e9e427eed201 not found: ID does not exist" Feb 02 14:28:43 crc kubenswrapper[4846]: I0202 14:28:43.439582 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1edc7e2-42b3-4279-a421-fc09eba93f63" path="/var/lib/kubelet/pods/c1edc7e2-42b3-4279-a421-fc09eba93f63/volumes" Feb 02 14:29:00 crc kubenswrapper[4846]: I0202 14:29:00.478793 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:29:00 crc kubenswrapper[4846]: I0202 14:29:00.479724 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:29:30 crc kubenswrapper[4846]: I0202 14:29:30.479491 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:29:30 crc kubenswrapper[4846]: I0202 14:29:30.481883 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.162805 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f"] Feb 02 14:30:00 crc kubenswrapper[4846]: E0202 14:30:00.163944 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerName="extract-content" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.163966 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerName="extract-content" Feb 02 14:30:00 crc kubenswrapper[4846]: E0202 14:30:00.163999 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerName="registry-server" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.164008 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerName="registry-server" Feb 02 14:30:00 crc kubenswrapper[4846]: E0202 14:30:00.164028 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerName="extract-utilities" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.164035 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerName="extract-utilities" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.164304 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1edc7e2-42b3-4279-a421-fc09eba93f63" containerName="registry-server" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.165299 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.171957 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.171989 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.191360 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f"] Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.266068 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e03a5b7-19c7-4a63-8144-709df73ef388-secret-volume\") pod \"collect-profiles-29500710-lgt6f\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.266138 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72mvv\" (UniqueName: \"kubernetes.io/projected/1e03a5b7-19c7-4a63-8144-709df73ef388-kube-api-access-72mvv\") pod \"collect-profiles-29500710-lgt6f\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.266172 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e03a5b7-19c7-4a63-8144-709df73ef388-config-volume\") pod \"collect-profiles-29500710-lgt6f\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.369148 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e03a5b7-19c7-4a63-8144-709df73ef388-secret-volume\") pod \"collect-profiles-29500710-lgt6f\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.369431 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72mvv\" (UniqueName: \"kubernetes.io/projected/1e03a5b7-19c7-4a63-8144-709df73ef388-kube-api-access-72mvv\") pod \"collect-profiles-29500710-lgt6f\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.369477 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e03a5b7-19c7-4a63-8144-709df73ef388-config-volume\") pod \"collect-profiles-29500710-lgt6f\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.371032 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e03a5b7-19c7-4a63-8144-709df73ef388-config-volume\") pod \"collect-profiles-29500710-lgt6f\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.381792 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e03a5b7-19c7-4a63-8144-709df73ef388-secret-volume\") pod \"collect-profiles-29500710-lgt6f\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.390837 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72mvv\" (UniqueName: \"kubernetes.io/projected/1e03a5b7-19c7-4a63-8144-709df73ef388-kube-api-access-72mvv\") pod \"collect-profiles-29500710-lgt6f\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.482111 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.482185 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.482260 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.483275 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.483361 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" gracePeriod=600 Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.488405 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:00 crc kubenswrapper[4846]: E0202 14:30:00.614213 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:30:00 crc kubenswrapper[4846]: I0202 14:30:00.960834 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f"] Feb 02 14:30:01 crc kubenswrapper[4846]: I0202 14:30:01.506944 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" event={"ID":"1e03a5b7-19c7-4a63-8144-709df73ef388","Type":"ContainerStarted","Data":"edcac77820bf06b5ef1583dbf04bc8c3d7412b0ef348c7a4428e6a06a7577d49"} Feb 02 14:30:01 crc kubenswrapper[4846]: I0202 14:30:01.507435 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" event={"ID":"1e03a5b7-19c7-4a63-8144-709df73ef388","Type":"ContainerStarted","Data":"6d829b3fa07c4ad7d07b58c8b107c06dc08e0c76b8b96879ebf9e8ae4905a42e"} Feb 02 14:30:01 crc kubenswrapper[4846]: I0202 14:30:01.511667 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" exitCode=0 Feb 02 14:30:01 crc kubenswrapper[4846]: I0202 14:30:01.511808 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03"} Feb 02 14:30:01 crc kubenswrapper[4846]: I0202 14:30:01.512052 4846 scope.go:117] "RemoveContainer" containerID="142167fe3c3a3e3800b91c942706509a1f1ffc663317249d41a0425b6ba59540" Feb 02 14:30:01 crc kubenswrapper[4846]: I0202 14:30:01.512939 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:30:01 crc kubenswrapper[4846]: E0202 14:30:01.513588 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:30:01 crc kubenswrapper[4846]: I0202 14:30:01.564019 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" podStartSLOduration=1.5639941130000001 podStartE2EDuration="1.563994113s" podCreationTimestamp="2026-02-02 14:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 14:30:01.534737572 +0000 UTC m=+8432.763324445" watchObservedRunningTime="2026-02-02 14:30:01.563994113 +0000 UTC m=+8432.792580996" Feb 02 14:30:02 crc kubenswrapper[4846]: I0202 14:30:02.526380 4846 generic.go:334] "Generic (PLEG): container finished" podID="1e03a5b7-19c7-4a63-8144-709df73ef388" containerID="edcac77820bf06b5ef1583dbf04bc8c3d7412b0ef348c7a4428e6a06a7577d49" exitCode=0 Feb 02 14:30:02 crc kubenswrapper[4846]: I0202 14:30:02.526435 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" event={"ID":"1e03a5b7-19c7-4a63-8144-709df73ef388","Type":"ContainerDied","Data":"edcac77820bf06b5ef1583dbf04bc8c3d7412b0ef348c7a4428e6a06a7577d49"} Feb 02 14:30:03 crc kubenswrapper[4846]: I0202 14:30:03.929556 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.054251 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72mvv\" (UniqueName: \"kubernetes.io/projected/1e03a5b7-19c7-4a63-8144-709df73ef388-kube-api-access-72mvv\") pod \"1e03a5b7-19c7-4a63-8144-709df73ef388\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.054404 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e03a5b7-19c7-4a63-8144-709df73ef388-secret-volume\") pod \"1e03a5b7-19c7-4a63-8144-709df73ef388\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.054470 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e03a5b7-19c7-4a63-8144-709df73ef388-config-volume\") pod \"1e03a5b7-19c7-4a63-8144-709df73ef388\" (UID: \"1e03a5b7-19c7-4a63-8144-709df73ef388\") " Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.055393 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e03a5b7-19c7-4a63-8144-709df73ef388-config-volume" (OuterVolumeSpecName: "config-volume") pod "1e03a5b7-19c7-4a63-8144-709df73ef388" (UID: "1e03a5b7-19c7-4a63-8144-709df73ef388"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.064506 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e03a5b7-19c7-4a63-8144-709df73ef388-kube-api-access-72mvv" (OuterVolumeSpecName: "kube-api-access-72mvv") pod "1e03a5b7-19c7-4a63-8144-709df73ef388" (UID: "1e03a5b7-19c7-4a63-8144-709df73ef388"). InnerVolumeSpecName "kube-api-access-72mvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.064872 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e03a5b7-19c7-4a63-8144-709df73ef388-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1e03a5b7-19c7-4a63-8144-709df73ef388" (UID: "1e03a5b7-19c7-4a63-8144-709df73ef388"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.158132 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72mvv\" (UniqueName: \"kubernetes.io/projected/1e03a5b7-19c7-4a63-8144-709df73ef388-kube-api-access-72mvv\") on node \"crc\" DevicePath \"\"" Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.158217 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e03a5b7-19c7-4a63-8144-709df73ef388-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.158236 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e03a5b7-19c7-4a63-8144-709df73ef388-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.552919 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" event={"ID":"1e03a5b7-19c7-4a63-8144-709df73ef388","Type":"ContainerDied","Data":"6d829b3fa07c4ad7d07b58c8b107c06dc08e0c76b8b96879ebf9e8ae4905a42e"} Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.553211 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d829b3fa07c4ad7d07b58c8b107c06dc08e0c76b8b96879ebf9e8ae4905a42e" Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.552984 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500710-lgt6f" Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.612783 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd"] Feb 02 14:30:04 crc kubenswrapper[4846]: I0202 14:30:04.621295 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500665-c82nd"] Feb 02 14:30:05 crc kubenswrapper[4846]: I0202 14:30:05.439495 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="061cc83a-9da1-4990-bf23-663267b2a816" path="/var/lib/kubelet/pods/061cc83a-9da1-4990-bf23-663267b2a816/volumes" Feb 02 14:30:13 crc kubenswrapper[4846]: I0202 14:30:13.425032 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:30:13 crc kubenswrapper[4846]: E0202 14:30:13.426158 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:30:27 crc kubenswrapper[4846]: I0202 14:30:27.423919 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:30:27 crc kubenswrapper[4846]: E0202 14:30:27.424608 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:30:28 crc kubenswrapper[4846]: I0202 14:30:28.857329 4846 generic.go:334] "Generic (PLEG): container finished" podID="7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" containerID="9d4e6dd631f34ed36c69a8f7fc079a01b587b772f408806d4a44b3bc863f7554" exitCode=0 Feb 02 14:30:28 crc kubenswrapper[4846]: I0202 14:30:28.857462 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" event={"ID":"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545","Type":"ContainerDied","Data":"9d4e6dd631f34ed36c69a8f7fc079a01b587b772f408806d4a44b3bc863f7554"} Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.334411 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.407931 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-0\") pod \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.408029 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-1\") pod \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.408134 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw8p4\" (UniqueName: \"kubernetes.io/projected/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-kube-api-access-zw8p4\") pod \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.408177 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-2\") pod \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.408237 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ssh-key-openstack-cell1\") pod \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.408264 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-inventory\") pod \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.408281 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-telemetry-combined-ca-bundle\") pod \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\" (UID: \"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545\") " Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.413777 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" (UID: "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.424867 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-kube-api-access-zw8p4" (OuterVolumeSpecName: "kube-api-access-zw8p4") pod "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" (UID: "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545"). InnerVolumeSpecName "kube-api-access-zw8p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.443999 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" (UID: "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.446227 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" (UID: "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.452247 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" (UID: "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.453786 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" (UID: "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.455810 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-inventory" (OuterVolumeSpecName: "inventory") pod "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" (UID: "7fdfa2d0-e63f-4659-8c81-7bb3dfc07545"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.522314 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw8p4\" (UniqueName: \"kubernetes.io/projected/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-kube-api-access-zw8p4\") on node \"crc\" DevicePath \"\"" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.522348 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.522358 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.522368 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.522377 4846 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.522386 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.522395 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7fdfa2d0-e63f-4659-8c81-7bb3dfc07545-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.882223 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" event={"ID":"7fdfa2d0-e63f-4659-8c81-7bb3dfc07545","Type":"ContainerDied","Data":"ffef1dc80878b236062833c989dfada7980c3332d9db1fb284dd9cf987eb6798"} Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.882657 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffef1dc80878b236062833c989dfada7980c3332d9db1fb284dd9cf987eb6798" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.882361 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-b5j8z" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.982950 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-7n5sr"] Feb 02 14:30:30 crc kubenswrapper[4846]: E0202 14:30:30.983433 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e03a5b7-19c7-4a63-8144-709df73ef388" containerName="collect-profiles" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.983454 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e03a5b7-19c7-4a63-8144-709df73ef388" containerName="collect-profiles" Feb 02 14:30:30 crc kubenswrapper[4846]: E0202 14:30:30.983477 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" containerName="telemetry-openstack-openstack-cell1" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.983483 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" containerName="telemetry-openstack-openstack-cell1" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.983679 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fdfa2d0-e63f-4659-8c81-7bb3dfc07545" containerName="telemetry-openstack-openstack-cell1" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.983692 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e03a5b7-19c7-4a63-8144-709df73ef388" containerName="collect-profiles" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.984408 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.989909 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.989944 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.990388 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.990736 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:30:30 crc kubenswrapper[4846]: I0202 14:30:30.996638 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.006311 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-7n5sr"] Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.043786 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.044098 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.044208 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.044356 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.044516 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdrpx\" (UniqueName: \"kubernetes.io/projected/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-kube-api-access-rdrpx\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.146200 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.146288 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.146358 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.146502 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdrpx\" (UniqueName: \"kubernetes.io/projected/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-kube-api-access-rdrpx\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.146568 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.150968 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.151073 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.151534 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.152323 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.170996 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdrpx\" (UniqueName: \"kubernetes.io/projected/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-kube-api-access-rdrpx\") pod \"neutron-sriov-openstack-openstack-cell1-7n5sr\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.352863 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.934795 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-7n5sr"] Feb 02 14:30:31 crc kubenswrapper[4846]: I0202 14:30:31.966070 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 14:30:32 crc kubenswrapper[4846]: I0202 14:30:32.905454 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" event={"ID":"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc","Type":"ContainerStarted","Data":"42c672ce07c3c8ee74663acd72f0ffdf181f0370de67a605e2d4aa62e25039d1"} Feb 02 14:30:33 crc kubenswrapper[4846]: I0202 14:30:33.923276 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" event={"ID":"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc","Type":"ContainerStarted","Data":"34047f1db91775be97324329522fd00c1ce5bf74a5641b85dd45abea9bf14f24"} Feb 02 14:30:33 crc kubenswrapper[4846]: I0202 14:30:33.947518 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" podStartSLOduration=3.264396672 podStartE2EDuration="3.947500231s" podCreationTimestamp="2026-02-02 14:30:30 +0000 UTC" firstStartedPulling="2026-02-02 14:30:31.96581565 +0000 UTC m=+8463.194402513" lastFinishedPulling="2026-02-02 14:30:32.648919209 +0000 UTC m=+8463.877506072" observedRunningTime="2026-02-02 14:30:33.946528738 +0000 UTC m=+8465.175115631" watchObservedRunningTime="2026-02-02 14:30:33.947500231 +0000 UTC m=+8465.176087094" Feb 02 14:30:42 crc kubenswrapper[4846]: I0202 14:30:42.431187 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:30:42 crc kubenswrapper[4846]: E0202 14:30:42.432567 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:30:45 crc kubenswrapper[4846]: I0202 14:30:45.330614 4846 scope.go:117] "RemoveContainer" containerID="e00862ba6b7cda66a4e976c3aef34084c74f57957a876f2f21606a410a64a845" Feb 02 14:30:53 crc kubenswrapper[4846]: I0202 14:30:53.423896 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:30:53 crc kubenswrapper[4846]: E0202 14:30:53.424530 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:31:06 crc kubenswrapper[4846]: I0202 14:31:06.424066 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:31:06 crc kubenswrapper[4846]: E0202 14:31:06.425015 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.436026 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rw5rp"] Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.441296 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.448856 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rw5rp"] Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.569749 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lx48\" (UniqueName: \"kubernetes.io/projected/47a7e4dd-d847-4611-b249-ac9b36e99d4e-kube-api-access-7lx48\") pod \"community-operators-rw5rp\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.569950 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-utilities\") pod \"community-operators-rw5rp\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.570022 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-catalog-content\") pod \"community-operators-rw5rp\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.671797 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-utilities\") pod \"community-operators-rw5rp\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.672273 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-catalog-content\") pod \"community-operators-rw5rp\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.672454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lx48\" (UniqueName: \"kubernetes.io/projected/47a7e4dd-d847-4611-b249-ac9b36e99d4e-kube-api-access-7lx48\") pod \"community-operators-rw5rp\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.673806 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-utilities\") pod \"community-operators-rw5rp\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.674441 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-catalog-content\") pod \"community-operators-rw5rp\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.693552 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lx48\" (UniqueName: \"kubernetes.io/projected/47a7e4dd-d847-4611-b249-ac9b36e99d4e-kube-api-access-7lx48\") pod \"community-operators-rw5rp\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:10 crc kubenswrapper[4846]: I0202 14:31:10.776496 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:11 crc kubenswrapper[4846]: I0202 14:31:11.301514 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rw5rp"] Feb 02 14:31:12 crc kubenswrapper[4846]: I0202 14:31:12.318658 4846 generic.go:334] "Generic (PLEG): container finished" podID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerID="da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595" exitCode=0 Feb 02 14:31:12 crc kubenswrapper[4846]: I0202 14:31:12.318702 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw5rp" event={"ID":"47a7e4dd-d847-4611-b249-ac9b36e99d4e","Type":"ContainerDied","Data":"da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595"} Feb 02 14:31:12 crc kubenswrapper[4846]: I0202 14:31:12.318958 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw5rp" event={"ID":"47a7e4dd-d847-4611-b249-ac9b36e99d4e","Type":"ContainerStarted","Data":"5dab19060b3d4c617943a00300ab6cc00287bea6d86ccf669e13b2bff6ce149f"} Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.231459 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5c2fg"] Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.234603 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.247431 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5c2fg"] Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.338382 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2bzs\" (UniqueName: \"kubernetes.io/projected/c0269b7a-21e4-493b-9930-261800880764-kube-api-access-x2bzs\") pod \"redhat-operators-5c2fg\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.338876 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-utilities\") pod \"redhat-operators-5c2fg\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.339205 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-catalog-content\") pod \"redhat-operators-5c2fg\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.441350 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2bzs\" (UniqueName: \"kubernetes.io/projected/c0269b7a-21e4-493b-9930-261800880764-kube-api-access-x2bzs\") pod \"redhat-operators-5c2fg\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.441392 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-utilities\") pod \"redhat-operators-5c2fg\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.441548 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-catalog-content\") pod \"redhat-operators-5c2fg\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.442026 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-utilities\") pod \"redhat-operators-5c2fg\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.442125 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-catalog-content\") pod \"redhat-operators-5c2fg\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.462519 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2bzs\" (UniqueName: \"kubernetes.io/projected/c0269b7a-21e4-493b-9930-261800880764-kube-api-access-x2bzs\") pod \"redhat-operators-5c2fg\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:13 crc kubenswrapper[4846]: I0202 14:31:13.641973 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:14 crc kubenswrapper[4846]: I0202 14:31:14.126502 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5c2fg"] Feb 02 14:31:14 crc kubenswrapper[4846]: I0202 14:31:14.344854 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw5rp" event={"ID":"47a7e4dd-d847-4611-b249-ac9b36e99d4e","Type":"ContainerStarted","Data":"d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60"} Feb 02 14:31:14 crc kubenswrapper[4846]: I0202 14:31:14.350792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5c2fg" event={"ID":"c0269b7a-21e4-493b-9930-261800880764","Type":"ContainerStarted","Data":"a3d0ebdde643ac6893fd1d5ef3bcfc0a5161c6ed88e7bac4cb8e4c0128f53b59"} Feb 02 14:31:15 crc kubenswrapper[4846]: I0202 14:31:15.374734 4846 generic.go:334] "Generic (PLEG): container finished" podID="c0269b7a-21e4-493b-9930-261800880764" containerID="f9adf5db1b07b9c1602fc2f61a65900ca0fb2450e4dff2a6af35d3aadcd34ff9" exitCode=0 Feb 02 14:31:15 crc kubenswrapper[4846]: I0202 14:31:15.375259 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5c2fg" event={"ID":"c0269b7a-21e4-493b-9930-261800880764","Type":"ContainerDied","Data":"f9adf5db1b07b9c1602fc2f61a65900ca0fb2450e4dff2a6af35d3aadcd34ff9"} Feb 02 14:31:15 crc kubenswrapper[4846]: I0202 14:31:15.382316 4846 generic.go:334] "Generic (PLEG): container finished" podID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerID="d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60" exitCode=0 Feb 02 14:31:15 crc kubenswrapper[4846]: I0202 14:31:15.382366 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw5rp" event={"ID":"47a7e4dd-d847-4611-b249-ac9b36e99d4e","Type":"ContainerDied","Data":"d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60"} Feb 02 14:31:16 crc kubenswrapper[4846]: I0202 14:31:16.396055 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw5rp" event={"ID":"47a7e4dd-d847-4611-b249-ac9b36e99d4e","Type":"ContainerStarted","Data":"48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8"} Feb 02 14:31:16 crc kubenswrapper[4846]: I0202 14:31:16.424049 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rw5rp" podStartSLOduration=2.9286716 podStartE2EDuration="6.424024995s" podCreationTimestamp="2026-02-02 14:31:10 +0000 UTC" firstStartedPulling="2026-02-02 14:31:12.321132999 +0000 UTC m=+8503.549719862" lastFinishedPulling="2026-02-02 14:31:15.816486394 +0000 UTC m=+8507.045073257" observedRunningTime="2026-02-02 14:31:16.416133093 +0000 UTC m=+8507.644719946" watchObservedRunningTime="2026-02-02 14:31:16.424024995 +0000 UTC m=+8507.652611878" Feb 02 14:31:17 crc kubenswrapper[4846]: I0202 14:31:17.411071 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5c2fg" event={"ID":"c0269b7a-21e4-493b-9930-261800880764","Type":"ContainerStarted","Data":"e2d9c33be71cace6558a9ee40a1c4d8e91a134f9cf1636175a0ebfaf7dfbc247"} Feb 02 14:31:20 crc kubenswrapper[4846]: I0202 14:31:20.777554 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:20 crc kubenswrapper[4846]: I0202 14:31:20.778229 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:21 crc kubenswrapper[4846]: I0202 14:31:21.424536 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:31:21 crc kubenswrapper[4846]: E0202 14:31:21.424854 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:31:21 crc kubenswrapper[4846]: I0202 14:31:21.848691 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rw5rp" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerName="registry-server" probeResult="failure" output=< Feb 02 14:31:21 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:31:21 crc kubenswrapper[4846]: > Feb 02 14:31:27 crc kubenswrapper[4846]: I0202 14:31:27.527308 4846 generic.go:334] "Generic (PLEG): container finished" podID="c0269b7a-21e4-493b-9930-261800880764" containerID="e2d9c33be71cace6558a9ee40a1c4d8e91a134f9cf1636175a0ebfaf7dfbc247" exitCode=0 Feb 02 14:31:27 crc kubenswrapper[4846]: I0202 14:31:27.527407 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5c2fg" event={"ID":"c0269b7a-21e4-493b-9930-261800880764","Type":"ContainerDied","Data":"e2d9c33be71cace6558a9ee40a1c4d8e91a134f9cf1636175a0ebfaf7dfbc247"} Feb 02 14:31:28 crc kubenswrapper[4846]: I0202 14:31:28.543330 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5c2fg" event={"ID":"c0269b7a-21e4-493b-9930-261800880764","Type":"ContainerStarted","Data":"49bdbf3f4cf2ae8546b9a02e6d4609eafbd31dd57d331e8fb54596cd67e00b2c"} Feb 02 14:31:28 crc kubenswrapper[4846]: I0202 14:31:28.573472 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5c2fg" podStartSLOduration=2.8177663539999998 podStartE2EDuration="15.573451847s" podCreationTimestamp="2026-02-02 14:31:13 +0000 UTC" firstStartedPulling="2026-02-02 14:31:15.3832612 +0000 UTC m=+8506.611848063" lastFinishedPulling="2026-02-02 14:31:28.138946693 +0000 UTC m=+8519.367533556" observedRunningTime="2026-02-02 14:31:28.562725606 +0000 UTC m=+8519.791312469" watchObservedRunningTime="2026-02-02 14:31:28.573451847 +0000 UTC m=+8519.802038710" Feb 02 14:31:31 crc kubenswrapper[4846]: I0202 14:31:31.829342 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rw5rp" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerName="registry-server" probeResult="failure" output=< Feb 02 14:31:31 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:31:31 crc kubenswrapper[4846]: > Feb 02 14:31:33 crc kubenswrapper[4846]: I0202 14:31:33.601348 4846 generic.go:334] "Generic (PLEG): container finished" podID="23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc" containerID="34047f1db91775be97324329522fd00c1ce5bf74a5641b85dd45abea9bf14f24" exitCode=0 Feb 02 14:31:33 crc kubenswrapper[4846]: I0202 14:31:33.601433 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" event={"ID":"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc","Type":"ContainerDied","Data":"34047f1db91775be97324329522fd00c1ce5bf74a5641b85dd45abea9bf14f24"} Feb 02 14:31:33 crc kubenswrapper[4846]: I0202 14:31:33.642534 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:33 crc kubenswrapper[4846]: I0202 14:31:33.642899 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:31:34 crc kubenswrapper[4846]: I0202 14:31:34.424054 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:31:34 crc kubenswrapper[4846]: E0202 14:31:34.424806 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:31:34 crc kubenswrapper[4846]: I0202 14:31:34.694051 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5c2fg" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="registry-server" probeResult="failure" output=< Feb 02 14:31:34 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:31:34 crc kubenswrapper[4846]: > Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.089902 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.188784 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-ssh-key-openstack-cell1\") pod \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.188972 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-agent-neutron-config-0\") pod \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.189099 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdrpx\" (UniqueName: \"kubernetes.io/projected/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-kube-api-access-rdrpx\") pod \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.189194 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-inventory\") pod \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.189283 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-combined-ca-bundle\") pod \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\" (UID: \"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc\") " Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.194192 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc" (UID: "23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.194428 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-kube-api-access-rdrpx" (OuterVolumeSpecName: "kube-api-access-rdrpx") pod "23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc" (UID: "23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc"). InnerVolumeSpecName "kube-api-access-rdrpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.220297 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc" (UID: "23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.224427 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc" (UID: "23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.227366 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-inventory" (OuterVolumeSpecName: "inventory") pod "23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc" (UID: "23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.292742 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.292785 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdrpx\" (UniqueName: \"kubernetes.io/projected/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-kube-api-access-rdrpx\") on node \"crc\" DevicePath \"\"" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.292796 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.292807 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.292816 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.623940 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" event={"ID":"23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc","Type":"ContainerDied","Data":"42c672ce07c3c8ee74663acd72f0ffdf181f0370de67a605e2d4aa62e25039d1"} Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.623996 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42c672ce07c3c8ee74663acd72f0ffdf181f0370de67a605e2d4aa62e25039d1" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.624061 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-7n5sr" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.770240 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6"] Feb 02 14:31:35 crc kubenswrapper[4846]: E0202 14:31:35.770785 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc" containerName="neutron-sriov-openstack-openstack-cell1" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.770802 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc" containerName="neutron-sriov-openstack-openstack-cell1" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.771106 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc" containerName="neutron-sriov-openstack-openstack-cell1" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.772367 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.776231 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.777337 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.777642 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.777830 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.779123 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.793180 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6"] Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.912504 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.912880 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx5r2\" (UniqueName: \"kubernetes.io/projected/7bea9696-bedb-46b6-a458-51e7776a7922-kube-api-access-rx5r2\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.912916 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.912965 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:35 crc kubenswrapper[4846]: I0202 14:31:35.913015 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.015763 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.015847 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx5r2\" (UniqueName: \"kubernetes.io/projected/7bea9696-bedb-46b6-a458-51e7776a7922-kube-api-access-rx5r2\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.015883 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.015948 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.015982 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.023339 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.023608 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.024780 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.035180 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.040472 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx5r2\" (UniqueName: \"kubernetes.io/projected/7bea9696-bedb-46b6-a458-51e7776a7922-kube-api-access-rx5r2\") pod \"neutron-dhcp-openstack-openstack-cell1-zpkt6\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.094590 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:31:36 crc kubenswrapper[4846]: I0202 14:31:36.689174 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6"] Feb 02 14:31:37 crc kubenswrapper[4846]: I0202 14:31:37.652499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" event={"ID":"7bea9696-bedb-46b6-a458-51e7776a7922","Type":"ContainerStarted","Data":"cec3334572a564b4a419dd66fd9aba32f5e0e2b4aa6bacdf0841f06eaad468e2"} Feb 02 14:31:37 crc kubenswrapper[4846]: I0202 14:31:37.653289 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" event={"ID":"7bea9696-bedb-46b6-a458-51e7776a7922","Type":"ContainerStarted","Data":"8a804956698c9e5189789822eb79bd33a7102298199f62f977f5cd2c144d2d9a"} Feb 02 14:31:37 crc kubenswrapper[4846]: I0202 14:31:37.685137 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" podStartSLOduration=2.077916477 podStartE2EDuration="2.685114501s" podCreationTimestamp="2026-02-02 14:31:35 +0000 UTC" firstStartedPulling="2026-02-02 14:31:36.697073778 +0000 UTC m=+8527.925660631" lastFinishedPulling="2026-02-02 14:31:37.304271792 +0000 UTC m=+8528.532858655" observedRunningTime="2026-02-02 14:31:37.67193991 +0000 UTC m=+8528.900526773" watchObservedRunningTime="2026-02-02 14:31:37.685114501 +0000 UTC m=+8528.913701364" Feb 02 14:31:40 crc kubenswrapper[4846]: I0202 14:31:40.827271 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:40 crc kubenswrapper[4846]: I0202 14:31:40.879166 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:41 crc kubenswrapper[4846]: I0202 14:31:41.644232 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rw5rp"] Feb 02 14:31:42 crc kubenswrapper[4846]: I0202 14:31:42.697050 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rw5rp" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerName="registry-server" containerID="cri-o://48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8" gracePeriod=2 Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.194062 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.300659 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-utilities\") pod \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.300980 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-catalog-content\") pod \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.301164 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lx48\" (UniqueName: \"kubernetes.io/projected/47a7e4dd-d847-4611-b249-ac9b36e99d4e-kube-api-access-7lx48\") pod \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\" (UID: \"47a7e4dd-d847-4611-b249-ac9b36e99d4e\") " Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.301619 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-utilities" (OuterVolumeSpecName: "utilities") pod "47a7e4dd-d847-4611-b249-ac9b36e99d4e" (UID: "47a7e4dd-d847-4611-b249-ac9b36e99d4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.311130 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47a7e4dd-d847-4611-b249-ac9b36e99d4e-kube-api-access-7lx48" (OuterVolumeSpecName: "kube-api-access-7lx48") pod "47a7e4dd-d847-4611-b249-ac9b36e99d4e" (UID: "47a7e4dd-d847-4611-b249-ac9b36e99d4e"). InnerVolumeSpecName "kube-api-access-7lx48". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.343742 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47a7e4dd-d847-4611-b249-ac9b36e99d4e" (UID: "47a7e4dd-d847-4611-b249-ac9b36e99d4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.403636 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.403944 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47a7e4dd-d847-4611-b249-ac9b36e99d4e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.403959 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lx48\" (UniqueName: \"kubernetes.io/projected/47a7e4dd-d847-4611-b249-ac9b36e99d4e-kube-api-access-7lx48\") on node \"crc\" DevicePath \"\"" Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.712367 4846 generic.go:334] "Generic (PLEG): container finished" podID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerID="48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8" exitCode=0 Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.712418 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw5rp" event={"ID":"47a7e4dd-d847-4611-b249-ac9b36e99d4e","Type":"ContainerDied","Data":"48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8"} Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.712453 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw5rp" event={"ID":"47a7e4dd-d847-4611-b249-ac9b36e99d4e","Type":"ContainerDied","Data":"5dab19060b3d4c617943a00300ab6cc00287bea6d86ccf669e13b2bff6ce149f"} Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.712464 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw5rp" Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.712474 4846 scope.go:117] "RemoveContainer" containerID="48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8" Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.738228 4846 scope.go:117] "RemoveContainer" containerID="d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60" Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.746825 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rw5rp"] Feb 02 14:31:43 crc kubenswrapper[4846]: I0202 14:31:43.759614 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rw5rp"] Feb 02 14:31:44 crc kubenswrapper[4846]: I0202 14:31:44.270657 4846 scope.go:117] "RemoveContainer" containerID="da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595" Feb 02 14:31:44 crc kubenswrapper[4846]: I0202 14:31:44.302254 4846 scope.go:117] "RemoveContainer" containerID="48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8" Feb 02 14:31:44 crc kubenswrapper[4846]: E0202 14:31:44.302785 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8\": container with ID starting with 48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8 not found: ID does not exist" containerID="48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8" Feb 02 14:31:44 crc kubenswrapper[4846]: I0202 14:31:44.302839 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8"} err="failed to get container status \"48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8\": rpc error: code = NotFound desc = could not find container \"48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8\": container with ID starting with 48c6159aafd68c4ae967c25e37bb2568eac7c3269e64a306aa4057846d29ccb8 not found: ID does not exist" Feb 02 14:31:44 crc kubenswrapper[4846]: I0202 14:31:44.302872 4846 scope.go:117] "RemoveContainer" containerID="d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60" Feb 02 14:31:44 crc kubenswrapper[4846]: E0202 14:31:44.303399 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60\": container with ID starting with d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60 not found: ID does not exist" containerID="d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60" Feb 02 14:31:44 crc kubenswrapper[4846]: I0202 14:31:44.303461 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60"} err="failed to get container status \"d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60\": rpc error: code = NotFound desc = could not find container \"d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60\": container with ID starting with d8e5a35a58841d79414971a6ec38b479952db1b5f5b58a6285d5a1e939d44a60 not found: ID does not exist" Feb 02 14:31:44 crc kubenswrapper[4846]: I0202 14:31:44.303500 4846 scope.go:117] "RemoveContainer" containerID="da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595" Feb 02 14:31:44 crc kubenswrapper[4846]: E0202 14:31:44.303877 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595\": container with ID starting with da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595 not found: ID does not exist" containerID="da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595" Feb 02 14:31:44 crc kubenswrapper[4846]: I0202 14:31:44.303949 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595"} err="failed to get container status \"da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595\": rpc error: code = NotFound desc = could not find container \"da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595\": container with ID starting with da68dc4cd24dfa697d715f53755ea2bc16cc3b16ee7dcf3e12661d4b11c1e595 not found: ID does not exist" Feb 02 14:31:44 crc kubenswrapper[4846]: I0202 14:31:44.715188 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5c2fg" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="registry-server" probeResult="failure" output=< Feb 02 14:31:44 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:31:44 crc kubenswrapper[4846]: > Feb 02 14:31:45 crc kubenswrapper[4846]: I0202 14:31:45.437005 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" path="/var/lib/kubelet/pods/47a7e4dd-d847-4611-b249-ac9b36e99d4e/volumes" Feb 02 14:31:47 crc kubenswrapper[4846]: I0202 14:31:47.425395 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:31:47 crc kubenswrapper[4846]: E0202 14:31:47.425986 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:31:54 crc kubenswrapper[4846]: I0202 14:31:54.722042 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5c2fg" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="registry-server" probeResult="failure" output=< Feb 02 14:31:54 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:31:54 crc kubenswrapper[4846]: > Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.261149 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x7lp5"] Feb 02 14:31:56 crc kubenswrapper[4846]: E0202 14:31:56.261996 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerName="extract-utilities" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.262013 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerName="extract-utilities" Feb 02 14:31:56 crc kubenswrapper[4846]: E0202 14:31:56.262041 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerName="extract-content" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.262047 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerName="extract-content" Feb 02 14:31:56 crc kubenswrapper[4846]: E0202 14:31:56.262080 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerName="registry-server" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.262087 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerName="registry-server" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.262372 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="47a7e4dd-d847-4611-b249-ac9b36e99d4e" containerName="registry-server" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.264045 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.279834 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7lp5"] Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.376668 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-utilities\") pod \"redhat-marketplace-x7lp5\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.377047 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-catalog-content\") pod \"redhat-marketplace-x7lp5\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.377098 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd6x7\" (UniqueName: \"kubernetes.io/projected/60d812f8-e6e0-471f-8c61-015e86fbf6eb-kube-api-access-xd6x7\") pod \"redhat-marketplace-x7lp5\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.478961 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-catalog-content\") pod \"redhat-marketplace-x7lp5\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.479046 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd6x7\" (UniqueName: \"kubernetes.io/projected/60d812f8-e6e0-471f-8c61-015e86fbf6eb-kube-api-access-xd6x7\") pod \"redhat-marketplace-x7lp5\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.479301 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-utilities\") pod \"redhat-marketplace-x7lp5\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.479557 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-catalog-content\") pod \"redhat-marketplace-x7lp5\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.480099 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-utilities\") pod \"redhat-marketplace-x7lp5\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.502973 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd6x7\" (UniqueName: \"kubernetes.io/projected/60d812f8-e6e0-471f-8c61-015e86fbf6eb-kube-api-access-xd6x7\") pod \"redhat-marketplace-x7lp5\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:56 crc kubenswrapper[4846]: I0202 14:31:56.590320 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:31:57 crc kubenswrapper[4846]: I0202 14:31:57.101189 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7lp5"] Feb 02 14:31:57 crc kubenswrapper[4846]: W0202 14:31:57.105894 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60d812f8_e6e0_471f_8c61_015e86fbf6eb.slice/crio-d2c125d0b3bdad0f4951ed7988602dcd2f8d50e5b10eb016d74231811b85c9c6 WatchSource:0}: Error finding container d2c125d0b3bdad0f4951ed7988602dcd2f8d50e5b10eb016d74231811b85c9c6: Status 404 returned error can't find the container with id d2c125d0b3bdad0f4951ed7988602dcd2f8d50e5b10eb016d74231811b85c9c6 Feb 02 14:31:57 crc kubenswrapper[4846]: I0202 14:31:57.878261 4846 generic.go:334] "Generic (PLEG): container finished" podID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerID="6314e75b1fad2fba88c697eff1e6fe848638172c7310af91253414a38a087d2b" exitCode=0 Feb 02 14:31:57 crc kubenswrapper[4846]: I0202 14:31:57.878311 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7lp5" event={"ID":"60d812f8-e6e0-471f-8c61-015e86fbf6eb","Type":"ContainerDied","Data":"6314e75b1fad2fba88c697eff1e6fe848638172c7310af91253414a38a087d2b"} Feb 02 14:31:57 crc kubenswrapper[4846]: I0202 14:31:57.878339 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7lp5" event={"ID":"60d812f8-e6e0-471f-8c61-015e86fbf6eb","Type":"ContainerStarted","Data":"d2c125d0b3bdad0f4951ed7988602dcd2f8d50e5b10eb016d74231811b85c9c6"} Feb 02 14:32:00 crc kubenswrapper[4846]: I0202 14:32:00.919747 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7lp5" event={"ID":"60d812f8-e6e0-471f-8c61-015e86fbf6eb","Type":"ContainerStarted","Data":"798660d8e6bebfd034c98b5c82db482ce29795ab11b516642687298a7f3c65bc"} Feb 02 14:32:01 crc kubenswrapper[4846]: I0202 14:32:01.933043 4846 generic.go:334] "Generic (PLEG): container finished" podID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerID="798660d8e6bebfd034c98b5c82db482ce29795ab11b516642687298a7f3c65bc" exitCode=0 Feb 02 14:32:01 crc kubenswrapper[4846]: I0202 14:32:01.933161 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7lp5" event={"ID":"60d812f8-e6e0-471f-8c61-015e86fbf6eb","Type":"ContainerDied","Data":"798660d8e6bebfd034c98b5c82db482ce29795ab11b516642687298a7f3c65bc"} Feb 02 14:32:02 crc kubenswrapper[4846]: I0202 14:32:02.423925 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:32:02 crc kubenswrapper[4846]: E0202 14:32:02.424264 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:32:04 crc kubenswrapper[4846]: I0202 14:32:04.689217 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5c2fg" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="registry-server" probeResult="failure" output=< Feb 02 14:32:04 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:32:04 crc kubenswrapper[4846]: > Feb 02 14:32:04 crc kubenswrapper[4846]: I0202 14:32:04.960847 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7lp5" event={"ID":"60d812f8-e6e0-471f-8c61-015e86fbf6eb","Type":"ContainerStarted","Data":"2fad826f9548936e50e1a5f042c96581fbd92bec5d935c7a10a8d55688d6c3f2"} Feb 02 14:32:04 crc kubenswrapper[4846]: I0202 14:32:04.981374 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x7lp5" podStartSLOduration=2.560308324 podStartE2EDuration="8.981355841s" podCreationTimestamp="2026-02-02 14:31:56 +0000 UTC" firstStartedPulling="2026-02-02 14:31:57.882424153 +0000 UTC m=+8549.111011016" lastFinishedPulling="2026-02-02 14:32:04.30347167 +0000 UTC m=+8555.532058533" observedRunningTime="2026-02-02 14:32:04.977810745 +0000 UTC m=+8556.206397608" watchObservedRunningTime="2026-02-02 14:32:04.981355841 +0000 UTC m=+8556.209942704" Feb 02 14:32:06 crc kubenswrapper[4846]: I0202 14:32:06.590799 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:32:06 crc kubenswrapper[4846]: I0202 14:32:06.591347 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:32:06 crc kubenswrapper[4846]: I0202 14:32:06.638893 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:32:13 crc kubenswrapper[4846]: I0202 14:32:13.693340 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:32:13 crc kubenswrapper[4846]: I0202 14:32:13.744715 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:32:13 crc kubenswrapper[4846]: I0202 14:32:13.939077 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5c2fg"] Feb 02 14:32:15 crc kubenswrapper[4846]: I0202 14:32:15.085604 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5c2fg" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="registry-server" containerID="cri-o://49bdbf3f4cf2ae8546b9a02e6d4609eafbd31dd57d331e8fb54596cd67e00b2c" gracePeriod=2 Feb 02 14:32:15 crc kubenswrapper[4846]: I0202 14:32:15.423902 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:32:15 crc kubenswrapper[4846]: E0202 14:32:15.424747 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.100902 4846 generic.go:334] "Generic (PLEG): container finished" podID="c0269b7a-21e4-493b-9930-261800880764" containerID="49bdbf3f4cf2ae8546b9a02e6d4609eafbd31dd57d331e8fb54596cd67e00b2c" exitCode=0 Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.101072 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5c2fg" event={"ID":"c0269b7a-21e4-493b-9930-261800880764","Type":"ContainerDied","Data":"49bdbf3f4cf2ae8546b9a02e6d4609eafbd31dd57d331e8fb54596cd67e00b2c"} Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.101320 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5c2fg" event={"ID":"c0269b7a-21e4-493b-9930-261800880764","Type":"ContainerDied","Data":"a3d0ebdde643ac6893fd1d5ef3bcfc0a5161c6ed88e7bac4cb8e4c0128f53b59"} Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.101338 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3d0ebdde643ac6893fd1d5ef3bcfc0a5161c6ed88e7bac4cb8e4c0128f53b59" Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.158278 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.255396 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-catalog-content\") pod \"c0269b7a-21e4-493b-9930-261800880764\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.255565 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2bzs\" (UniqueName: \"kubernetes.io/projected/c0269b7a-21e4-493b-9930-261800880764-kube-api-access-x2bzs\") pod \"c0269b7a-21e4-493b-9930-261800880764\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.255754 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-utilities\") pod \"c0269b7a-21e4-493b-9930-261800880764\" (UID: \"c0269b7a-21e4-493b-9930-261800880764\") " Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.256675 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-utilities" (OuterVolumeSpecName: "utilities") pod "c0269b7a-21e4-493b-9930-261800880764" (UID: "c0269b7a-21e4-493b-9930-261800880764"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.261465 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0269b7a-21e4-493b-9930-261800880764-kube-api-access-x2bzs" (OuterVolumeSpecName: "kube-api-access-x2bzs") pod "c0269b7a-21e4-493b-9930-261800880764" (UID: "c0269b7a-21e4-493b-9930-261800880764"). InnerVolumeSpecName "kube-api-access-x2bzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.358122 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.358158 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2bzs\" (UniqueName: \"kubernetes.io/projected/c0269b7a-21e4-493b-9930-261800880764-kube-api-access-x2bzs\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.370055 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0269b7a-21e4-493b-9930-261800880764" (UID: "c0269b7a-21e4-493b-9930-261800880764"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.460691 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0269b7a-21e4-493b-9930-261800880764-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:16 crc kubenswrapper[4846]: I0202 14:32:16.670340 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:32:17 crc kubenswrapper[4846]: I0202 14:32:17.119464 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5c2fg" Feb 02 14:32:17 crc kubenswrapper[4846]: I0202 14:32:17.171556 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5c2fg"] Feb 02 14:32:17 crc kubenswrapper[4846]: I0202 14:32:17.180473 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5c2fg"] Feb 02 14:32:17 crc kubenswrapper[4846]: I0202 14:32:17.445053 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0269b7a-21e4-493b-9930-261800880764" path="/var/lib/kubelet/pods/c0269b7a-21e4-493b-9930-261800880764/volumes" Feb 02 14:32:18 crc kubenswrapper[4846]: I0202 14:32:18.941677 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7lp5"] Feb 02 14:32:18 crc kubenswrapper[4846]: I0202 14:32:18.942339 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x7lp5" podUID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerName="registry-server" containerID="cri-o://2fad826f9548936e50e1a5f042c96581fbd92bec5d935c7a10a8d55688d6c3f2" gracePeriod=2 Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.142732 4846 generic.go:334] "Generic (PLEG): container finished" podID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerID="2fad826f9548936e50e1a5f042c96581fbd92bec5d935c7a10a8d55688d6c3f2" exitCode=0 Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.142790 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7lp5" event={"ID":"60d812f8-e6e0-471f-8c61-015e86fbf6eb","Type":"ContainerDied","Data":"2fad826f9548936e50e1a5f042c96581fbd92bec5d935c7a10a8d55688d6c3f2"} Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.464696 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.639255 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-catalog-content\") pod \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.639388 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd6x7\" (UniqueName: \"kubernetes.io/projected/60d812f8-e6e0-471f-8c61-015e86fbf6eb-kube-api-access-xd6x7\") pod \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.639593 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-utilities\") pod \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\" (UID: \"60d812f8-e6e0-471f-8c61-015e86fbf6eb\") " Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.640458 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-utilities" (OuterVolumeSpecName: "utilities") pod "60d812f8-e6e0-471f-8c61-015e86fbf6eb" (UID: "60d812f8-e6e0-471f-8c61-015e86fbf6eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.645906 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60d812f8-e6e0-471f-8c61-015e86fbf6eb-kube-api-access-xd6x7" (OuterVolumeSpecName: "kube-api-access-xd6x7") pod "60d812f8-e6e0-471f-8c61-015e86fbf6eb" (UID: "60d812f8-e6e0-471f-8c61-015e86fbf6eb"). InnerVolumeSpecName "kube-api-access-xd6x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.663996 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60d812f8-e6e0-471f-8c61-015e86fbf6eb" (UID: "60d812f8-e6e0-471f-8c61-015e86fbf6eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.743025 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.743079 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60d812f8-e6e0-471f-8c61-015e86fbf6eb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:19 crc kubenswrapper[4846]: I0202 14:32:19.743096 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd6x7\" (UniqueName: \"kubernetes.io/projected/60d812f8-e6e0-471f-8c61-015e86fbf6eb-kube-api-access-xd6x7\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:20 crc kubenswrapper[4846]: I0202 14:32:20.167219 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x7lp5" event={"ID":"60d812f8-e6e0-471f-8c61-015e86fbf6eb","Type":"ContainerDied","Data":"d2c125d0b3bdad0f4951ed7988602dcd2f8d50e5b10eb016d74231811b85c9c6"} Feb 02 14:32:20 crc kubenswrapper[4846]: I0202 14:32:20.169582 4846 scope.go:117] "RemoveContainer" containerID="2fad826f9548936e50e1a5f042c96581fbd92bec5d935c7a10a8d55688d6c3f2" Feb 02 14:32:20 crc kubenswrapper[4846]: I0202 14:32:20.170040 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x7lp5" Feb 02 14:32:20 crc kubenswrapper[4846]: I0202 14:32:20.207571 4846 scope.go:117] "RemoveContainer" containerID="798660d8e6bebfd034c98b5c82db482ce29795ab11b516642687298a7f3c65bc" Feb 02 14:32:20 crc kubenswrapper[4846]: I0202 14:32:20.219614 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7lp5"] Feb 02 14:32:20 crc kubenswrapper[4846]: I0202 14:32:20.228087 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x7lp5"] Feb 02 14:32:20 crc kubenswrapper[4846]: I0202 14:32:20.247273 4846 scope.go:117] "RemoveContainer" containerID="6314e75b1fad2fba88c697eff1e6fe848638172c7310af91253414a38a087d2b" Feb 02 14:32:21 crc kubenswrapper[4846]: I0202 14:32:21.448239 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" path="/var/lib/kubelet/pods/60d812f8-e6e0-471f-8c61-015e86fbf6eb/volumes" Feb 02 14:32:30 crc kubenswrapper[4846]: I0202 14:32:30.423927 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:32:30 crc kubenswrapper[4846]: E0202 14:32:30.425010 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:32:41 crc kubenswrapper[4846]: I0202 14:32:41.423680 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:32:41 crc kubenswrapper[4846]: E0202 14:32:41.426531 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:32:44 crc kubenswrapper[4846]: I0202 14:32:44.416217 4846 generic.go:334] "Generic (PLEG): container finished" podID="7bea9696-bedb-46b6-a458-51e7776a7922" containerID="cec3334572a564b4a419dd66fd9aba32f5e0e2b4aa6bacdf0841f06eaad468e2" exitCode=0 Feb 02 14:32:44 crc kubenswrapper[4846]: I0202 14:32:44.416325 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" event={"ID":"7bea9696-bedb-46b6-a458-51e7776a7922","Type":"ContainerDied","Data":"cec3334572a564b4a419dd66fd9aba32f5e0e2b4aa6bacdf0841f06eaad468e2"} Feb 02 14:32:45 crc kubenswrapper[4846]: I0202 14:32:45.902912 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:32:45 crc kubenswrapper[4846]: I0202 14:32:45.979934 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-ssh-key-openstack-cell1\") pod \"7bea9696-bedb-46b6-a458-51e7776a7922\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " Feb 02 14:32:45 crc kubenswrapper[4846]: I0202 14:32:45.980136 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-inventory\") pod \"7bea9696-bedb-46b6-a458-51e7776a7922\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " Feb 02 14:32:45 crc kubenswrapper[4846]: I0202 14:32:45.980205 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-agent-neutron-config-0\") pod \"7bea9696-bedb-46b6-a458-51e7776a7922\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " Feb 02 14:32:45 crc kubenswrapper[4846]: I0202 14:32:45.980241 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-combined-ca-bundle\") pod \"7bea9696-bedb-46b6-a458-51e7776a7922\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " Feb 02 14:32:45 crc kubenswrapper[4846]: I0202 14:32:45.980431 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rx5r2\" (UniqueName: \"kubernetes.io/projected/7bea9696-bedb-46b6-a458-51e7776a7922-kube-api-access-rx5r2\") pod \"7bea9696-bedb-46b6-a458-51e7776a7922\" (UID: \"7bea9696-bedb-46b6-a458-51e7776a7922\") " Feb 02 14:32:45 crc kubenswrapper[4846]: I0202 14:32:45.988571 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "7bea9696-bedb-46b6-a458-51e7776a7922" (UID: "7bea9696-bedb-46b6-a458-51e7776a7922"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.009817 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bea9696-bedb-46b6-a458-51e7776a7922-kube-api-access-rx5r2" (OuterVolumeSpecName: "kube-api-access-rx5r2") pod "7bea9696-bedb-46b6-a458-51e7776a7922" (UID: "7bea9696-bedb-46b6-a458-51e7776a7922"). InnerVolumeSpecName "kube-api-access-rx5r2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.020699 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "7bea9696-bedb-46b6-a458-51e7776a7922" (UID: "7bea9696-bedb-46b6-a458-51e7776a7922"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.021944 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-inventory" (OuterVolumeSpecName: "inventory") pod "7bea9696-bedb-46b6-a458-51e7776a7922" (UID: "7bea9696-bedb-46b6-a458-51e7776a7922"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.028596 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "7bea9696-bedb-46b6-a458-51e7776a7922" (UID: "7bea9696-bedb-46b6-a458-51e7776a7922"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.083534 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.083579 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.083612 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.083663 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rx5r2\" (UniqueName: \"kubernetes.io/projected/7bea9696-bedb-46b6-a458-51e7776a7922-kube-api-access-rx5r2\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.083678 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7bea9696-bedb-46b6-a458-51e7776a7922-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.440159 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.440125 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-zpkt6" event={"ID":"7bea9696-bedb-46b6-a458-51e7776a7922","Type":"ContainerDied","Data":"8a804956698c9e5189789822eb79bd33a7102298199f62f977f5cd2c144d2d9a"} Feb 02 14:32:46 crc kubenswrapper[4846]: I0202 14:32:46.440606 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a804956698c9e5189789822eb79bd33a7102298199f62f977f5cd2c144d2d9a" Feb 02 14:32:55 crc kubenswrapper[4846]: I0202 14:32:55.424275 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:32:55 crc kubenswrapper[4846]: E0202 14:32:55.426269 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:33:00 crc kubenswrapper[4846]: I0202 14:33:00.092057 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 14:33:00 crc kubenswrapper[4846]: I0202 14:33:00.092747 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="4b608189-194f-4aae-bd70-9c6ea33f2d44" containerName="nova-cell0-conductor-conductor" containerID="cri-o://143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb" gracePeriod=30 Feb 02 14:33:00 crc kubenswrapper[4846]: I0202 14:33:00.136999 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 14:33:00 crc kubenswrapper[4846]: I0202 14:33:00.137488 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="83c7cd73-a72d-4f4f-9b36-b7677290ecfa" containerName="nova-cell1-conductor-conductor" containerID="cri-o://e515de17b99aabcf65b27fd30dc2a3f94580b7ba2de58e83857ffaf644ca0ad8" gracePeriod=30 Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.377823 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.379246 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1977e95e-eac6-462d-b7fd-caacae087fee" containerName="nova-scheduler-scheduler" containerID="cri-o://861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923" gracePeriod=30 Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.444721 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.445073 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerName="nova-metadata-log" containerID="cri-o://211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490" gracePeriod=30 Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.445738 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerName="nova-metadata-metadata" containerID="cri-o://34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb" gracePeriod=30 Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.468984 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.478108 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerName="nova-api-log" containerID="cri-o://a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3" gracePeriod=30 Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.478252 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerName="nova-api-api" containerID="cri-o://42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c" gracePeriod=30 Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.605951 4846 generic.go:334] "Generic (PLEG): container finished" podID="83c7cd73-a72d-4f4f-9b36-b7677290ecfa" containerID="e515de17b99aabcf65b27fd30dc2a3f94580b7ba2de58e83857ffaf644ca0ad8" exitCode=0 Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.606083 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"83c7cd73-a72d-4f4f-9b36-b7677290ecfa","Type":"ContainerDied","Data":"e515de17b99aabcf65b27fd30dc2a3f94580b7ba2de58e83857ffaf644ca0ad8"} Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.610038 4846 generic.go:334] "Generic (PLEG): container finished" podID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerID="a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3" exitCode=143 Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.610291 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"39fdb60a-d87e-438c-8de9-63e49fdda9b8","Type":"ContainerDied","Data":"a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3"} Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.612683 4846 generic.go:334] "Generic (PLEG): container finished" podID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerID="211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490" exitCode=143 Feb 02 14:33:01 crc kubenswrapper[4846]: I0202 14:33:01.612730 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23d3a089-c3dc-411d-a099-3ff80dd40d2d","Type":"ContainerDied","Data":"211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490"} Feb 02 14:33:01 crc kubenswrapper[4846]: E0202 14:33:01.647783 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 14:33:01 crc kubenswrapper[4846]: E0202 14:33:01.651938 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 14:33:01 crc kubenswrapper[4846]: E0202 14:33:01.653753 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 02 14:33:01 crc kubenswrapper[4846]: E0202 14:33:01.653843 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1977e95e-eac6-462d-b7fd-caacae087fee" containerName="nova-scheduler-scheduler" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.141926 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.290678 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-combined-ca-bundle\") pod \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.291187 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c785t\" (UniqueName: \"kubernetes.io/projected/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-kube-api-access-c785t\") pod \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.291240 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-config-data\") pod \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\" (UID: \"83c7cd73-a72d-4f4f-9b36-b7677290ecfa\") " Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.298799 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-kube-api-access-c785t" (OuterVolumeSpecName: "kube-api-access-c785t") pod "83c7cd73-a72d-4f4f-9b36-b7677290ecfa" (UID: "83c7cd73-a72d-4f4f-9b36-b7677290ecfa"). InnerVolumeSpecName "kube-api-access-c785t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.334735 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83c7cd73-a72d-4f4f-9b36-b7677290ecfa" (UID: "83c7cd73-a72d-4f4f-9b36-b7677290ecfa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.334750 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-config-data" (OuterVolumeSpecName: "config-data") pod "83c7cd73-a72d-4f4f-9b36-b7677290ecfa" (UID: "83c7cd73-a72d-4f4f-9b36-b7677290ecfa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.394272 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.394313 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c785t\" (UniqueName: \"kubernetes.io/projected/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-kube-api-access-c785t\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.394329 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c7cd73-a72d-4f4f-9b36-b7677290ecfa-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.633152 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"83c7cd73-a72d-4f4f-9b36-b7677290ecfa","Type":"ContainerDied","Data":"af7ddccd8d7a27c0629f75c22f8d8fb745372e2a4ede5cfb1b8b51a9c2764859"} Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.633207 4846 scope.go:117] "RemoveContainer" containerID="e515de17b99aabcf65b27fd30dc2a3f94580b7ba2de58e83857ffaf644ca0ad8" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.633247 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.676924 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.688667 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.704465 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 14:33:02 crc kubenswrapper[4846]: E0202 14:33:02.705125 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerName="extract-content" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.705148 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerName="extract-content" Feb 02 14:33:02 crc kubenswrapper[4846]: E0202 14:33:02.705170 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="extract-content" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.705181 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="extract-content" Feb 02 14:33:02 crc kubenswrapper[4846]: E0202 14:33:02.705200 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="extract-utilities" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.705208 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="extract-utilities" Feb 02 14:33:02 crc kubenswrapper[4846]: E0202 14:33:02.705330 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerName="extract-utilities" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.705342 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerName="extract-utilities" Feb 02 14:33:02 crc kubenswrapper[4846]: E0202 14:33:02.705358 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerName="registry-server" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.705365 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerName="registry-server" Feb 02 14:33:02 crc kubenswrapper[4846]: E0202 14:33:02.705460 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="registry-server" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.705498 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="registry-server" Feb 02 14:33:02 crc kubenswrapper[4846]: E0202 14:33:02.705520 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83c7cd73-a72d-4f4f-9b36-b7677290ecfa" containerName="nova-cell1-conductor-conductor" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.705529 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="83c7cd73-a72d-4f4f-9b36-b7677290ecfa" containerName="nova-cell1-conductor-conductor" Feb 02 14:33:02 crc kubenswrapper[4846]: E0202 14:33:02.705589 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bea9696-bedb-46b6-a458-51e7776a7922" containerName="neutron-dhcp-openstack-openstack-cell1" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.705599 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bea9696-bedb-46b6-a458-51e7776a7922" containerName="neutron-dhcp-openstack-openstack-cell1" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.706007 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0269b7a-21e4-493b-9930-261800880764" containerName="registry-server" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.706073 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="83c7cd73-a72d-4f4f-9b36-b7677290ecfa" containerName="nova-cell1-conductor-conductor" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.706103 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="60d812f8-e6e0-471f-8c61-015e86fbf6eb" containerName="registry-server" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.706121 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bea9696-bedb-46b6-a458-51e7776a7922" containerName="neutron-dhcp-openstack-openstack-cell1" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.707064 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.713248 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.717129 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.801763 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c241eff7-d5c6-4a5f-ae44-a28e16a4e609-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c241eff7-d5c6-4a5f-ae44-a28e16a4e609\") " pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.802005 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmfwq\" (UniqueName: \"kubernetes.io/projected/c241eff7-d5c6-4a5f-ae44-a28e16a4e609-kube-api-access-lmfwq\") pod \"nova-cell1-conductor-0\" (UID: \"c241eff7-d5c6-4a5f-ae44-a28e16a4e609\") " pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.802510 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c241eff7-d5c6-4a5f-ae44-a28e16a4e609-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c241eff7-d5c6-4a5f-ae44-a28e16a4e609\") " pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.904729 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c241eff7-d5c6-4a5f-ae44-a28e16a4e609-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c241eff7-d5c6-4a5f-ae44-a28e16a4e609\") " pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.904849 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c241eff7-d5c6-4a5f-ae44-a28e16a4e609-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c241eff7-d5c6-4a5f-ae44-a28e16a4e609\") " pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.904925 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmfwq\" (UniqueName: \"kubernetes.io/projected/c241eff7-d5c6-4a5f-ae44-a28e16a4e609-kube-api-access-lmfwq\") pod \"nova-cell1-conductor-0\" (UID: \"c241eff7-d5c6-4a5f-ae44-a28e16a4e609\") " pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.910048 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c241eff7-d5c6-4a5f-ae44-a28e16a4e609-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c241eff7-d5c6-4a5f-ae44-a28e16a4e609\") " pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.910441 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c241eff7-d5c6-4a5f-ae44-a28e16a4e609-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c241eff7-d5c6-4a5f-ae44-a28e16a4e609\") " pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:02 crc kubenswrapper[4846]: I0202 14:33:02.926661 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmfwq\" (UniqueName: \"kubernetes.io/projected/c241eff7-d5c6-4a5f-ae44-a28e16a4e609-kube-api-access-lmfwq\") pod \"nova-cell1-conductor-0\" (UID: \"c241eff7-d5c6-4a5f-ae44-a28e16a4e609\") " pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:03 crc kubenswrapper[4846]: I0202 14:33:03.030151 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:03 crc kubenswrapper[4846]: I0202 14:33:03.436579 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83c7cd73-a72d-4f4f-9b36-b7677290ecfa" path="/var/lib/kubelet/pods/83c7cd73-a72d-4f4f-9b36-b7677290ecfa/volumes" Feb 02 14:33:03 crc kubenswrapper[4846]: I0202 14:33:03.561453 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 02 14:33:03 crc kubenswrapper[4846]: I0202 14:33:03.653069 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c241eff7-d5c6-4a5f-ae44-a28e16a4e609","Type":"ContainerStarted","Data":"25e349ebe28d22b907469792c9e0deaba3b640168831ba244008ad7148920ee0"} Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.024350 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.141522 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96n7m\" (UniqueName: \"kubernetes.io/projected/4b608189-194f-4aae-bd70-9c6ea33f2d44-kube-api-access-96n7m\") pod \"4b608189-194f-4aae-bd70-9c6ea33f2d44\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.141872 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-config-data\") pod \"4b608189-194f-4aae-bd70-9c6ea33f2d44\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.141937 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-combined-ca-bundle\") pod \"4b608189-194f-4aae-bd70-9c6ea33f2d44\" (UID: \"4b608189-194f-4aae-bd70-9c6ea33f2d44\") " Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.149572 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b608189-194f-4aae-bd70-9c6ea33f2d44-kube-api-access-96n7m" (OuterVolumeSpecName: "kube-api-access-96n7m") pod "4b608189-194f-4aae-bd70-9c6ea33f2d44" (UID: "4b608189-194f-4aae-bd70-9c6ea33f2d44"). InnerVolumeSpecName "kube-api-access-96n7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.181817 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b608189-194f-4aae-bd70-9c6ea33f2d44" (UID: "4b608189-194f-4aae-bd70-9c6ea33f2d44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.201749 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-config-data" (OuterVolumeSpecName: "config-data") pod "4b608189-194f-4aae-bd70-9c6ea33f2d44" (UID: "4b608189-194f-4aae-bd70-9c6ea33f2d44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.244044 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.244094 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b608189-194f-4aae-bd70-9c6ea33f2d44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.244106 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96n7m\" (UniqueName: \"kubernetes.io/projected/4b608189-194f-4aae-bd70-9c6ea33f2d44-kube-api-access-96n7m\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.667427 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c241eff7-d5c6-4a5f-ae44-a28e16a4e609","Type":"ContainerStarted","Data":"3785ec7372503ee8a4fc204c16d7590bbe36322a39829d71d1b8a82e9d261fad"} Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.668331 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.670992 4846 generic.go:334] "Generic (PLEG): container finished" podID="4b608189-194f-4aae-bd70-9c6ea33f2d44" containerID="143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb" exitCode=0 Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.671055 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4b608189-194f-4aae-bd70-9c6ea33f2d44","Type":"ContainerDied","Data":"143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb"} Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.671090 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4b608189-194f-4aae-bd70-9c6ea33f2d44","Type":"ContainerDied","Data":"16383ed8055255e688a36e31acf8ed4f437f6c216c4564fb304a79a9c67c39e2"} Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.671112 4846 scope.go:117] "RemoveContainer" containerID="143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.671259 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.826741 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.826723737 podStartE2EDuration="2.826723737s" podCreationTimestamp="2026-02-02 14:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 14:33:04.690576417 +0000 UTC m=+8615.919163290" watchObservedRunningTime="2026-02-02 14:33:04.826723737 +0000 UTC m=+8616.055310610" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.836756 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.866003 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.887728 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 14:33:04 crc kubenswrapper[4846]: E0202 14:33:04.888371 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b608189-194f-4aae-bd70-9c6ea33f2d44" containerName="nova-cell0-conductor-conductor" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.888393 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b608189-194f-4aae-bd70-9c6ea33f2d44" containerName="nova-cell0-conductor-conductor" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.888648 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b608189-194f-4aae-bd70-9c6ea33f2d44" containerName="nova-cell0-conductor-conductor" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.889397 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.895593 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.899395 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.917375 4846 scope.go:117] "RemoveContainer" containerID="143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb" Feb 02 14:33:04 crc kubenswrapper[4846]: E0202 14:33:04.917738 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb\": container with ID starting with 143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb not found: ID does not exist" containerID="143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.917764 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb"} err="failed to get container status \"143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb\": rpc error: code = NotFound desc = could not find container \"143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb\": container with ID starting with 143cd75ef3258f59ca67c6709b6c8c8daf143f21cbf640dc09823257db35b1fb not found: ID does not exist" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.970686 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e\") " pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.970814 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl2kk\" (UniqueName: \"kubernetes.io/projected/7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e-kube-api-access-kl2kk\") pod \"nova-cell0-conductor-0\" (UID: \"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e\") " pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:04 crc kubenswrapper[4846]: I0202 14:33:04.970855 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e\") " pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.073016 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e\") " pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.073228 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e\") " pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.074110 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl2kk\" (UniqueName: \"kubernetes.io/projected/7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e-kube-api-access-kl2kk\") pod \"nova-cell0-conductor-0\" (UID: \"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e\") " pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.082520 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e\") " pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.098330 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e\") " pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.100198 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl2kk\" (UniqueName: \"kubernetes.io/projected/7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e-kube-api-access-kl2kk\") pod \"nova-cell0-conductor-0\" (UID: \"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e\") " pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.231894 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.377609 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.378837 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.458369 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b608189-194f-4aae-bd70-9c6ea33f2d44" path="/var/lib/kubelet/pods/4b608189-194f-4aae-bd70-9c6ea33f2d44/volumes" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.484403 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-combined-ca-bundle\") pod \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.484447 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-config-data\") pod \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.484498 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-internal-tls-certs\") pod \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.484602 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-public-tls-certs\") pod \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.485332 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czqdn\" (UniqueName: \"kubernetes.io/projected/39fdb60a-d87e-438c-8de9-63e49fdda9b8-kube-api-access-czqdn\") pod \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.485429 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-config-data\") pod \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.485462 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpm5f\" (UniqueName: \"kubernetes.io/projected/23d3a089-c3dc-411d-a099-3ff80dd40d2d-kube-api-access-fpm5f\") pod \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.485510 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-nova-metadata-tls-certs\") pod \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.485564 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d3a089-c3dc-411d-a099-3ff80dd40d2d-logs\") pod \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\" (UID: \"23d3a089-c3dc-411d-a099-3ff80dd40d2d\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.485611 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-combined-ca-bundle\") pod \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.485695 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39fdb60a-d87e-438c-8de9-63e49fdda9b8-logs\") pod \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\" (UID: \"39fdb60a-d87e-438c-8de9-63e49fdda9b8\") " Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.488353 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39fdb60a-d87e-438c-8de9-63e49fdda9b8-logs" (OuterVolumeSpecName: "logs") pod "39fdb60a-d87e-438c-8de9-63e49fdda9b8" (UID: "39fdb60a-d87e-438c-8de9-63e49fdda9b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.490761 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23d3a089-c3dc-411d-a099-3ff80dd40d2d-logs" (OuterVolumeSpecName: "logs") pod "23d3a089-c3dc-411d-a099-3ff80dd40d2d" (UID: "23d3a089-c3dc-411d-a099-3ff80dd40d2d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.514935 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23d3a089-c3dc-411d-a099-3ff80dd40d2d-kube-api-access-fpm5f" (OuterVolumeSpecName: "kube-api-access-fpm5f") pod "23d3a089-c3dc-411d-a099-3ff80dd40d2d" (UID: "23d3a089-c3dc-411d-a099-3ff80dd40d2d"). InnerVolumeSpecName "kube-api-access-fpm5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.518689 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39fdb60a-d87e-438c-8de9-63e49fdda9b8-kube-api-access-czqdn" (OuterVolumeSpecName: "kube-api-access-czqdn") pod "39fdb60a-d87e-438c-8de9-63e49fdda9b8" (UID: "39fdb60a-d87e-438c-8de9-63e49fdda9b8"). InnerVolumeSpecName "kube-api-access-czqdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.542097 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23d3a089-c3dc-411d-a099-3ff80dd40d2d" (UID: "23d3a089-c3dc-411d-a099-3ff80dd40d2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.543545 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-config-data" (OuterVolumeSpecName: "config-data") pod "39fdb60a-d87e-438c-8de9-63e49fdda9b8" (UID: "39fdb60a-d87e-438c-8de9-63e49fdda9b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.558405 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39fdb60a-d87e-438c-8de9-63e49fdda9b8" (UID: "39fdb60a-d87e-438c-8de9-63e49fdda9b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.561986 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-config-data" (OuterVolumeSpecName: "config-data") pod "23d3a089-c3dc-411d-a099-3ff80dd40d2d" (UID: "23d3a089-c3dc-411d-a099-3ff80dd40d2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.597194 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czqdn\" (UniqueName: \"kubernetes.io/projected/39fdb60a-d87e-438c-8de9-63e49fdda9b8-kube-api-access-czqdn\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.597238 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.597252 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpm5f\" (UniqueName: \"kubernetes.io/projected/23d3a089-c3dc-411d-a099-3ff80dd40d2d-kube-api-access-fpm5f\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.597266 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d3a089-c3dc-411d-a099-3ff80dd40d2d-logs\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.597279 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.597291 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39fdb60a-d87e-438c-8de9-63e49fdda9b8-logs\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.597301 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.597312 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.609347 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "39fdb60a-d87e-438c-8de9-63e49fdda9b8" (UID: "39fdb60a-d87e-438c-8de9-63e49fdda9b8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.617518 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "23d3a089-c3dc-411d-a099-3ff80dd40d2d" (UID: "23d3a089-c3dc-411d-a099-3ff80dd40d2d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.631606 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "39fdb60a-d87e-438c-8de9-63e49fdda9b8" (UID: "39fdb60a-d87e-438c-8de9-63e49fdda9b8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.700164 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.700205 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3a089-c3dc-411d-a099-3ff80dd40d2d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.700220 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fdb60a-d87e-438c-8de9-63e49fdda9b8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.717883 4846 generic.go:334] "Generic (PLEG): container finished" podID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerID="34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb" exitCode=0 Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.718014 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.718150 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23d3a089-c3dc-411d-a099-3ff80dd40d2d","Type":"ContainerDied","Data":"34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb"} Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.718184 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"23d3a089-c3dc-411d-a099-3ff80dd40d2d","Type":"ContainerDied","Data":"389c37c72a1c48a690c38947ca2eae59fc7924f19c28c4edd35e18d3cb4d678f"} Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.718201 4846 scope.go:117] "RemoveContainer" containerID="34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.744272 4846 generic.go:334] "Generic (PLEG): container finished" podID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerID="42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c" exitCode=0 Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.744970 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"39fdb60a-d87e-438c-8de9-63e49fdda9b8","Type":"ContainerDied","Data":"42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c"} Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.745082 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"39fdb60a-d87e-438c-8de9-63e49fdda9b8","Type":"ContainerDied","Data":"f7b53f6664b4d1c22e2151def81b3e2de229c85f54cfba100ed9cdd7b2f467b3"} Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.745238 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.778898 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.800327 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.814151 4846 scope.go:117] "RemoveContainer" containerID="211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.835731 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 02 14:33:05 crc kubenswrapper[4846]: E0202 14:33:05.836259 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerName="nova-api-log" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.836277 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerName="nova-api-log" Feb 02 14:33:05 crc kubenswrapper[4846]: E0202 14:33:05.836289 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerName="nova-api-api" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.836297 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerName="nova-api-api" Feb 02 14:33:05 crc kubenswrapper[4846]: E0202 14:33:05.836317 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerName="nova-metadata-log" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.836327 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerName="nova-metadata-log" Feb 02 14:33:05 crc kubenswrapper[4846]: E0202 14:33:05.836341 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerName="nova-metadata-metadata" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.836349 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerName="nova-metadata-metadata" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.836688 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerName="nova-metadata-log" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.836744 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerName="nova-api-api" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.836770 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" containerName="nova-api-log" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.836784 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" containerName="nova-metadata-metadata" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.838184 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.843486 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.844014 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.869818 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.886361 4846 scope.go:117] "RemoveContainer" containerID="34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.886444 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 02 14:33:05 crc kubenswrapper[4846]: E0202 14:33:05.887041 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb\": container with ID starting with 34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb not found: ID does not exist" containerID="34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.887080 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb"} err="failed to get container status \"34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb\": rpc error: code = NotFound desc = could not find container \"34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb\": container with ID starting with 34628593381e208839e97a4ee0ddfe58105671532a0321c3ae2e8d0c91e9f2bb not found: ID does not exist" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.887117 4846 scope.go:117] "RemoveContainer" containerID="211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490" Feb 02 14:33:05 crc kubenswrapper[4846]: E0202 14:33:05.887922 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490\": container with ID starting with 211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490 not found: ID does not exist" containerID="211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.887965 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490"} err="failed to get container status \"211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490\": rpc error: code = NotFound desc = could not find container \"211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490\": container with ID starting with 211eadbcf774ecf87326b9a1ea3358bf9b684a5079482edc701b28beef266490 not found: ID does not exist" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.887981 4846 scope.go:117] "RemoveContainer" containerID="42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c" Feb 02 14:33:05 crc kubenswrapper[4846]: W0202 14:33:05.888144 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ab6bdf6_72fd_46d7_9c26_bb7a11063b5e.slice/crio-d6752f7efe2abf9a8b53366713bef9bcb32c592123363ce535dd671f0f23077d WatchSource:0}: Error finding container d6752f7efe2abf9a8b53366713bef9bcb32c592123363ce535dd671f0f23077d: Status 404 returned error can't find the container with id d6752f7efe2abf9a8b53366713bef9bcb32c592123363ce535dd671f0f23077d Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.899803 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.910449 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-logs\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.910520 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.910573 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg945\" (UniqueName: \"kubernetes.io/projected/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-kube-api-access-fg945\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.910646 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-config-data\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.910678 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.921371 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.938644 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.938839 4846 scope.go:117] "RemoveContainer" containerID="a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.941507 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.944037 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.944206 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.944327 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.952263 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.983618 4846 scope.go:117] "RemoveContainer" containerID="42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c" Feb 02 14:33:05 crc kubenswrapper[4846]: E0202 14:33:05.984144 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c\": container with ID starting with 42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c not found: ID does not exist" containerID="42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.984184 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c"} err="failed to get container status \"42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c\": rpc error: code = NotFound desc = could not find container \"42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c\": container with ID starting with 42b2c95d33a5a0c04e7323f62804f2aff2e43c6509aeb9a8dd8253f8203e335c not found: ID does not exist" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.984214 4846 scope.go:117] "RemoveContainer" containerID="a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3" Feb 02 14:33:05 crc kubenswrapper[4846]: E0202 14:33:05.987236 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3\": container with ID starting with a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3 not found: ID does not exist" containerID="a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3" Feb 02 14:33:05 crc kubenswrapper[4846]: I0202 14:33:05.987270 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3"} err="failed to get container status \"a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3\": rpc error: code = NotFound desc = could not find container \"a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3\": container with ID starting with a1c1a35893e567b474e09319b69a4f685936567f633a21ad490041a16ac3f0c3 not found: ID does not exist" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.014083 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-config-data\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.014565 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.014733 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-logs\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.014785 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.014851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg945\" (UniqueName: \"kubernetes.io/projected/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-kube-api-access-fg945\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.014927 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.014956 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-config-data\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.014991 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.015058 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wst5t\" (UniqueName: \"kubernetes.io/projected/d7dd5959-c237-42de-9578-e5450cfe6092-kube-api-access-wst5t\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.015110 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7dd5959-c237-42de-9578-e5450cfe6092-logs\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.015149 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-public-tls-certs\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.015151 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-logs\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.019465 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-config-data\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.021494 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.025482 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.036566 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg945\" (UniqueName: \"kubernetes.io/projected/f1b2b0df-c8bd-4dad-a9bc-a91edc63d246-kube-api-access-fg945\") pod \"nova-metadata-0\" (UID: \"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246\") " pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.120347 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wst5t\" (UniqueName: \"kubernetes.io/projected/d7dd5959-c237-42de-9578-e5450cfe6092-kube-api-access-wst5t\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.121990 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7dd5959-c237-42de-9578-e5450cfe6092-logs\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.122180 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-public-tls-certs\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.122511 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-config-data\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.122602 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7dd5959-c237-42de-9578-e5450cfe6092-logs\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.122665 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.123647 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.127582 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-config-data\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.127895 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.131807 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.142687 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wst5t\" (UniqueName: \"kubernetes.io/projected/d7dd5959-c237-42de-9578-e5450cfe6092-kube-api-access-wst5t\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.145745 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7dd5959-c237-42de-9578-e5450cfe6092-public-tls-certs\") pod \"nova-api-0\" (UID: \"d7dd5959-c237-42de-9578-e5450cfe6092\") " pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.175539 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.278060 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.539175 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.638327 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22qxf\" (UniqueName: \"kubernetes.io/projected/1977e95e-eac6-462d-b7fd-caacae087fee-kube-api-access-22qxf\") pod \"1977e95e-eac6-462d-b7fd-caacae087fee\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.638509 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-combined-ca-bundle\") pod \"1977e95e-eac6-462d-b7fd-caacae087fee\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.639924 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-config-data\") pod \"1977e95e-eac6-462d-b7fd-caacae087fee\" (UID: \"1977e95e-eac6-462d-b7fd-caacae087fee\") " Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.651865 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1977e95e-eac6-462d-b7fd-caacae087fee-kube-api-access-22qxf" (OuterVolumeSpecName: "kube-api-access-22qxf") pod "1977e95e-eac6-462d-b7fd-caacae087fee" (UID: "1977e95e-eac6-462d-b7fd-caacae087fee"). InnerVolumeSpecName "kube-api-access-22qxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.684309 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-config-data" (OuterVolumeSpecName: "config-data") pod "1977e95e-eac6-462d-b7fd-caacae087fee" (UID: "1977e95e-eac6-462d-b7fd-caacae087fee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.702265 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1977e95e-eac6-462d-b7fd-caacae087fee" (UID: "1977e95e-eac6-462d-b7fd-caacae087fee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.744823 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.744878 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1977e95e-eac6-462d-b7fd-caacae087fee-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.744888 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22qxf\" (UniqueName: \"kubernetes.io/projected/1977e95e-eac6-462d-b7fd-caacae087fee-kube-api-access-22qxf\") on node \"crc\" DevicePath \"\"" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.755058 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e","Type":"ContainerStarted","Data":"0027034b2c897d201c5134d594465dd4dedf70d383bf994ecccdff28187d8a07"} Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.755104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e","Type":"ContainerStarted","Data":"d6752f7efe2abf9a8b53366713bef9bcb32c592123363ce535dd671f0f23077d"} Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.755254 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.761861 4846 generic.go:334] "Generic (PLEG): container finished" podID="1977e95e-eac6-462d-b7fd-caacae087fee" containerID="861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923" exitCode=0 Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.761935 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.761953 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1977e95e-eac6-462d-b7fd-caacae087fee","Type":"ContainerDied","Data":"861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923"} Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.761978 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1977e95e-eac6-462d-b7fd-caacae087fee","Type":"ContainerDied","Data":"ef7013589d21129eb3f31931366d245b7cb70e18a51784ea043b1ad249742e49"} Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.761996 4846 scope.go:117] "RemoveContainer" containerID="861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.773392 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.7733739870000003 podStartE2EDuration="2.773373987s" podCreationTimestamp="2026-02-02 14:33:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 14:33:06.769053982 +0000 UTC m=+8617.997640845" watchObservedRunningTime="2026-02-02 14:33:06.773373987 +0000 UTC m=+8618.001960850" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.802705 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.822523 4846 scope.go:117] "RemoveContainer" containerID="861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923" Feb 02 14:33:06 crc kubenswrapper[4846]: E0202 14:33:06.823512 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923\": container with ID starting with 861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923 not found: ID does not exist" containerID="861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.823634 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923"} err="failed to get container status \"861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923\": rpc error: code = NotFound desc = could not find container \"861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923\": container with ID starting with 861b4101473290b55110c60ba5b6c8c6d0ac02595d4c2d5208cccff6dfc3c923 not found: ID does not exist" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.829380 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.856136 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 14:33:06 crc kubenswrapper[4846]: E0202 14:33:06.856919 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1977e95e-eac6-462d-b7fd-caacae087fee" containerName="nova-scheduler-scheduler" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.856947 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1977e95e-eac6-462d-b7fd-caacae087fee" containerName="nova-scheduler-scheduler" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.857225 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1977e95e-eac6-462d-b7fd-caacae087fee" containerName="nova-scheduler-scheduler" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.858087 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 14:33:06 crc kubenswrapper[4846]: W0202 14:33:06.869276 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1b2b0df_c8bd_4dad_a9bc_a91edc63d246.slice/crio-375f1e77d2e1b3da19e3e59bb4690c1fed695e67850151445b6600beeddfe1fb WatchSource:0}: Error finding container 375f1e77d2e1b3da19e3e59bb4690c1fed695e67850151445b6600beeddfe1fb: Status 404 returned error can't find the container with id 375f1e77d2e1b3da19e3e59bb4690c1fed695e67850151445b6600beeddfe1fb Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.869751 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.887176 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.894514 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.948495 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92pg5\" (UniqueName: \"kubernetes.io/projected/fe78d901-4a4a-4db5-9ef3-c15b7501cc11-kube-api-access-92pg5\") pod \"nova-scheduler-0\" (UID: \"fe78d901-4a4a-4db5-9ef3-c15b7501cc11\") " pod="openstack/nova-scheduler-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.948565 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe78d901-4a4a-4db5-9ef3-c15b7501cc11-config-data\") pod \"nova-scheduler-0\" (UID: \"fe78d901-4a4a-4db5-9ef3-c15b7501cc11\") " pod="openstack/nova-scheduler-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.951008 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe78d901-4a4a-4db5-9ef3-c15b7501cc11-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fe78d901-4a4a-4db5-9ef3-c15b7501cc11\") " pod="openstack/nova-scheduler-0" Feb 02 14:33:06 crc kubenswrapper[4846]: I0202 14:33:06.998272 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 02 14:33:07 crc kubenswrapper[4846]: W0202 14:33:07.013264 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7dd5959_c237_42de_9578_e5450cfe6092.slice/crio-d1f1c292de7fc6c8096346d69ca15cdcceffa3a0b2d4398ed427c168f432c17e WatchSource:0}: Error finding container d1f1c292de7fc6c8096346d69ca15cdcceffa3a0b2d4398ed427c168f432c17e: Status 404 returned error can't find the container with id d1f1c292de7fc6c8096346d69ca15cdcceffa3a0b2d4398ed427c168f432c17e Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.052510 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92pg5\" (UniqueName: \"kubernetes.io/projected/fe78d901-4a4a-4db5-9ef3-c15b7501cc11-kube-api-access-92pg5\") pod \"nova-scheduler-0\" (UID: \"fe78d901-4a4a-4db5-9ef3-c15b7501cc11\") " pod="openstack/nova-scheduler-0" Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.052574 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe78d901-4a4a-4db5-9ef3-c15b7501cc11-config-data\") pod \"nova-scheduler-0\" (UID: \"fe78d901-4a4a-4db5-9ef3-c15b7501cc11\") " pod="openstack/nova-scheduler-0" Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.052743 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe78d901-4a4a-4db5-9ef3-c15b7501cc11-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fe78d901-4a4a-4db5-9ef3-c15b7501cc11\") " pod="openstack/nova-scheduler-0" Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.057865 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe78d901-4a4a-4db5-9ef3-c15b7501cc11-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fe78d901-4a4a-4db5-9ef3-c15b7501cc11\") " pod="openstack/nova-scheduler-0" Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.059650 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe78d901-4a4a-4db5-9ef3-c15b7501cc11-config-data\") pod \"nova-scheduler-0\" (UID: \"fe78d901-4a4a-4db5-9ef3-c15b7501cc11\") " pod="openstack/nova-scheduler-0" Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.076268 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92pg5\" (UniqueName: \"kubernetes.io/projected/fe78d901-4a4a-4db5-9ef3-c15b7501cc11-kube-api-access-92pg5\") pod \"nova-scheduler-0\" (UID: \"fe78d901-4a4a-4db5-9ef3-c15b7501cc11\") " pod="openstack/nova-scheduler-0" Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.258053 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.449032 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1977e95e-eac6-462d-b7fd-caacae087fee" path="/var/lib/kubelet/pods/1977e95e-eac6-462d-b7fd-caacae087fee/volumes" Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.449955 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23d3a089-c3dc-411d-a099-3ff80dd40d2d" path="/var/lib/kubelet/pods/23d3a089-c3dc-411d-a099-3ff80dd40d2d/volumes" Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.452246 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39fdb60a-d87e-438c-8de9-63e49fdda9b8" path="/var/lib/kubelet/pods/39fdb60a-d87e-438c-8de9-63e49fdda9b8/volumes" Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.565748 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.800809 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fe78d901-4a4a-4db5-9ef3-c15b7501cc11","Type":"ContainerStarted","Data":"71d2f04700593c600afbb8d60ae0c6878ad25a9f7f9348854b967e18b27b5515"} Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.804721 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246","Type":"ContainerStarted","Data":"0118e1bee408f15f5bb468aebcabe6f513b4bf257ed62d36bb950cb9307d0bce"} Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.804761 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246","Type":"ContainerStarted","Data":"26e0602c23cd749f29a0cf4a3437d176968d88d56fbea74864db94c0081310f6"} Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.804772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1b2b0df-c8bd-4dad-a9bc-a91edc63d246","Type":"ContainerStarted","Data":"375f1e77d2e1b3da19e3e59bb4690c1fed695e67850151445b6600beeddfe1fb"} Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.808355 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7dd5959-c237-42de-9578-e5450cfe6092","Type":"ContainerStarted","Data":"bf216aa010187448afecf5473af105a8da5c91025a9aaa6e7cba68f55f010ebf"} Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.808414 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7dd5959-c237-42de-9578-e5450cfe6092","Type":"ContainerStarted","Data":"d1f1c292de7fc6c8096346d69ca15cdcceffa3a0b2d4398ed427c168f432c17e"} Feb 02 14:33:07 crc kubenswrapper[4846]: I0202 14:33:07.839306 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.8392860730000002 podStartE2EDuration="2.839286073s" podCreationTimestamp="2026-02-02 14:33:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 14:33:07.834907117 +0000 UTC m=+8619.063493980" watchObservedRunningTime="2026-02-02 14:33:07.839286073 +0000 UTC m=+8619.067872936" Feb 02 14:33:08 crc kubenswrapper[4846]: I0202 14:33:08.064380 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 02 14:33:08 crc kubenswrapper[4846]: I0202 14:33:08.825233 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fe78d901-4a4a-4db5-9ef3-c15b7501cc11","Type":"ContainerStarted","Data":"720370c272c93bd1b3fce97f638d60a0c0b85a4c20cde265f5eaecab119f87c3"} Feb 02 14:33:08 crc kubenswrapper[4846]: I0202 14:33:08.828787 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7dd5959-c237-42de-9578-e5450cfe6092","Type":"ContainerStarted","Data":"3507fa5bc52e450e01901f2d9579e04fbbb00035c4a13164f2630b7431bd186a"} Feb 02 14:33:08 crc kubenswrapper[4846]: I0202 14:33:08.859867 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.8598435159999998 podStartE2EDuration="2.859843516s" podCreationTimestamp="2026-02-02 14:33:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 14:33:08.844321339 +0000 UTC m=+8620.072908242" watchObservedRunningTime="2026-02-02 14:33:08.859843516 +0000 UTC m=+8620.088430379" Feb 02 14:33:08 crc kubenswrapper[4846]: I0202 14:33:08.893925 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.893905154 podStartE2EDuration="3.893905154s" podCreationTimestamp="2026-02-02 14:33:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 14:33:08.877636529 +0000 UTC m=+8620.106223392" watchObservedRunningTime="2026-02-02 14:33:08.893905154 +0000 UTC m=+8620.122492017" Feb 02 14:33:09 crc kubenswrapper[4846]: I0202 14:33:09.433360 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:33:09 crc kubenswrapper[4846]: E0202 14:33:09.433960 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:33:11 crc kubenswrapper[4846]: I0202 14:33:11.179583 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 14:33:11 crc kubenswrapper[4846]: I0202 14:33:11.179951 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 02 14:33:12 crc kubenswrapper[4846]: I0202 14:33:12.259032 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 02 14:33:15 crc kubenswrapper[4846]: I0202 14:33:15.261155 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 02 14:33:16 crc kubenswrapper[4846]: I0202 14:33:16.179249 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 14:33:16 crc kubenswrapper[4846]: I0202 14:33:16.179559 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 02 14:33:16 crc kubenswrapper[4846]: I0202 14:33:16.279306 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 14:33:16 crc kubenswrapper[4846]: I0202 14:33:16.279358 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 02 14:33:17 crc kubenswrapper[4846]: I0202 14:33:17.194047 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f1b2b0df-c8bd-4dad-a9bc-a91edc63d246" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 14:33:17 crc kubenswrapper[4846]: I0202 14:33:17.194084 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f1b2b0df-c8bd-4dad-a9bc-a91edc63d246" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 14:33:17 crc kubenswrapper[4846]: I0202 14:33:17.259183 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 02 14:33:17 crc kubenswrapper[4846]: I0202 14:33:17.297303 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d7dd5959-c237-42de-9578-e5450cfe6092" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 14:33:17 crc kubenswrapper[4846]: I0202 14:33:17.297351 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d7dd5959-c237-42de-9578-e5450cfe6092" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 02 14:33:17 crc kubenswrapper[4846]: I0202 14:33:17.298971 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 02 14:33:17 crc kubenswrapper[4846]: I0202 14:33:17.958698 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 02 14:33:21 crc kubenswrapper[4846]: I0202 14:33:21.426316 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:33:21 crc kubenswrapper[4846]: E0202 14:33:21.428366 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:33:26 crc kubenswrapper[4846]: I0202 14:33:26.188533 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 14:33:26 crc kubenswrapper[4846]: I0202 14:33:26.189483 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 02 14:33:26 crc kubenswrapper[4846]: I0202 14:33:26.199188 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 14:33:26 crc kubenswrapper[4846]: I0202 14:33:26.199982 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 02 14:33:26 crc kubenswrapper[4846]: I0202 14:33:26.304568 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 14:33:26 crc kubenswrapper[4846]: I0202 14:33:26.305157 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 14:33:26 crc kubenswrapper[4846]: I0202 14:33:26.306053 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 02 14:33:26 crc kubenswrapper[4846]: I0202 14:33:26.313508 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 14:33:26 crc kubenswrapper[4846]: I0202 14:33:26.995688 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 02 14:33:27 crc kubenswrapper[4846]: I0202 14:33:27.002341 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.213219 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x"] Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.215036 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.228654 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x"] Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.229202 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.229201 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dvqhw" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.229513 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.229605 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.229672 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.229751 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.229890 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.279624 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnxp2\" (UniqueName: \"kubernetes.io/projected/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-kube-api-access-lnxp2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.279699 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.279795 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.279874 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.279975 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.280022 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.280077 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.280184 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.280248 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.381735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.381818 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.381851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.381887 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.381947 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.381980 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.382014 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnxp2\" (UniqueName: \"kubernetes.io/projected/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-kube-api-access-lnxp2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.382041 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.382061 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.383196 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.388727 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.389469 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.389583 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.389846 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.390450 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.390818 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.397291 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.408679 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnxp2\" (UniqueName: \"kubernetes.io/projected/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-kube-api-access-lnxp2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:28 crc kubenswrapper[4846]: I0202 14:33:28.546437 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:33:29 crc kubenswrapper[4846]: W0202 14:33:29.123750 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96fbf45d_7c8a_48e4_9d49_c6498105c3a5.slice/crio-02e9dd0c31ac385f80c9e1a879a284635c41d2a3646e59a778906a1d34017160 WatchSource:0}: Error finding container 02e9dd0c31ac385f80c9e1a879a284635c41d2a3646e59a778906a1d34017160: Status 404 returned error can't find the container with id 02e9dd0c31ac385f80c9e1a879a284635c41d2a3646e59a778906a1d34017160 Feb 02 14:33:29 crc kubenswrapper[4846]: I0202 14:33:29.127792 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x"] Feb 02 14:33:29 crc kubenswrapper[4846]: I0202 14:33:29.744607 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 02 14:33:30 crc kubenswrapper[4846]: I0202 14:33:30.021562 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" event={"ID":"96fbf45d-7c8a-48e4-9d49-c6498105c3a5","Type":"ContainerStarted","Data":"02e9dd0c31ac385f80c9e1a879a284635c41d2a3646e59a778906a1d34017160"} Feb 02 14:33:31 crc kubenswrapper[4846]: I0202 14:33:31.032137 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" event={"ID":"96fbf45d-7c8a-48e4-9d49-c6498105c3a5","Type":"ContainerStarted","Data":"9ec82a9ca6c60679fc1afca1da93320d138fce0d85e0225e314fdb08061c2123"} Feb 02 14:33:31 crc kubenswrapper[4846]: I0202 14:33:31.065034 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" podStartSLOduration=2.4516075219999998 podStartE2EDuration="3.065013056s" podCreationTimestamp="2026-02-02 14:33:28 +0000 UTC" firstStartedPulling="2026-02-02 14:33:29.126378062 +0000 UTC m=+8640.354964925" lastFinishedPulling="2026-02-02 14:33:29.739783566 +0000 UTC m=+8640.968370459" observedRunningTime="2026-02-02 14:33:31.05122736 +0000 UTC m=+8642.279814263" watchObservedRunningTime="2026-02-02 14:33:31.065013056 +0000 UTC m=+8642.293599929" Feb 02 14:33:34 crc kubenswrapper[4846]: I0202 14:33:34.423862 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:33:34 crc kubenswrapper[4846]: E0202 14:33:34.424743 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:33:49 crc kubenswrapper[4846]: I0202 14:33:49.432172 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:33:49 crc kubenswrapper[4846]: E0202 14:33:49.432881 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:34:04 crc kubenswrapper[4846]: I0202 14:34:04.425819 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:34:04 crc kubenswrapper[4846]: E0202 14:34:04.426717 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:34:19 crc kubenswrapper[4846]: I0202 14:34:19.434985 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:34:19 crc kubenswrapper[4846]: E0202 14:34:19.435820 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:34:33 crc kubenswrapper[4846]: I0202 14:34:33.423673 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:34:33 crc kubenswrapper[4846]: E0202 14:34:33.424443 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:34:48 crc kubenswrapper[4846]: I0202 14:34:48.424209 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:34:48 crc kubenswrapper[4846]: E0202 14:34:48.425873 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:35:02 crc kubenswrapper[4846]: I0202 14:35:02.423920 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:35:02 crc kubenswrapper[4846]: I0202 14:35:02.981524 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"37bf28f1711c5fec0c8da9946daaf65e3b2d086105d1e1a5ccc396efadba69af"} Feb 02 14:36:19 crc kubenswrapper[4846]: I0202 14:36:19.729321 4846 generic.go:334] "Generic (PLEG): container finished" podID="96fbf45d-7c8a-48e4-9d49-c6498105c3a5" containerID="9ec82a9ca6c60679fc1afca1da93320d138fce0d85e0225e314fdb08061c2123" exitCode=0 Feb 02 14:36:19 crc kubenswrapper[4846]: I0202 14:36:19.729443 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" event={"ID":"96fbf45d-7c8a-48e4-9d49-c6498105c3a5","Type":"ContainerDied","Data":"9ec82a9ca6c60679fc1afca1da93320d138fce0d85e0225e314fdb08061c2123"} Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.214283 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.321754 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-1\") pod \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.321827 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-0\") pod \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.321936 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-inventory\") pod \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.321982 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnxp2\" (UniqueName: \"kubernetes.io/projected/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-kube-api-access-lnxp2\") pod \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.322127 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-0\") pod \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.322272 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-combined-ca-bundle\") pod \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.322320 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-1\") pod \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.322352 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-ssh-key-openstack-cell1\") pod \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.322412 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cells-global-config-0\") pod \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\" (UID: \"96fbf45d-7c8a-48e4-9d49-c6498105c3a5\") " Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.333063 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-kube-api-access-lnxp2" (OuterVolumeSpecName: "kube-api-access-lnxp2") pod "96fbf45d-7c8a-48e4-9d49-c6498105c3a5" (UID: "96fbf45d-7c8a-48e4-9d49-c6498105c3a5"). InnerVolumeSpecName "kube-api-access-lnxp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.335735 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "96fbf45d-7c8a-48e4-9d49-c6498105c3a5" (UID: "96fbf45d-7c8a-48e4-9d49-c6498105c3a5"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.352944 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "96fbf45d-7c8a-48e4-9d49-c6498105c3a5" (UID: "96fbf45d-7c8a-48e4-9d49-c6498105c3a5"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.361012 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-inventory" (OuterVolumeSpecName: "inventory") pod "96fbf45d-7c8a-48e4-9d49-c6498105c3a5" (UID: "96fbf45d-7c8a-48e4-9d49-c6498105c3a5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.361981 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "96fbf45d-7c8a-48e4-9d49-c6498105c3a5" (UID: "96fbf45d-7c8a-48e4-9d49-c6498105c3a5"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.362201 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "96fbf45d-7c8a-48e4-9d49-c6498105c3a5" (UID: "96fbf45d-7c8a-48e4-9d49-c6498105c3a5"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.363211 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "96fbf45d-7c8a-48e4-9d49-c6498105c3a5" (UID: "96fbf45d-7c8a-48e4-9d49-c6498105c3a5"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.367939 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "96fbf45d-7c8a-48e4-9d49-c6498105c3a5" (UID: "96fbf45d-7c8a-48e4-9d49-c6498105c3a5"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.370612 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "96fbf45d-7c8a-48e4-9d49-c6498105c3a5" (UID: "96fbf45d-7c8a-48e4-9d49-c6498105c3a5"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.425450 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.425485 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.425500 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.425512 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.425525 4846 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.425538 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.425550 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-inventory\") on node \"crc\" DevicePath \"\"" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.425562 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnxp2\" (UniqueName: \"kubernetes.io/projected/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-kube-api-access-lnxp2\") on node \"crc\" DevicePath \"\"" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.425573 4846 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/96fbf45d-7c8a-48e4-9d49-c6498105c3a5-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.749849 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" event={"ID":"96fbf45d-7c8a-48e4-9d49-c6498105c3a5","Type":"ContainerDied","Data":"02e9dd0c31ac385f80c9e1a879a284635c41d2a3646e59a778906a1d34017160"} Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.750114 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02e9dd0c31ac385f80c9e1a879a284635c41d2a3646e59a778906a1d34017160" Feb 02 14:36:21 crc kubenswrapper[4846]: I0202 14:36:21.749895 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x" Feb 02 14:37:30 crc kubenswrapper[4846]: I0202 14:37:30.479293 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:37:30 crc kubenswrapper[4846]: I0202 14:37:30.480205 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:37:45 crc kubenswrapper[4846]: I0202 14:37:45.939507 4846 scope.go:117] "RemoveContainer" containerID="49bdbf3f4cf2ae8546b9a02e6d4609eafbd31dd57d331e8fb54596cd67e00b2c" Feb 02 14:37:45 crc kubenswrapper[4846]: I0202 14:37:45.963315 4846 scope.go:117] "RemoveContainer" containerID="e2d9c33be71cace6558a9ee40a1c4d8e91a134f9cf1636175a0ebfaf7dfbc247" Feb 02 14:37:45 crc kubenswrapper[4846]: I0202 14:37:45.991181 4846 scope.go:117] "RemoveContainer" containerID="f9adf5db1b07b9c1602fc2f61a65900ca0fb2450e4dff2a6af35d3aadcd34ff9" Feb 02 14:38:00 crc kubenswrapper[4846]: I0202 14:38:00.479135 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:38:00 crc kubenswrapper[4846]: I0202 14:38:00.479700 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:38:11 crc kubenswrapper[4846]: I0202 14:38:11.981188 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Feb 02 14:38:11 crc kubenswrapper[4846]: I0202 14:38:11.981937 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="a5799c70-3207-43d8-a2d6-495aa207445c" containerName="adoption" containerID="cri-o://48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e" gracePeriod=30 Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.535908 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cnvdv"] Feb 02 14:38:22 crc kubenswrapper[4846]: E0202 14:38:22.537215 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96fbf45d-7c8a-48e4-9d49-c6498105c3a5" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.537249 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="96fbf45d-7c8a-48e4-9d49-c6498105c3a5" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.537519 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="96fbf45d-7c8a-48e4-9d49-c6498105c3a5" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.539512 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.552734 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cnvdv"] Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.657592 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8b6l\" (UniqueName: \"kubernetes.io/projected/214db85a-630c-47dd-b366-cb4fc4b93673-kube-api-access-w8b6l\") pod \"certified-operators-cnvdv\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.657841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-catalog-content\") pod \"certified-operators-cnvdv\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.657888 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-utilities\") pod \"certified-operators-cnvdv\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.760718 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-catalog-content\") pod \"certified-operators-cnvdv\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.760790 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-utilities\") pod \"certified-operators-cnvdv\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.760949 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8b6l\" (UniqueName: \"kubernetes.io/projected/214db85a-630c-47dd-b366-cb4fc4b93673-kube-api-access-w8b6l\") pod \"certified-operators-cnvdv\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.761249 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-catalog-content\") pod \"certified-operators-cnvdv\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.761459 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-utilities\") pod \"certified-operators-cnvdv\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.781016 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8b6l\" (UniqueName: \"kubernetes.io/projected/214db85a-630c-47dd-b366-cb4fc4b93673-kube-api-access-w8b6l\") pod \"certified-operators-cnvdv\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:22 crc kubenswrapper[4846]: I0202 14:38:22.863108 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:23 crc kubenswrapper[4846]: I0202 14:38:23.480367 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cnvdv"] Feb 02 14:38:23 crc kubenswrapper[4846]: I0202 14:38:23.960869 4846 generic.go:334] "Generic (PLEG): container finished" podID="214db85a-630c-47dd-b366-cb4fc4b93673" containerID="1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb" exitCode=0 Feb 02 14:38:23 crc kubenswrapper[4846]: I0202 14:38:23.960927 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnvdv" event={"ID":"214db85a-630c-47dd-b366-cb4fc4b93673","Type":"ContainerDied","Data":"1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb"} Feb 02 14:38:23 crc kubenswrapper[4846]: I0202 14:38:23.961154 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnvdv" event={"ID":"214db85a-630c-47dd-b366-cb4fc4b93673","Type":"ContainerStarted","Data":"f7dbe6b30436f10413b0b71f3c7c43fc3a647cfbf3a11ba82679c27d63469d0f"} Feb 02 14:38:23 crc kubenswrapper[4846]: I0202 14:38:23.963349 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 14:38:25 crc kubenswrapper[4846]: I0202 14:38:25.984792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnvdv" event={"ID":"214db85a-630c-47dd-b366-cb4fc4b93673","Type":"ContainerStarted","Data":"276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6"} Feb 02 14:38:26 crc kubenswrapper[4846]: I0202 14:38:26.996834 4846 generic.go:334] "Generic (PLEG): container finished" podID="214db85a-630c-47dd-b366-cb4fc4b93673" containerID="276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6" exitCode=0 Feb 02 14:38:26 crc kubenswrapper[4846]: I0202 14:38:26.996885 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnvdv" event={"ID":"214db85a-630c-47dd-b366-cb4fc4b93673","Type":"ContainerDied","Data":"276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6"} Feb 02 14:38:28 crc kubenswrapper[4846]: I0202 14:38:28.007284 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnvdv" event={"ID":"214db85a-630c-47dd-b366-cb4fc4b93673","Type":"ContainerStarted","Data":"65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d"} Feb 02 14:38:28 crc kubenswrapper[4846]: I0202 14:38:28.030044 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cnvdv" podStartSLOduration=2.54087978 podStartE2EDuration="6.030025963s" podCreationTimestamp="2026-02-02 14:38:22 +0000 UTC" firstStartedPulling="2026-02-02 14:38:23.96299812 +0000 UTC m=+8935.191585003" lastFinishedPulling="2026-02-02 14:38:27.452144333 +0000 UTC m=+8938.680731186" observedRunningTime="2026-02-02 14:38:28.026036726 +0000 UTC m=+8939.254623589" watchObservedRunningTime="2026-02-02 14:38:28.030025963 +0000 UTC m=+8939.258612826" Feb 02 14:38:30 crc kubenswrapper[4846]: I0202 14:38:30.478892 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:38:30 crc kubenswrapper[4846]: I0202 14:38:30.479350 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:38:30 crc kubenswrapper[4846]: I0202 14:38:30.479393 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 14:38:30 crc kubenswrapper[4846]: I0202 14:38:30.480256 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"37bf28f1711c5fec0c8da9946daaf65e3b2d086105d1e1a5ccc396efadba69af"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 14:38:30 crc kubenswrapper[4846]: I0202 14:38:30.480322 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://37bf28f1711c5fec0c8da9946daaf65e3b2d086105d1e1a5ccc396efadba69af" gracePeriod=600 Feb 02 14:38:32 crc kubenswrapper[4846]: I0202 14:38:32.048363 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="37bf28f1711c5fec0c8da9946daaf65e3b2d086105d1e1a5ccc396efadba69af" exitCode=0 Feb 02 14:38:32 crc kubenswrapper[4846]: I0202 14:38:32.048414 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"37bf28f1711c5fec0c8da9946daaf65e3b2d086105d1e1a5ccc396efadba69af"} Feb 02 14:38:32 crc kubenswrapper[4846]: I0202 14:38:32.048750 4846 scope.go:117] "RemoveContainer" containerID="264d20539b72829650f431b079702b436e3bf5bcc5a477288a4a05ec2ecf5d03" Feb 02 14:38:32 crc kubenswrapper[4846]: I0202 14:38:32.863993 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:32 crc kubenswrapper[4846]: I0202 14:38:32.864481 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:32 crc kubenswrapper[4846]: I0202 14:38:32.939398 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:33 crc kubenswrapper[4846]: I0202 14:38:33.065656 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288"} Feb 02 14:38:33 crc kubenswrapper[4846]: I0202 14:38:33.111450 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:33 crc kubenswrapper[4846]: I0202 14:38:33.179546 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cnvdv"] Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.083478 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cnvdv" podUID="214db85a-630c-47dd-b366-cb4fc4b93673" containerName="registry-server" containerID="cri-o://65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d" gracePeriod=2 Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.613133 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.663650 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8b6l\" (UniqueName: \"kubernetes.io/projected/214db85a-630c-47dd-b366-cb4fc4b93673-kube-api-access-w8b6l\") pod \"214db85a-630c-47dd-b366-cb4fc4b93673\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.663824 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-catalog-content\") pod \"214db85a-630c-47dd-b366-cb4fc4b93673\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.664085 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-utilities\") pod \"214db85a-630c-47dd-b366-cb4fc4b93673\" (UID: \"214db85a-630c-47dd-b366-cb4fc4b93673\") " Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.665042 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-utilities" (OuterVolumeSpecName: "utilities") pod "214db85a-630c-47dd-b366-cb4fc4b93673" (UID: "214db85a-630c-47dd-b366-cb4fc4b93673"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.669569 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/214db85a-630c-47dd-b366-cb4fc4b93673-kube-api-access-w8b6l" (OuterVolumeSpecName: "kube-api-access-w8b6l") pod "214db85a-630c-47dd-b366-cb4fc4b93673" (UID: "214db85a-630c-47dd-b366-cb4fc4b93673"). InnerVolumeSpecName "kube-api-access-w8b6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.712653 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "214db85a-630c-47dd-b366-cb4fc4b93673" (UID: "214db85a-630c-47dd-b366-cb4fc4b93673"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.766955 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.766999 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8b6l\" (UniqueName: \"kubernetes.io/projected/214db85a-630c-47dd-b366-cb4fc4b93673-kube-api-access-w8b6l\") on node \"crc\" DevicePath \"\"" Feb 02 14:38:35 crc kubenswrapper[4846]: I0202 14:38:35.767014 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/214db85a-630c-47dd-b366-cb4fc4b93673-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.093578 4846 generic.go:334] "Generic (PLEG): container finished" podID="214db85a-630c-47dd-b366-cb4fc4b93673" containerID="65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d" exitCode=0 Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.093651 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnvdv" event={"ID":"214db85a-630c-47dd-b366-cb4fc4b93673","Type":"ContainerDied","Data":"65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d"} Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.093907 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnvdv" event={"ID":"214db85a-630c-47dd-b366-cb4fc4b93673","Type":"ContainerDied","Data":"f7dbe6b30436f10413b0b71f3c7c43fc3a647cfbf3a11ba82679c27d63469d0f"} Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.093938 4846 scope.go:117] "RemoveContainer" containerID="65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d" Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.093701 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cnvdv" Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.127506 4846 scope.go:117] "RemoveContainer" containerID="276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6" Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.133881 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cnvdv"] Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.145082 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cnvdv"] Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.157423 4846 scope.go:117] "RemoveContainer" containerID="1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb" Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.209812 4846 scope.go:117] "RemoveContainer" containerID="65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d" Feb 02 14:38:36 crc kubenswrapper[4846]: E0202 14:38:36.211483 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d\": container with ID starting with 65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d not found: ID does not exist" containerID="65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d" Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.211520 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d"} err="failed to get container status \"65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d\": rpc error: code = NotFound desc = could not find container \"65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d\": container with ID starting with 65fd7b681efd236425f30cb5f3bf77e4595eeca7de51b41a84f2470c8b2c450d not found: ID does not exist" Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.211548 4846 scope.go:117] "RemoveContainer" containerID="276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6" Feb 02 14:38:36 crc kubenswrapper[4846]: E0202 14:38:36.211989 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6\": container with ID starting with 276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6 not found: ID does not exist" containerID="276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6" Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.212020 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6"} err="failed to get container status \"276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6\": rpc error: code = NotFound desc = could not find container \"276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6\": container with ID starting with 276e677807895dd27bc4ec3f8fe0bb30beb889d1d9f0a9085d229729755503b6 not found: ID does not exist" Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.212040 4846 scope.go:117] "RemoveContainer" containerID="1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb" Feb 02 14:38:36 crc kubenswrapper[4846]: E0202 14:38:36.212334 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb\": container with ID starting with 1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb not found: ID does not exist" containerID="1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb" Feb 02 14:38:36 crc kubenswrapper[4846]: I0202 14:38:36.212361 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb"} err="failed to get container status \"1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb\": rpc error: code = NotFound desc = could not find container \"1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb\": container with ID starting with 1c82c5366c0c6bfe6e8961485446ebfcda8828f2c21a3b5382282590ddc6aacb not found: ID does not exist" Feb 02 14:38:37 crc kubenswrapper[4846]: I0202 14:38:37.436487 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="214db85a-630c-47dd-b366-cb4fc4b93673" path="/var/lib/kubelet/pods/214db85a-630c-47dd-b366-cb4fc4b93673/volumes" Feb 02 14:38:42 crc kubenswrapper[4846]: I0202 14:38:42.695167 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Feb 02 14:38:42 crc kubenswrapper[4846]: I0202 14:38:42.730535 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\") pod \"a5799c70-3207-43d8-a2d6-495aa207445c\" (UID: \"a5799c70-3207-43d8-a2d6-495aa207445c\") " Feb 02 14:38:42 crc kubenswrapper[4846]: I0202 14:38:42.730758 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phz4d\" (UniqueName: \"kubernetes.io/projected/a5799c70-3207-43d8-a2d6-495aa207445c-kube-api-access-phz4d\") pod \"a5799c70-3207-43d8-a2d6-495aa207445c\" (UID: \"a5799c70-3207-43d8-a2d6-495aa207445c\") " Feb 02 14:38:42 crc kubenswrapper[4846]: I0202 14:38:42.737882 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5799c70-3207-43d8-a2d6-495aa207445c-kube-api-access-phz4d" (OuterVolumeSpecName: "kube-api-access-phz4d") pod "a5799c70-3207-43d8-a2d6-495aa207445c" (UID: "a5799c70-3207-43d8-a2d6-495aa207445c"). InnerVolumeSpecName "kube-api-access-phz4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:38:42 crc kubenswrapper[4846]: I0202 14:38:42.753000 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00af887f-a708-4da5-80d9-12cc047e0fb3" (OuterVolumeSpecName: "mariadb-data") pod "a5799c70-3207-43d8-a2d6-495aa207445c" (UID: "a5799c70-3207-43d8-a2d6-495aa207445c"). InnerVolumeSpecName "pvc-00af887f-a708-4da5-80d9-12cc047e0fb3". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 14:38:42 crc kubenswrapper[4846]: I0202 14:38:42.834340 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\") on node \"crc\" " Feb 02 14:38:42 crc kubenswrapper[4846]: I0202 14:38:42.834434 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phz4d\" (UniqueName: \"kubernetes.io/projected/a5799c70-3207-43d8-a2d6-495aa207445c-kube-api-access-phz4d\") on node \"crc\" DevicePath \"\"" Feb 02 14:38:42 crc kubenswrapper[4846]: I0202 14:38:42.861934 4846 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 02 14:38:42 crc kubenswrapper[4846]: I0202 14:38:42.862122 4846 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-00af887f-a708-4da5-80d9-12cc047e0fb3" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00af887f-a708-4da5-80d9-12cc047e0fb3") on node "crc" Feb 02 14:38:42 crc kubenswrapper[4846]: I0202 14:38:42.936340 4846 reconciler_common.go:293] "Volume detached for volume \"pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00af887f-a708-4da5-80d9-12cc047e0fb3\") on node \"crc\" DevicePath \"\"" Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.167056 4846 generic.go:334] "Generic (PLEG): container finished" podID="a5799c70-3207-43d8-a2d6-495aa207445c" containerID="48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e" exitCode=137 Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.167103 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"a5799c70-3207-43d8-a2d6-495aa207445c","Type":"ContainerDied","Data":"48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e"} Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.167121 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.167131 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"a5799c70-3207-43d8-a2d6-495aa207445c","Type":"ContainerDied","Data":"62d327e126d7cc804be6a8d3f410f5dcca719f1bd24600110532d97bf69c6e32"} Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.167151 4846 scope.go:117] "RemoveContainer" containerID="48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e" Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.191750 4846 scope.go:117] "RemoveContainer" containerID="48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e" Feb 02 14:38:43 crc kubenswrapper[4846]: E0202 14:38:43.192685 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e\": container with ID starting with 48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e not found: ID does not exist" containerID="48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e" Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.192778 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e"} err="failed to get container status \"48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e\": rpc error: code = NotFound desc = could not find container \"48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e\": container with ID starting with 48e06df965cceadc90d911a8152adb1effe2b86ee2aaaa72580cb8ca2320545e not found: ID does not exist" Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.213971 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.223449 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.437424 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5799c70-3207-43d8-a2d6-495aa207445c" path="/var/lib/kubelet/pods/a5799c70-3207-43d8-a2d6-495aa207445c/volumes" Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.778105 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Feb 02 14:38:43 crc kubenswrapper[4846]: I0202 14:38:43.778315 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="9f811429-9750-4803-b9e6-931cac4216aa" containerName="adoption" containerID="cri-o://a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4" gracePeriod=30 Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.278007 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.423340 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\") pod \"9f811429-9750-4803-b9e6-931cac4216aa\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.423606 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/9f811429-9750-4803-b9e6-931cac4216aa-ovn-data-cert\") pod \"9f811429-9750-4803-b9e6-931cac4216aa\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.423685 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcpmf\" (UniqueName: \"kubernetes.io/projected/9f811429-9750-4803-b9e6-931cac4216aa-kube-api-access-lcpmf\") pod \"9f811429-9750-4803-b9e6-931cac4216aa\" (UID: \"9f811429-9750-4803-b9e6-931cac4216aa\") " Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.432219 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f811429-9750-4803-b9e6-931cac4216aa-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "9f811429-9750-4803-b9e6-931cac4216aa" (UID: "9f811429-9750-4803-b9e6-931cac4216aa"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.435259 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f811429-9750-4803-b9e6-931cac4216aa-kube-api-access-lcpmf" (OuterVolumeSpecName: "kube-api-access-lcpmf") pod "9f811429-9750-4803-b9e6-931cac4216aa" (UID: "9f811429-9750-4803-b9e6-931cac4216aa"). InnerVolumeSpecName "kube-api-access-lcpmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.461441 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a" (OuterVolumeSpecName: "ovn-data") pod "9f811429-9750-4803-b9e6-931cac4216aa" (UID: "9f811429-9750-4803-b9e6-931cac4216aa"). InnerVolumeSpecName "pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.501658 4846 generic.go:334] "Generic (PLEG): container finished" podID="9f811429-9750-4803-b9e6-931cac4216aa" containerID="a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4" exitCode=137 Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.501711 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"9f811429-9750-4803-b9e6-931cac4216aa","Type":"ContainerDied","Data":"a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4"} Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.501723 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.501742 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"9f811429-9750-4803-b9e6-931cac4216aa","Type":"ContainerDied","Data":"9b01284e644d1c116e5e7cb8ecb4fc3a4ea449f93092dadd648c3dbc2f6b49ab"} Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.501761 4846 scope.go:117] "RemoveContainer" containerID="a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.533493 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\") on node \"crc\" " Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.533540 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/9f811429-9750-4803-b9e6-931cac4216aa-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.533555 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcpmf\" (UniqueName: \"kubernetes.io/projected/9f811429-9750-4803-b9e6-931cac4216aa-kube-api-access-lcpmf\") on node \"crc\" DevicePath \"\"" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.558177 4846 scope.go:117] "RemoveContainer" containerID="a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4" Feb 02 14:39:14 crc kubenswrapper[4846]: E0202 14:39:14.558761 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4\": container with ID starting with a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4 not found: ID does not exist" containerID="a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.558812 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4"} err="failed to get container status \"a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4\": rpc error: code = NotFound desc = could not find container \"a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4\": container with ID starting with a192126c45785892a653540264d22f4d369425881513248dd79adfd92bdc2ce4 not found: ID does not exist" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.563989 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.565167 4846 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.565409 4846 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a") on node "crc" Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.574408 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Feb 02 14:39:14 crc kubenswrapper[4846]: I0202 14:39:14.635930 4846 reconciler_common.go:293] "Volume detached for volume \"pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c95aab54-195f-46f4-8cc9-b719b05b3a4a\") on node \"crc\" DevicePath \"\"" Feb 02 14:39:15 crc kubenswrapper[4846]: I0202 14:39:15.439491 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f811429-9750-4803-b9e6-931cac4216aa" path="/var/lib/kubelet/pods/9f811429-9750-4803-b9e6-931cac4216aa/volumes" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.204903 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bbhbc/must-gather-ps8nn"] Feb 02 14:40:18 crc kubenswrapper[4846]: E0202 14:40:18.205800 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f811429-9750-4803-b9e6-931cac4216aa" containerName="adoption" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.205814 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f811429-9750-4803-b9e6-931cac4216aa" containerName="adoption" Feb 02 14:40:18 crc kubenswrapper[4846]: E0202 14:40:18.205827 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="214db85a-630c-47dd-b366-cb4fc4b93673" containerName="extract-utilities" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.205833 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="214db85a-630c-47dd-b366-cb4fc4b93673" containerName="extract-utilities" Feb 02 14:40:18 crc kubenswrapper[4846]: E0202 14:40:18.205844 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="214db85a-630c-47dd-b366-cb4fc4b93673" containerName="registry-server" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.205851 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="214db85a-630c-47dd-b366-cb4fc4b93673" containerName="registry-server" Feb 02 14:40:18 crc kubenswrapper[4846]: E0202 14:40:18.205886 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="214db85a-630c-47dd-b366-cb4fc4b93673" containerName="extract-content" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.205891 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="214db85a-630c-47dd-b366-cb4fc4b93673" containerName="extract-content" Feb 02 14:40:18 crc kubenswrapper[4846]: E0202 14:40:18.205904 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5799c70-3207-43d8-a2d6-495aa207445c" containerName="adoption" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.205911 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5799c70-3207-43d8-a2d6-495aa207445c" containerName="adoption" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.206103 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5799c70-3207-43d8-a2d6-495aa207445c" containerName="adoption" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.206127 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f811429-9750-4803-b9e6-931cac4216aa" containerName="adoption" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.206135 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="214db85a-630c-47dd-b366-cb4fc4b93673" containerName="registry-server" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.207332 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/must-gather-ps8nn" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.211150 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-bbhbc"/"default-dockercfg-4hh4d" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.211252 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bbhbc"/"openshift-service-ca.crt" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.212504 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bbhbc"/"kube-root-ca.crt" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.231943 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bbhbc/must-gather-ps8nn"] Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.245762 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-must-gather-output\") pod \"must-gather-ps8nn\" (UID: \"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a\") " pod="openshift-must-gather-bbhbc/must-gather-ps8nn" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.245834 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8hl9\" (UniqueName: \"kubernetes.io/projected/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-kube-api-access-s8hl9\") pod \"must-gather-ps8nn\" (UID: \"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a\") " pod="openshift-must-gather-bbhbc/must-gather-ps8nn" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.347722 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-must-gather-output\") pod \"must-gather-ps8nn\" (UID: \"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a\") " pod="openshift-must-gather-bbhbc/must-gather-ps8nn" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.347786 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8hl9\" (UniqueName: \"kubernetes.io/projected/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-kube-api-access-s8hl9\") pod \"must-gather-ps8nn\" (UID: \"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a\") " pod="openshift-must-gather-bbhbc/must-gather-ps8nn" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.349251 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-must-gather-output\") pod \"must-gather-ps8nn\" (UID: \"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a\") " pod="openshift-must-gather-bbhbc/must-gather-ps8nn" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.368591 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8hl9\" (UniqueName: \"kubernetes.io/projected/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-kube-api-access-s8hl9\") pod \"must-gather-ps8nn\" (UID: \"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a\") " pod="openshift-must-gather-bbhbc/must-gather-ps8nn" Feb 02 14:40:18 crc kubenswrapper[4846]: I0202 14:40:18.544712 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/must-gather-ps8nn" Feb 02 14:40:19 crc kubenswrapper[4846]: I0202 14:40:19.023166 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bbhbc/must-gather-ps8nn"] Feb 02 14:40:19 crc kubenswrapper[4846]: I0202 14:40:19.176506 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/must-gather-ps8nn" event={"ID":"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a","Type":"ContainerStarted","Data":"e03cda34c54f4fc2ae78a1e8ee6a01e4e58a3f3b92eefeb3ab2e5361c578b62c"} Feb 02 14:40:25 crc kubenswrapper[4846]: I0202 14:40:25.290237 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/must-gather-ps8nn" event={"ID":"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a","Type":"ContainerStarted","Data":"5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24"} Feb 02 14:40:26 crc kubenswrapper[4846]: I0202 14:40:26.302181 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/must-gather-ps8nn" event={"ID":"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a","Type":"ContainerStarted","Data":"568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc"} Feb 02 14:40:26 crc kubenswrapper[4846]: I0202 14:40:26.328349 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bbhbc/must-gather-ps8nn" podStartSLOduration=3.464998067 podStartE2EDuration="8.32832717s" podCreationTimestamp="2026-02-02 14:40:18 +0000 UTC" firstStartedPulling="2026-02-02 14:40:19.032063234 +0000 UTC m=+9050.260650097" lastFinishedPulling="2026-02-02 14:40:23.895392327 +0000 UTC m=+9055.123979200" observedRunningTime="2026-02-02 14:40:26.317048446 +0000 UTC m=+9057.545635309" watchObservedRunningTime="2026-02-02 14:40:26.32832717 +0000 UTC m=+9057.556914033" Feb 02 14:40:29 crc kubenswrapper[4846]: I0202 14:40:29.223778 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bbhbc/crc-debug-qw5qx"] Feb 02 14:40:29 crc kubenswrapper[4846]: I0202 14:40:29.226591 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" Feb 02 14:40:29 crc kubenswrapper[4846]: I0202 14:40:29.329008 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6lzl\" (UniqueName: \"kubernetes.io/projected/7f755d3f-6310-4119-9022-833aa19e5cac-kube-api-access-m6lzl\") pod \"crc-debug-qw5qx\" (UID: \"7f755d3f-6310-4119-9022-833aa19e5cac\") " pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" Feb 02 14:40:29 crc kubenswrapper[4846]: I0202 14:40:29.329344 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f755d3f-6310-4119-9022-833aa19e5cac-host\") pod \"crc-debug-qw5qx\" (UID: \"7f755d3f-6310-4119-9022-833aa19e5cac\") " pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" Feb 02 14:40:29 crc kubenswrapper[4846]: I0202 14:40:29.431046 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f755d3f-6310-4119-9022-833aa19e5cac-host\") pod \"crc-debug-qw5qx\" (UID: \"7f755d3f-6310-4119-9022-833aa19e5cac\") " pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" Feb 02 14:40:29 crc kubenswrapper[4846]: I0202 14:40:29.431459 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6lzl\" (UniqueName: \"kubernetes.io/projected/7f755d3f-6310-4119-9022-833aa19e5cac-kube-api-access-m6lzl\") pod \"crc-debug-qw5qx\" (UID: \"7f755d3f-6310-4119-9022-833aa19e5cac\") " pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" Feb 02 14:40:29 crc kubenswrapper[4846]: I0202 14:40:29.431258 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f755d3f-6310-4119-9022-833aa19e5cac-host\") pod \"crc-debug-qw5qx\" (UID: \"7f755d3f-6310-4119-9022-833aa19e5cac\") " pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" Feb 02 14:40:29 crc kubenswrapper[4846]: I0202 14:40:29.458146 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6lzl\" (UniqueName: \"kubernetes.io/projected/7f755d3f-6310-4119-9022-833aa19e5cac-kube-api-access-m6lzl\") pod \"crc-debug-qw5qx\" (UID: \"7f755d3f-6310-4119-9022-833aa19e5cac\") " pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" Feb 02 14:40:29 crc kubenswrapper[4846]: I0202 14:40:29.545168 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" Feb 02 14:40:29 crc kubenswrapper[4846]: W0202 14:40:29.611469 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f755d3f_6310_4119_9022_833aa19e5cac.slice/crio-41f8185c2794c40993ed1315ba683fdc3bbf66ad579ef926c7a0ec87faa77599 WatchSource:0}: Error finding container 41f8185c2794c40993ed1315ba683fdc3bbf66ad579ef926c7a0ec87faa77599: Status 404 returned error can't find the container with id 41f8185c2794c40993ed1315ba683fdc3bbf66ad579ef926c7a0ec87faa77599 Feb 02 14:40:30 crc kubenswrapper[4846]: I0202 14:40:30.351170 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" event={"ID":"7f755d3f-6310-4119-9022-833aa19e5cac","Type":"ContainerStarted","Data":"41f8185c2794c40993ed1315ba683fdc3bbf66ad579ef926c7a0ec87faa77599"} Feb 02 14:40:31 crc kubenswrapper[4846]: E0202 14:40:31.087521 4846 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.223:48112->38.102.83.223:44473: read tcp 38.102.83.223:48112->38.102.83.223:44473: read: connection reset by peer Feb 02 14:40:43 crc kubenswrapper[4846]: I0202 14:40:43.490518 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" event={"ID":"7f755d3f-6310-4119-9022-833aa19e5cac","Type":"ContainerStarted","Data":"680c8d7e76ec2bc066e8b78c1d2768031f249da9bddf734871cdef15fa790971"} Feb 02 14:40:43 crc kubenswrapper[4846]: I0202 14:40:43.513593 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" podStartSLOduration=1.5386885769999998 podStartE2EDuration="14.513575999s" podCreationTimestamp="2026-02-02 14:40:29 +0000 UTC" firstStartedPulling="2026-02-02 14:40:29.613452722 +0000 UTC m=+9060.842039585" lastFinishedPulling="2026-02-02 14:40:42.588340134 +0000 UTC m=+9073.816927007" observedRunningTime="2026-02-02 14:40:43.50781757 +0000 UTC m=+9074.736404433" watchObservedRunningTime="2026-02-02 14:40:43.513575999 +0000 UTC m=+9074.742162862" Feb 02 14:41:00 crc kubenswrapper[4846]: I0202 14:41:00.479116 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:41:00 crc kubenswrapper[4846]: I0202 14:41:00.479683 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:41:30 crc kubenswrapper[4846]: I0202 14:41:30.478837 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:41:30 crc kubenswrapper[4846]: I0202 14:41:30.479397 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:41:32 crc kubenswrapper[4846]: I0202 14:41:32.958577 4846 generic.go:334] "Generic (PLEG): container finished" podID="7f755d3f-6310-4119-9022-833aa19e5cac" containerID="680c8d7e76ec2bc066e8b78c1d2768031f249da9bddf734871cdef15fa790971" exitCode=0 Feb 02 14:41:32 crc kubenswrapper[4846]: I0202 14:41:32.958711 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" event={"ID":"7f755d3f-6310-4119-9022-833aa19e5cac","Type":"ContainerDied","Data":"680c8d7e76ec2bc066e8b78c1d2768031f249da9bddf734871cdef15fa790971"} Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.112289 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.167999 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bbhbc/crc-debug-qw5qx"] Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.180367 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bbhbc/crc-debug-qw5qx"] Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.185464 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f755d3f-6310-4119-9022-833aa19e5cac-host\") pod \"7f755d3f-6310-4119-9022-833aa19e5cac\" (UID: \"7f755d3f-6310-4119-9022-833aa19e5cac\") " Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.185598 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f755d3f-6310-4119-9022-833aa19e5cac-host" (OuterVolumeSpecName: "host") pod "7f755d3f-6310-4119-9022-833aa19e5cac" (UID: "7f755d3f-6310-4119-9022-833aa19e5cac"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.185829 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6lzl\" (UniqueName: \"kubernetes.io/projected/7f755d3f-6310-4119-9022-833aa19e5cac-kube-api-access-m6lzl\") pod \"7f755d3f-6310-4119-9022-833aa19e5cac\" (UID: \"7f755d3f-6310-4119-9022-833aa19e5cac\") " Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.186290 4846 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7f755d3f-6310-4119-9022-833aa19e5cac-host\") on node \"crc\" DevicePath \"\"" Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.201096 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f755d3f-6310-4119-9022-833aa19e5cac-kube-api-access-m6lzl" (OuterVolumeSpecName: "kube-api-access-m6lzl") pod "7f755d3f-6310-4119-9022-833aa19e5cac" (UID: "7f755d3f-6310-4119-9022-833aa19e5cac"). InnerVolumeSpecName "kube-api-access-m6lzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.290156 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6lzl\" (UniqueName: \"kubernetes.io/projected/7f755d3f-6310-4119-9022-833aa19e5cac-kube-api-access-m6lzl\") on node \"crc\" DevicePath \"\"" Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.982803 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41f8185c2794c40993ed1315ba683fdc3bbf66ad579ef926c7a0ec87faa77599" Feb 02 14:41:34 crc kubenswrapper[4846]: I0202 14:41:34.982864 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-qw5qx" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.362514 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bbhbc/crc-debug-ck8nv"] Feb 02 14:41:35 crc kubenswrapper[4846]: E0202 14:41:35.362957 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f755d3f-6310-4119-9022-833aa19e5cac" containerName="container-00" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.362968 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f755d3f-6310-4119-9022-833aa19e5cac" containerName="container-00" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.363185 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f755d3f-6310-4119-9022-833aa19e5cac" containerName="container-00" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.364137 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.421448 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95d8f151-14bd-41b8-b975-c0c999633b89-host\") pod \"crc-debug-ck8nv\" (UID: \"95d8f151-14bd-41b8-b975-c0c999633b89\") " pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.421494 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ghqx\" (UniqueName: \"kubernetes.io/projected/95d8f151-14bd-41b8-b975-c0c999633b89-kube-api-access-4ghqx\") pod \"crc-debug-ck8nv\" (UID: \"95d8f151-14bd-41b8-b975-c0c999633b89\") " pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.436218 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f755d3f-6310-4119-9022-833aa19e5cac" path="/var/lib/kubelet/pods/7f755d3f-6310-4119-9022-833aa19e5cac/volumes" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.523077 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95d8f151-14bd-41b8-b975-c0c999633b89-host\") pod \"crc-debug-ck8nv\" (UID: \"95d8f151-14bd-41b8-b975-c0c999633b89\") " pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.523169 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ghqx\" (UniqueName: \"kubernetes.io/projected/95d8f151-14bd-41b8-b975-c0c999633b89-kube-api-access-4ghqx\") pod \"crc-debug-ck8nv\" (UID: \"95d8f151-14bd-41b8-b975-c0c999633b89\") " pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.523243 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95d8f151-14bd-41b8-b975-c0c999633b89-host\") pod \"crc-debug-ck8nv\" (UID: \"95d8f151-14bd-41b8-b975-c0c999633b89\") " pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.541310 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ghqx\" (UniqueName: \"kubernetes.io/projected/95d8f151-14bd-41b8-b975-c0c999633b89-kube-api-access-4ghqx\") pod \"crc-debug-ck8nv\" (UID: \"95d8f151-14bd-41b8-b975-c0c999633b89\") " pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.681794 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" Feb 02 14:41:35 crc kubenswrapper[4846]: W0202 14:41:35.739647 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95d8f151_14bd_41b8_b975_c0c999633b89.slice/crio-fcdadbc0d9d55b73e714328dbe13d974a649b060fd18b802c32092263449e8c5 WatchSource:0}: Error finding container fcdadbc0d9d55b73e714328dbe13d974a649b060fd18b802c32092263449e8c5: Status 404 returned error can't find the container with id fcdadbc0d9d55b73e714328dbe13d974a649b060fd18b802c32092263449e8c5 Feb 02 14:41:35 crc kubenswrapper[4846]: I0202 14:41:35.993010 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" event={"ID":"95d8f151-14bd-41b8-b975-c0c999633b89","Type":"ContainerStarted","Data":"fcdadbc0d9d55b73e714328dbe13d974a649b060fd18b802c32092263449e8c5"} Feb 02 14:41:37 crc kubenswrapper[4846]: I0202 14:41:37.003796 4846 generic.go:334] "Generic (PLEG): container finished" podID="95d8f151-14bd-41b8-b975-c0c999633b89" containerID="78dcc4bea3ae34c2d813a0c1993a0b74eebd80094c6a374969aad16c73e3577f" exitCode=0 Feb 02 14:41:37 crc kubenswrapper[4846]: I0202 14:41:37.003874 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" event={"ID":"95d8f151-14bd-41b8-b975-c0c999633b89","Type":"ContainerDied","Data":"78dcc4bea3ae34c2d813a0c1993a0b74eebd80094c6a374969aad16c73e3577f"} Feb 02 14:41:37 crc kubenswrapper[4846]: I0202 14:41:37.408748 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bbhbc/crc-debug-ck8nv"] Feb 02 14:41:37 crc kubenswrapper[4846]: I0202 14:41:37.417413 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bbhbc/crc-debug-ck8nv"] Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.125449 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.174886 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ghqx\" (UniqueName: \"kubernetes.io/projected/95d8f151-14bd-41b8-b975-c0c999633b89-kube-api-access-4ghqx\") pod \"95d8f151-14bd-41b8-b975-c0c999633b89\" (UID: \"95d8f151-14bd-41b8-b975-c0c999633b89\") " Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.175181 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95d8f151-14bd-41b8-b975-c0c999633b89-host\") pod \"95d8f151-14bd-41b8-b975-c0c999633b89\" (UID: \"95d8f151-14bd-41b8-b975-c0c999633b89\") " Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.175348 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95d8f151-14bd-41b8-b975-c0c999633b89-host" (OuterVolumeSpecName: "host") pod "95d8f151-14bd-41b8-b975-c0c999633b89" (UID: "95d8f151-14bd-41b8-b975-c0c999633b89"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.175767 4846 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/95d8f151-14bd-41b8-b975-c0c999633b89-host\") on node \"crc\" DevicePath \"\"" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.183081 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95d8f151-14bd-41b8-b975-c0c999633b89-kube-api-access-4ghqx" (OuterVolumeSpecName: "kube-api-access-4ghqx") pod "95d8f151-14bd-41b8-b975-c0c999633b89" (UID: "95d8f151-14bd-41b8-b975-c0c999633b89"). InnerVolumeSpecName "kube-api-access-4ghqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.277431 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ghqx\" (UniqueName: \"kubernetes.io/projected/95d8f151-14bd-41b8-b975-c0c999633b89-kube-api-access-4ghqx\") on node \"crc\" DevicePath \"\"" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.637465 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xjcgs"] Feb 02 14:41:38 crc kubenswrapper[4846]: E0202 14:41:38.637923 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d8f151-14bd-41b8-b975-c0c999633b89" containerName="container-00" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.637939 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d8f151-14bd-41b8-b975-c0c999633b89" containerName="container-00" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.638208 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="95d8f151-14bd-41b8-b975-c0c999633b89" containerName="container-00" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.639814 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.652021 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xjcgs"] Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.690308 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-utilities\") pod \"community-operators-xjcgs\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.690456 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg9dx\" (UniqueName: \"kubernetes.io/projected/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-kube-api-access-tg9dx\") pod \"community-operators-xjcgs\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.690765 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-catalog-content\") pod \"community-operators-xjcgs\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.792571 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-catalog-content\") pod \"community-operators-xjcgs\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.793006 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-utilities\") pod \"community-operators-xjcgs\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.793068 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg9dx\" (UniqueName: \"kubernetes.io/projected/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-kube-api-access-tg9dx\") pod \"community-operators-xjcgs\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.793066 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-catalog-content\") pod \"community-operators-xjcgs\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.793249 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-utilities\") pod \"community-operators-xjcgs\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.811471 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg9dx\" (UniqueName: \"kubernetes.io/projected/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-kube-api-access-tg9dx\") pod \"community-operators-xjcgs\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:38 crc kubenswrapper[4846]: I0202 14:41:38.965831 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.094945 4846 scope.go:117] "RemoveContainer" containerID="78dcc4bea3ae34c2d813a0c1993a0b74eebd80094c6a374969aad16c73e3577f" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.095120 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-ck8nv" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.300105 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bbhbc/crc-debug-8bzhm"] Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.301893 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.407690 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf8x9\" (UniqueName: \"kubernetes.io/projected/90e27bc8-77ac-492a-91d1-014da7374fb0-kube-api-access-xf8x9\") pod \"crc-debug-8bzhm\" (UID: \"90e27bc8-77ac-492a-91d1-014da7374fb0\") " pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.407836 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90e27bc8-77ac-492a-91d1-014da7374fb0-host\") pod \"crc-debug-8bzhm\" (UID: \"90e27bc8-77ac-492a-91d1-014da7374fb0\") " pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.450970 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95d8f151-14bd-41b8-b975-c0c999633b89" path="/var/lib/kubelet/pods/95d8f151-14bd-41b8-b975-c0c999633b89/volumes" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.509849 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf8x9\" (UniqueName: \"kubernetes.io/projected/90e27bc8-77ac-492a-91d1-014da7374fb0-kube-api-access-xf8x9\") pod \"crc-debug-8bzhm\" (UID: \"90e27bc8-77ac-492a-91d1-014da7374fb0\") " pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.510202 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90e27bc8-77ac-492a-91d1-014da7374fb0-host\") pod \"crc-debug-8bzhm\" (UID: \"90e27bc8-77ac-492a-91d1-014da7374fb0\") " pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.510829 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90e27bc8-77ac-492a-91d1-014da7374fb0-host\") pod \"crc-debug-8bzhm\" (UID: \"90e27bc8-77ac-492a-91d1-014da7374fb0\") " pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.533297 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf8x9\" (UniqueName: \"kubernetes.io/projected/90e27bc8-77ac-492a-91d1-014da7374fb0-kube-api-access-xf8x9\") pod \"crc-debug-8bzhm\" (UID: \"90e27bc8-77ac-492a-91d1-014da7374fb0\") " pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.622925 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" Feb 02 14:41:39 crc kubenswrapper[4846]: I0202 14:41:39.804492 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xjcgs"] Feb 02 14:41:39 crc kubenswrapper[4846]: W0202 14:41:39.812439 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97d78223_9f5c_4fb0_9d09_cae9a688a8c9.slice/crio-3c7c41c334af2ad1e13e95b3364fcdd7730676c7ca72146558f629091e5a4f5d WatchSource:0}: Error finding container 3c7c41c334af2ad1e13e95b3364fcdd7730676c7ca72146558f629091e5a4f5d: Status 404 returned error can't find the container with id 3c7c41c334af2ad1e13e95b3364fcdd7730676c7ca72146558f629091e5a4f5d Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.107193 4846 generic.go:334] "Generic (PLEG): container finished" podID="90e27bc8-77ac-492a-91d1-014da7374fb0" containerID="8b0e9f386e7b65bc43bd726fd2ba91d9d06ffd67ef0ed0188aac714a3da971f3" exitCode=0 Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.107268 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" event={"ID":"90e27bc8-77ac-492a-91d1-014da7374fb0","Type":"ContainerDied","Data":"8b0e9f386e7b65bc43bd726fd2ba91d9d06ffd67ef0ed0188aac714a3da971f3"} Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.107298 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" event={"ID":"90e27bc8-77ac-492a-91d1-014da7374fb0","Type":"ContainerStarted","Data":"b8e66e4a2d6ee48ba800a49d261fe93f9880b8fec71dd61b1521fa81c5c90426"} Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.109216 4846 generic.go:334] "Generic (PLEG): container finished" podID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerID="f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339" exitCode=0 Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.109253 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjcgs" event={"ID":"97d78223-9f5c-4fb0-9d09-cae9a688a8c9","Type":"ContainerDied","Data":"f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339"} Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.109286 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjcgs" event={"ID":"97d78223-9f5c-4fb0-9d09-cae9a688a8c9","Type":"ContainerStarted","Data":"3c7c41c334af2ad1e13e95b3364fcdd7730676c7ca72146558f629091e5a4f5d"} Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.161894 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bbhbc/crc-debug-8bzhm"] Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.172683 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bbhbc/crc-debug-8bzhm"] Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.421755 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8jjp5"] Feb 02 14:41:40 crc kubenswrapper[4846]: E0202 14:41:40.422780 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e27bc8-77ac-492a-91d1-014da7374fb0" containerName="container-00" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.422799 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e27bc8-77ac-492a-91d1-014da7374fb0" containerName="container-00" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.423120 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="90e27bc8-77ac-492a-91d1-014da7374fb0" containerName="container-00" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.425344 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.432538 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8jjp5"] Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.535149 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-utilities\") pod \"redhat-operators-8jjp5\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.535916 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx6w9\" (UniqueName: \"kubernetes.io/projected/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-kube-api-access-nx6w9\") pod \"redhat-operators-8jjp5\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.536604 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-catalog-content\") pod \"redhat-operators-8jjp5\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.638903 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-catalog-content\") pod \"redhat-operators-8jjp5\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.639029 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-utilities\") pod \"redhat-operators-8jjp5\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.639074 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx6w9\" (UniqueName: \"kubernetes.io/projected/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-kube-api-access-nx6w9\") pod \"redhat-operators-8jjp5\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.639585 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-catalog-content\") pod \"redhat-operators-8jjp5\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.639746 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-utilities\") pod \"redhat-operators-8jjp5\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.670126 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx6w9\" (UniqueName: \"kubernetes.io/projected/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-kube-api-access-nx6w9\") pod \"redhat-operators-8jjp5\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:40 crc kubenswrapper[4846]: I0202 14:41:40.780828 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:41:41 crc kubenswrapper[4846]: I0202 14:41:41.195341 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" Feb 02 14:41:41 crc kubenswrapper[4846]: I0202 14:41:41.251715 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf8x9\" (UniqueName: \"kubernetes.io/projected/90e27bc8-77ac-492a-91d1-014da7374fb0-kube-api-access-xf8x9\") pod \"90e27bc8-77ac-492a-91d1-014da7374fb0\" (UID: \"90e27bc8-77ac-492a-91d1-014da7374fb0\") " Feb 02 14:41:41 crc kubenswrapper[4846]: I0202 14:41:41.251925 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90e27bc8-77ac-492a-91d1-014da7374fb0-host\") pod \"90e27bc8-77ac-492a-91d1-014da7374fb0\" (UID: \"90e27bc8-77ac-492a-91d1-014da7374fb0\") " Feb 02 14:41:41 crc kubenswrapper[4846]: I0202 14:41:41.252731 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/90e27bc8-77ac-492a-91d1-014da7374fb0-host" (OuterVolumeSpecName: "host") pod "90e27bc8-77ac-492a-91d1-014da7374fb0" (UID: "90e27bc8-77ac-492a-91d1-014da7374fb0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 14:41:41 crc kubenswrapper[4846]: I0202 14:41:41.269148 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8jjp5"] Feb 02 14:41:41 crc kubenswrapper[4846]: I0202 14:41:41.273927 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90e27bc8-77ac-492a-91d1-014da7374fb0-kube-api-access-xf8x9" (OuterVolumeSpecName: "kube-api-access-xf8x9") pod "90e27bc8-77ac-492a-91d1-014da7374fb0" (UID: "90e27bc8-77ac-492a-91d1-014da7374fb0"). InnerVolumeSpecName "kube-api-access-xf8x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:41:41 crc kubenswrapper[4846]: I0202 14:41:41.353928 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf8x9\" (UniqueName: \"kubernetes.io/projected/90e27bc8-77ac-492a-91d1-014da7374fb0-kube-api-access-xf8x9\") on node \"crc\" DevicePath \"\"" Feb 02 14:41:41 crc kubenswrapper[4846]: I0202 14:41:41.353963 4846 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90e27bc8-77ac-492a-91d1-014da7374fb0-host\") on node \"crc\" DevicePath \"\"" Feb 02 14:41:41 crc kubenswrapper[4846]: I0202 14:41:41.457355 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90e27bc8-77ac-492a-91d1-014da7374fb0" path="/var/lib/kubelet/pods/90e27bc8-77ac-492a-91d1-014da7374fb0/volumes" Feb 02 14:41:42 crc kubenswrapper[4846]: I0202 14:41:42.128437 4846 generic.go:334] "Generic (PLEG): container finished" podID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerID="a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c" exitCode=0 Feb 02 14:41:42 crc kubenswrapper[4846]: I0202 14:41:42.128537 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jjp5" event={"ID":"8bf34605-f5f8-429c-8cfb-cac96fed3aa0","Type":"ContainerDied","Data":"a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c"} Feb 02 14:41:42 crc kubenswrapper[4846]: I0202 14:41:42.128858 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jjp5" event={"ID":"8bf34605-f5f8-429c-8cfb-cac96fed3aa0","Type":"ContainerStarted","Data":"bede4de80fe43ef56cf4f2d29243b67d0e229c81d8d62cac7fe4410eca15d5d2"} Feb 02 14:41:42 crc kubenswrapper[4846]: I0202 14:41:42.131342 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjcgs" event={"ID":"97d78223-9f5c-4fb0-9d09-cae9a688a8c9","Type":"ContainerStarted","Data":"3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e"} Feb 02 14:41:42 crc kubenswrapper[4846]: I0202 14:41:42.133742 4846 scope.go:117] "RemoveContainer" containerID="8b0e9f386e7b65bc43bd726fd2ba91d9d06ffd67ef0ed0188aac714a3da971f3" Feb 02 14:41:42 crc kubenswrapper[4846]: I0202 14:41:42.133850 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/crc-debug-8bzhm" Feb 02 14:41:44 crc kubenswrapper[4846]: I0202 14:41:44.154158 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jjp5" event={"ID":"8bf34605-f5f8-429c-8cfb-cac96fed3aa0","Type":"ContainerStarted","Data":"950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9"} Feb 02 14:41:48 crc kubenswrapper[4846]: I0202 14:41:48.212493 4846 generic.go:334] "Generic (PLEG): container finished" podID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerID="3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e" exitCode=0 Feb 02 14:41:48 crc kubenswrapper[4846]: I0202 14:41:48.212587 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjcgs" event={"ID":"97d78223-9f5c-4fb0-9d09-cae9a688a8c9","Type":"ContainerDied","Data":"3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e"} Feb 02 14:41:50 crc kubenswrapper[4846]: I0202 14:41:50.233807 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjcgs" event={"ID":"97d78223-9f5c-4fb0-9d09-cae9a688a8c9","Type":"ContainerStarted","Data":"971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e"} Feb 02 14:41:50 crc kubenswrapper[4846]: I0202 14:41:50.268039 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xjcgs" podStartSLOduration=3.344628889 podStartE2EDuration="12.268013418s" podCreationTimestamp="2026-02-02 14:41:38 +0000 UTC" firstStartedPulling="2026-02-02 14:41:40.11112899 +0000 UTC m=+9131.339715863" lastFinishedPulling="2026-02-02 14:41:49.034513529 +0000 UTC m=+9140.263100392" observedRunningTime="2026-02-02 14:41:50.253697229 +0000 UTC m=+9141.482284112" watchObservedRunningTime="2026-02-02 14:41:50.268013418 +0000 UTC m=+9141.496600301" Feb 02 14:41:55 crc kubenswrapper[4846]: I0202 14:41:55.280721 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jjp5" event={"ID":"8bf34605-f5f8-429c-8cfb-cac96fed3aa0","Type":"ContainerDied","Data":"950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9"} Feb 02 14:41:55 crc kubenswrapper[4846]: I0202 14:41:55.280727 4846 generic.go:334] "Generic (PLEG): container finished" podID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerID="950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9" exitCode=0 Feb 02 14:41:56 crc kubenswrapper[4846]: I0202 14:41:56.296274 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jjp5" event={"ID":"8bf34605-f5f8-429c-8cfb-cac96fed3aa0","Type":"ContainerStarted","Data":"52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf"} Feb 02 14:41:56 crc kubenswrapper[4846]: I0202 14:41:56.314484 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8jjp5" podStartSLOduration=2.4253188789999998 podStartE2EDuration="16.314465171s" podCreationTimestamp="2026-02-02 14:41:40 +0000 UTC" firstStartedPulling="2026-02-02 14:41:42.131002455 +0000 UTC m=+9133.359589308" lastFinishedPulling="2026-02-02 14:41:56.020148697 +0000 UTC m=+9147.248735600" observedRunningTime="2026-02-02 14:41:56.314009651 +0000 UTC m=+9147.542596524" watchObservedRunningTime="2026-02-02 14:41:56.314465171 +0000 UTC m=+9147.543052044" Feb 02 14:41:58 crc kubenswrapper[4846]: I0202 14:41:58.966544 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:41:58 crc kubenswrapper[4846]: I0202 14:41:58.966875 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:42:00 crc kubenswrapper[4846]: I0202 14:42:00.012685 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-xjcgs" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerName="registry-server" probeResult="failure" output=< Feb 02 14:42:00 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:42:00 crc kubenswrapper[4846]: > Feb 02 14:42:00 crc kubenswrapper[4846]: I0202 14:42:00.480197 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:42:00 crc kubenswrapper[4846]: I0202 14:42:00.480522 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:42:00 crc kubenswrapper[4846]: I0202 14:42:00.480572 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 14:42:00 crc kubenswrapper[4846]: I0202 14:42:00.481354 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 14:42:00 crc kubenswrapper[4846]: I0202 14:42:00.481410 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" gracePeriod=600 Feb 02 14:42:00 crc kubenswrapper[4846]: E0202 14:42:00.612662 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:42:00 crc kubenswrapper[4846]: I0202 14:42:00.781828 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:42:00 crc kubenswrapper[4846]: I0202 14:42:00.782185 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:42:01 crc kubenswrapper[4846]: I0202 14:42:01.352274 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" exitCode=0 Feb 02 14:42:01 crc kubenswrapper[4846]: I0202 14:42:01.352351 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288"} Feb 02 14:42:01 crc kubenswrapper[4846]: I0202 14:42:01.353007 4846 scope.go:117] "RemoveContainer" containerID="37bf28f1711c5fec0c8da9946daaf65e3b2d086105d1e1a5ccc396efadba69af" Feb 02 14:42:01 crc kubenswrapper[4846]: I0202 14:42:01.354517 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:42:01 crc kubenswrapper[4846]: E0202 14:42:01.354965 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:42:01 crc kubenswrapper[4846]: I0202 14:42:01.831670 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8jjp5" podUID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerName="registry-server" probeResult="failure" output=< Feb 02 14:42:01 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:42:01 crc kubenswrapper[4846]: > Feb 02 14:42:10 crc kubenswrapper[4846]: I0202 14:42:10.014232 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-xjcgs" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerName="registry-server" probeResult="failure" output=< Feb 02 14:42:10 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Feb 02 14:42:10 crc kubenswrapper[4846]: > Feb 02 14:42:10 crc kubenswrapper[4846]: I0202 14:42:10.830592 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:42:10 crc kubenswrapper[4846]: I0202 14:42:10.883296 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:42:11 crc kubenswrapper[4846]: I0202 14:42:11.072305 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8jjp5"] Feb 02 14:42:12 crc kubenswrapper[4846]: I0202 14:42:12.501039 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8jjp5" podUID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerName="registry-server" containerID="cri-o://52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf" gracePeriod=2 Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.027352 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.139917 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-catalog-content\") pod \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.140167 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx6w9\" (UniqueName: \"kubernetes.io/projected/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-kube-api-access-nx6w9\") pod \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.140201 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-utilities\") pod \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\" (UID: \"8bf34605-f5f8-429c-8cfb-cac96fed3aa0\") " Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.141987 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-utilities" (OuterVolumeSpecName: "utilities") pod "8bf34605-f5f8-429c-8cfb-cac96fed3aa0" (UID: "8bf34605-f5f8-429c-8cfb-cac96fed3aa0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.145382 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-kube-api-access-nx6w9" (OuterVolumeSpecName: "kube-api-access-nx6w9") pod "8bf34605-f5f8-429c-8cfb-cac96fed3aa0" (UID: "8bf34605-f5f8-429c-8cfb-cac96fed3aa0"). InnerVolumeSpecName "kube-api-access-nx6w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.243418 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx6w9\" (UniqueName: \"kubernetes.io/projected/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-kube-api-access-nx6w9\") on node \"crc\" DevicePath \"\"" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.243938 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.290558 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bf34605-f5f8-429c-8cfb-cac96fed3aa0" (UID: "8bf34605-f5f8-429c-8cfb-cac96fed3aa0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.345895 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf34605-f5f8-429c-8cfb-cac96fed3aa0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.514690 4846 generic.go:334] "Generic (PLEG): container finished" podID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerID="52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf" exitCode=0 Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.514739 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jjp5" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.514739 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jjp5" event={"ID":"8bf34605-f5f8-429c-8cfb-cac96fed3aa0","Type":"ContainerDied","Data":"52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf"} Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.514915 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jjp5" event={"ID":"8bf34605-f5f8-429c-8cfb-cac96fed3aa0","Type":"ContainerDied","Data":"bede4de80fe43ef56cf4f2d29243b67d0e229c81d8d62cac7fe4410eca15d5d2"} Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.514956 4846 scope.go:117] "RemoveContainer" containerID="52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.544702 4846 scope.go:117] "RemoveContainer" containerID="950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.551548 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8jjp5"] Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.561666 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8jjp5"] Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.570025 4846 scope.go:117] "RemoveContainer" containerID="a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.634735 4846 scope.go:117] "RemoveContainer" containerID="52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf" Feb 02 14:42:13 crc kubenswrapper[4846]: E0202 14:42:13.635256 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf\": container with ID starting with 52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf not found: ID does not exist" containerID="52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.635292 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf"} err="failed to get container status \"52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf\": rpc error: code = NotFound desc = could not find container \"52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf\": container with ID starting with 52fa092a83cbc7d2577459d434d50d2b7f3edc686e110733994464daecbfddcf not found: ID does not exist" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.635315 4846 scope.go:117] "RemoveContainer" containerID="950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9" Feb 02 14:42:13 crc kubenswrapper[4846]: E0202 14:42:13.635616 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9\": container with ID starting with 950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9 not found: ID does not exist" containerID="950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.635646 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9"} err="failed to get container status \"950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9\": rpc error: code = NotFound desc = could not find container \"950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9\": container with ID starting with 950e3ed438841e79666890c7d88cc6d822af4d7b51db9fc7bf0e3f164bec3dd9 not found: ID does not exist" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.635666 4846 scope.go:117] "RemoveContainer" containerID="a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c" Feb 02 14:42:13 crc kubenswrapper[4846]: E0202 14:42:13.636024 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c\": container with ID starting with a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c not found: ID does not exist" containerID="a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c" Feb 02 14:42:13 crc kubenswrapper[4846]: I0202 14:42:13.636055 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c"} err="failed to get container status \"a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c\": rpc error: code = NotFound desc = could not find container \"a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c\": container with ID starting with a85fef3e71f04bbc489f3f05e11532e9a949b2cfc8bc3da70b4081a81451202c not found: ID does not exist" Feb 02 14:42:15 crc kubenswrapper[4846]: I0202 14:42:15.423715 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:42:15 crc kubenswrapper[4846]: E0202 14:42:15.424220 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:42:15 crc kubenswrapper[4846]: I0202 14:42:15.444961 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" path="/var/lib/kubelet/pods/8bf34605-f5f8-429c-8cfb-cac96fed3aa0/volumes" Feb 02 14:42:19 crc kubenswrapper[4846]: I0202 14:42:19.045488 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:42:19 crc kubenswrapper[4846]: I0202 14:42:19.098731 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:42:19 crc kubenswrapper[4846]: I0202 14:42:19.290244 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xjcgs"] Feb 02 14:42:20 crc kubenswrapper[4846]: I0202 14:42:20.578262 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xjcgs" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerName="registry-server" containerID="cri-o://971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e" gracePeriod=2 Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.174401 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.295272 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg9dx\" (UniqueName: \"kubernetes.io/projected/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-kube-api-access-tg9dx\") pod \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.295367 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-utilities\") pod \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.295398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-catalog-content\") pod \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\" (UID: \"97d78223-9f5c-4fb0-9d09-cae9a688a8c9\") " Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.297055 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-utilities" (OuterVolumeSpecName: "utilities") pod "97d78223-9f5c-4fb0-9d09-cae9a688a8c9" (UID: "97d78223-9f5c-4fb0-9d09-cae9a688a8c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.307051 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-kube-api-access-tg9dx" (OuterVolumeSpecName: "kube-api-access-tg9dx") pod "97d78223-9f5c-4fb0-9d09-cae9a688a8c9" (UID: "97d78223-9f5c-4fb0-9d09-cae9a688a8c9"). InnerVolumeSpecName "kube-api-access-tg9dx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.347419 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97d78223-9f5c-4fb0-9d09-cae9a688a8c9" (UID: "97d78223-9f5c-4fb0-9d09-cae9a688a8c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.398270 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg9dx\" (UniqueName: \"kubernetes.io/projected/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-kube-api-access-tg9dx\") on node \"crc\" DevicePath \"\"" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.398310 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.398320 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97d78223-9f5c-4fb0-9d09-cae9a688a8c9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.590490 4846 generic.go:334] "Generic (PLEG): container finished" podID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerID="971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e" exitCode=0 Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.590832 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjcgs" event={"ID":"97d78223-9f5c-4fb0-9d09-cae9a688a8c9","Type":"ContainerDied","Data":"971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e"} Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.590858 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xjcgs" event={"ID":"97d78223-9f5c-4fb0-9d09-cae9a688a8c9","Type":"ContainerDied","Data":"3c7c41c334af2ad1e13e95b3364fcdd7730676c7ca72146558f629091e5a4f5d"} Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.590874 4846 scope.go:117] "RemoveContainer" containerID="971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.590972 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xjcgs" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.615219 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xjcgs"] Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.616414 4846 scope.go:117] "RemoveContainer" containerID="3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.636265 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xjcgs"] Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.638503 4846 scope.go:117] "RemoveContainer" containerID="f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.684765 4846 scope.go:117] "RemoveContainer" containerID="971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e" Feb 02 14:42:21 crc kubenswrapper[4846]: E0202 14:42:21.685614 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e\": container with ID starting with 971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e not found: ID does not exist" containerID="971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.685671 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e"} err="failed to get container status \"971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e\": rpc error: code = NotFound desc = could not find container \"971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e\": container with ID starting with 971456b5029fc64019416f92595fa2682a47f7eb9d5baaeb53da48c39d1d7a2e not found: ID does not exist" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.685699 4846 scope.go:117] "RemoveContainer" containerID="3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e" Feb 02 14:42:21 crc kubenswrapper[4846]: E0202 14:42:21.686085 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e\": container with ID starting with 3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e not found: ID does not exist" containerID="3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.686114 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e"} err="failed to get container status \"3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e\": rpc error: code = NotFound desc = could not find container \"3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e\": container with ID starting with 3afecead75ffac4792b1db37f80171a544e3796312e28176bbcc1355d275d32e not found: ID does not exist" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.686129 4846 scope.go:117] "RemoveContainer" containerID="f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339" Feb 02 14:42:21 crc kubenswrapper[4846]: E0202 14:42:21.686452 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339\": container with ID starting with f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339 not found: ID does not exist" containerID="f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339" Feb 02 14:42:21 crc kubenswrapper[4846]: I0202 14:42:21.686475 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339"} err="failed to get container status \"f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339\": rpc error: code = NotFound desc = could not find container \"f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339\": container with ID starting with f431578fb80deec170467ab12cd9b3710d315153c4afc4104ae6419bd770a339 not found: ID does not exist" Feb 02 14:42:23 crc kubenswrapper[4846]: I0202 14:42:23.454091 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" path="/var/lib/kubelet/pods/97d78223-9f5c-4fb0-9d09-cae9a688a8c9/volumes" Feb 02 14:42:30 crc kubenswrapper[4846]: I0202 14:42:30.424025 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:42:30 crc kubenswrapper[4846]: E0202 14:42:30.424868 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:42:43 crc kubenswrapper[4846]: I0202 14:42:43.636734 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:42:43 crc kubenswrapper[4846]: E0202 14:42:43.637894 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:42:55 crc kubenswrapper[4846]: I0202 14:42:55.424031 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:42:55 crc kubenswrapper[4846]: E0202 14:42:55.424801 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:43:07 crc kubenswrapper[4846]: I0202 14:43:07.424459 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:43:07 crc kubenswrapper[4846]: E0202 14:43:07.425450 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:43:22 crc kubenswrapper[4846]: I0202 14:43:22.425769 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:43:22 crc kubenswrapper[4846]: E0202 14:43:22.427074 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:43:37 crc kubenswrapper[4846]: I0202 14:43:37.423864 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:43:37 crc kubenswrapper[4846]: E0202 14:43:37.424849 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:43:49 crc kubenswrapper[4846]: I0202 14:43:49.440555 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:43:49 crc kubenswrapper[4846]: E0202 14:43:49.441481 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:44:02 crc kubenswrapper[4846]: I0202 14:44:02.423902 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:44:02 crc kubenswrapper[4846]: E0202 14:44:02.424663 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:44:14 crc kubenswrapper[4846]: I0202 14:44:14.423702 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:44:14 crc kubenswrapper[4846]: E0202 14:44:14.424569 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:44:26 crc kubenswrapper[4846]: I0202 14:44:26.423794 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:44:26 crc kubenswrapper[4846]: E0202 14:44:26.424470 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:44:39 crc kubenswrapper[4846]: I0202 14:44:39.432456 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:44:39 crc kubenswrapper[4846]: E0202 14:44:39.433338 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:44:48 crc kubenswrapper[4846]: I0202 14:44:48.899905 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-64b5b76f97-xk9n4" podUID="cb8f2427-7089-4ecf-8ae9-4ece48067de8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 02 14:44:54 crc kubenswrapper[4846]: I0202 14:44:54.423481 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:44:54 crc kubenswrapper[4846]: E0202 14:44:54.424332 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.299256 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z5cdw"] Feb 02 14:44:56 crc kubenswrapper[4846]: E0202 14:44:56.300082 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerName="extract-content" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.300097 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerName="extract-content" Feb 02 14:44:56 crc kubenswrapper[4846]: E0202 14:44:56.300130 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerName="extract-utilities" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.300140 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerName="extract-utilities" Feb 02 14:44:56 crc kubenswrapper[4846]: E0202 14:44:56.300149 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerName="registry-server" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.300156 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerName="registry-server" Feb 02 14:44:56 crc kubenswrapper[4846]: E0202 14:44:56.301100 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerName="extract-content" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.301114 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerName="extract-content" Feb 02 14:44:56 crc kubenswrapper[4846]: E0202 14:44:56.301168 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerName="registry-server" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.301182 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerName="registry-server" Feb 02 14:44:56 crc kubenswrapper[4846]: E0202 14:44:56.301205 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerName="extract-utilities" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.301211 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerName="extract-utilities" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.301454 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="97d78223-9f5c-4fb0-9d09-cae9a688a8c9" containerName="registry-server" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.301486 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf34605-f5f8-429c-8cfb-cac96fed3aa0" containerName="registry-server" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.305474 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.314225 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5cdw"] Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.360196 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-utilities\") pod \"redhat-marketplace-z5cdw\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.360491 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-catalog-content\") pod \"redhat-marketplace-z5cdw\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.360812 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klznh\" (UniqueName: \"kubernetes.io/projected/72962c33-2fa8-4d67-a3ad-81748efcb2e7-kube-api-access-klznh\") pod \"redhat-marketplace-z5cdw\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.463038 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klznh\" (UniqueName: \"kubernetes.io/projected/72962c33-2fa8-4d67-a3ad-81748efcb2e7-kube-api-access-klznh\") pod \"redhat-marketplace-z5cdw\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.463451 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-utilities\") pod \"redhat-marketplace-z5cdw\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.463835 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-catalog-content\") pod \"redhat-marketplace-z5cdw\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.463929 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-utilities\") pod \"redhat-marketplace-z5cdw\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.464450 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-catalog-content\") pod \"redhat-marketplace-z5cdw\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.484531 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klznh\" (UniqueName: \"kubernetes.io/projected/72962c33-2fa8-4d67-a3ad-81748efcb2e7-kube-api-access-klznh\") pod \"redhat-marketplace-z5cdw\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:56 crc kubenswrapper[4846]: I0202 14:44:56.632805 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:44:57 crc kubenswrapper[4846]: I0202 14:44:57.132637 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5cdw"] Feb 02 14:44:57 crc kubenswrapper[4846]: W0202 14:44:57.137123 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72962c33_2fa8_4d67_a3ad_81748efcb2e7.slice/crio-4fa7c49d4f4461686c6b68bc10cefdb74dccf02074c014c7ec585c779d73beb2 WatchSource:0}: Error finding container 4fa7c49d4f4461686c6b68bc10cefdb74dccf02074c014c7ec585c779d73beb2: Status 404 returned error can't find the container with id 4fa7c49d4f4461686c6b68bc10cefdb74dccf02074c014c7ec585c779d73beb2 Feb 02 14:44:58 crc kubenswrapper[4846]: I0202 14:44:58.141610 4846 generic.go:334] "Generic (PLEG): container finished" podID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerID="15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7" exitCode=0 Feb 02 14:44:58 crc kubenswrapper[4846]: I0202 14:44:58.141674 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5cdw" event={"ID":"72962c33-2fa8-4d67-a3ad-81748efcb2e7","Type":"ContainerDied","Data":"15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7"} Feb 02 14:44:58 crc kubenswrapper[4846]: I0202 14:44:58.141701 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5cdw" event={"ID":"72962c33-2fa8-4d67-a3ad-81748efcb2e7","Type":"ContainerStarted","Data":"4fa7c49d4f4461686c6b68bc10cefdb74dccf02074c014c7ec585c779d73beb2"} Feb 02 14:44:58 crc kubenswrapper[4846]: I0202 14:44:58.145590 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 14:44:59 crc kubenswrapper[4846]: I0202 14:44:59.150882 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5cdw" event={"ID":"72962c33-2fa8-4d67-a3ad-81748efcb2e7","Type":"ContainerStarted","Data":"e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0"} Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.151847 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f"] Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.153584 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.156487 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.156544 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.165249 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f"] Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.171870 4846 generic.go:334] "Generic (PLEG): container finished" podID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerID="e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0" exitCode=0 Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.172028 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5cdw" event={"ID":"72962c33-2fa8-4d67-a3ad-81748efcb2e7","Type":"ContainerDied","Data":"e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0"} Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.259391 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbs2s\" (UniqueName: \"kubernetes.io/projected/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-kube-api-access-hbs2s\") pod \"collect-profiles-29500725-fhn6f\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.259467 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-config-volume\") pod \"collect-profiles-29500725-fhn6f\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.259610 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-secret-volume\") pod \"collect-profiles-29500725-fhn6f\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.361326 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbs2s\" (UniqueName: \"kubernetes.io/projected/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-kube-api-access-hbs2s\") pod \"collect-profiles-29500725-fhn6f\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.361931 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-config-volume\") pod \"collect-profiles-29500725-fhn6f\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.362069 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-secret-volume\") pod \"collect-profiles-29500725-fhn6f\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.363342 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-config-volume\") pod \"collect-profiles-29500725-fhn6f\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.368135 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-secret-volume\") pod \"collect-profiles-29500725-fhn6f\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.378403 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbs2s\" (UniqueName: \"kubernetes.io/projected/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-kube-api-access-hbs2s\") pod \"collect-profiles-29500725-fhn6f\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.486809 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:00 crc kubenswrapper[4846]: I0202 14:45:00.964898 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f"] Feb 02 14:45:01 crc kubenswrapper[4846]: W0202 14:45:01.206718 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b0837e7_a782_4c89_9f6b_3fe6df3aecb9.slice/crio-975aaccb33ebf56fc409771c77028e4f35d41e8c086697bada6e6d0a576f87b2 WatchSource:0}: Error finding container 975aaccb33ebf56fc409771c77028e4f35d41e8c086697bada6e6d0a576f87b2: Status 404 returned error can't find the container with id 975aaccb33ebf56fc409771c77028e4f35d41e8c086697bada6e6d0a576f87b2 Feb 02 14:45:02 crc kubenswrapper[4846]: I0202 14:45:02.191584 4846 generic.go:334] "Generic (PLEG): container finished" podID="6b0837e7-a782-4c89-9f6b-3fe6df3aecb9" containerID="e9ba7f2141b796f3bdfc142ee0d26092afb1b33d5da259cf50ab9fc1c48ff8ac" exitCode=0 Feb 02 14:45:02 crc kubenswrapper[4846]: I0202 14:45:02.191713 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" event={"ID":"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9","Type":"ContainerDied","Data":"e9ba7f2141b796f3bdfc142ee0d26092afb1b33d5da259cf50ab9fc1c48ff8ac"} Feb 02 14:45:02 crc kubenswrapper[4846]: I0202 14:45:02.192219 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" event={"ID":"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9","Type":"ContainerStarted","Data":"975aaccb33ebf56fc409771c77028e4f35d41e8c086697bada6e6d0a576f87b2"} Feb 02 14:45:02 crc kubenswrapper[4846]: I0202 14:45:02.194860 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5cdw" event={"ID":"72962c33-2fa8-4d67-a3ad-81748efcb2e7","Type":"ContainerStarted","Data":"86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a"} Feb 02 14:45:02 crc kubenswrapper[4846]: I0202 14:45:02.238670 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z5cdw" podStartSLOduration=3.7753633239999997 podStartE2EDuration="6.238641623s" podCreationTimestamp="2026-02-02 14:44:56 +0000 UTC" firstStartedPulling="2026-02-02 14:44:58.145309284 +0000 UTC m=+9329.373896147" lastFinishedPulling="2026-02-02 14:45:00.608587583 +0000 UTC m=+9331.837174446" observedRunningTime="2026-02-02 14:45:02.228801043 +0000 UTC m=+9333.457387916" watchObservedRunningTime="2026-02-02 14:45:02.238641623 +0000 UTC m=+9333.467228506" Feb 02 14:45:03 crc kubenswrapper[4846]: I0202 14:45:03.618739 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:03 crc kubenswrapper[4846]: I0202 14:45:03.632205 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-secret-volume\") pod \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " Feb 02 14:45:03 crc kubenswrapper[4846]: I0202 14:45:03.632458 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbs2s\" (UniqueName: \"kubernetes.io/projected/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-kube-api-access-hbs2s\") pod \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " Feb 02 14:45:03 crc kubenswrapper[4846]: I0202 14:45:03.632494 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-config-volume\") pod \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\" (UID: \"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9\") " Feb 02 14:45:03 crc kubenswrapper[4846]: I0202 14:45:03.634982 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-config-volume" (OuterVolumeSpecName: "config-volume") pod "6b0837e7-a782-4c89-9f6b-3fe6df3aecb9" (UID: "6b0837e7-a782-4c89-9f6b-3fe6df3aecb9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 14:45:03 crc kubenswrapper[4846]: I0202 14:45:03.649145 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6b0837e7-a782-4c89-9f6b-3fe6df3aecb9" (UID: "6b0837e7-a782-4c89-9f6b-3fe6df3aecb9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 14:45:03 crc kubenswrapper[4846]: I0202 14:45:03.649240 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-kube-api-access-hbs2s" (OuterVolumeSpecName: "kube-api-access-hbs2s") pod "6b0837e7-a782-4c89-9f6b-3fe6df3aecb9" (UID: "6b0837e7-a782-4c89-9f6b-3fe6df3aecb9"). InnerVolumeSpecName "kube-api-access-hbs2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:45:03 crc kubenswrapper[4846]: I0202 14:45:03.735587 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbs2s\" (UniqueName: \"kubernetes.io/projected/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-kube-api-access-hbs2s\") on node \"crc\" DevicePath \"\"" Feb 02 14:45:03 crc kubenswrapper[4846]: I0202 14:45:03.735652 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 14:45:03 crc kubenswrapper[4846]: I0202 14:45:03.735662 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b0837e7-a782-4c89-9f6b-3fe6df3aecb9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 14:45:04 crc kubenswrapper[4846]: I0202 14:45:04.216051 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" event={"ID":"6b0837e7-a782-4c89-9f6b-3fe6df3aecb9","Type":"ContainerDied","Data":"975aaccb33ebf56fc409771c77028e4f35d41e8c086697bada6e6d0a576f87b2"} Feb 02 14:45:04 crc kubenswrapper[4846]: I0202 14:45:04.216097 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500725-fhn6f" Feb 02 14:45:04 crc kubenswrapper[4846]: I0202 14:45:04.216113 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="975aaccb33ebf56fc409771c77028e4f35d41e8c086697bada6e6d0a576f87b2" Feb 02 14:45:04 crc kubenswrapper[4846]: I0202 14:45:04.720668 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz"] Feb 02 14:45:04 crc kubenswrapper[4846]: I0202 14:45:04.734258 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500680-j9xzz"] Feb 02 14:45:05 crc kubenswrapper[4846]: I0202 14:45:05.440404 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82d62939-2d72-495d-a677-4dad13b4fab1" path="/var/lib/kubelet/pods/82d62939-2d72-495d-a677-4dad13b4fab1/volumes" Feb 02 14:45:06 crc kubenswrapper[4846]: I0202 14:45:06.633708 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:45:06 crc kubenswrapper[4846]: I0202 14:45:06.635202 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:45:06 crc kubenswrapper[4846]: I0202 14:45:06.704954 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:45:07 crc kubenswrapper[4846]: I0202 14:45:07.291230 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:45:07 crc kubenswrapper[4846]: I0202 14:45:07.342816 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5cdw"] Feb 02 14:45:07 crc kubenswrapper[4846]: I0202 14:45:07.423252 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:45:07 crc kubenswrapper[4846]: E0202 14:45:07.423653 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:45:09 crc kubenswrapper[4846]: I0202 14:45:09.262828 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z5cdw" podUID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerName="registry-server" containerID="cri-o://86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a" gracePeriod=2 Feb 02 14:45:09 crc kubenswrapper[4846]: E0202 14:45:09.446052 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72962c33_2fa8_4d67_a3ad_81748efcb2e7.slice/crio-conmon-86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a.scope\": RecentStats: unable to find data in memory cache]" Feb 02 14:45:09 crc kubenswrapper[4846]: I0202 14:45:09.814609 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:45:09 crc kubenswrapper[4846]: I0202 14:45:09.968493 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-catalog-content\") pod \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " Feb 02 14:45:09 crc kubenswrapper[4846]: I0202 14:45:09.968547 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-utilities\") pod \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " Feb 02 14:45:09 crc kubenswrapper[4846]: I0202 14:45:09.968585 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klznh\" (UniqueName: \"kubernetes.io/projected/72962c33-2fa8-4d67-a3ad-81748efcb2e7-kube-api-access-klznh\") pod \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\" (UID: \"72962c33-2fa8-4d67-a3ad-81748efcb2e7\") " Feb 02 14:45:09 crc kubenswrapper[4846]: I0202 14:45:09.970503 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-utilities" (OuterVolumeSpecName: "utilities") pod "72962c33-2fa8-4d67-a3ad-81748efcb2e7" (UID: "72962c33-2fa8-4d67-a3ad-81748efcb2e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:45:09 crc kubenswrapper[4846]: I0202 14:45:09.985584 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72962c33-2fa8-4d67-a3ad-81748efcb2e7-kube-api-access-klznh" (OuterVolumeSpecName: "kube-api-access-klznh") pod "72962c33-2fa8-4d67-a3ad-81748efcb2e7" (UID: "72962c33-2fa8-4d67-a3ad-81748efcb2e7"). InnerVolumeSpecName "kube-api-access-klznh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.014224 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72962c33-2fa8-4d67-a3ad-81748efcb2e7" (UID: "72962c33-2fa8-4d67-a3ad-81748efcb2e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.071238 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.071296 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72962c33-2fa8-4d67-a3ad-81748efcb2e7-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.071306 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klznh\" (UniqueName: \"kubernetes.io/projected/72962c33-2fa8-4d67-a3ad-81748efcb2e7-kube-api-access-klznh\") on node \"crc\" DevicePath \"\"" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.275577 4846 generic.go:334] "Generic (PLEG): container finished" podID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerID="86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a" exitCode=0 Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.275616 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5cdw" event={"ID":"72962c33-2fa8-4d67-a3ad-81748efcb2e7","Type":"ContainerDied","Data":"86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a"} Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.275657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z5cdw" event={"ID":"72962c33-2fa8-4d67-a3ad-81748efcb2e7","Type":"ContainerDied","Data":"4fa7c49d4f4461686c6b68bc10cefdb74dccf02074c014c7ec585c779d73beb2"} Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.275665 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z5cdw" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.275678 4846 scope.go:117] "RemoveContainer" containerID="86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.303262 4846 scope.go:117] "RemoveContainer" containerID="e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.310397 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5cdw"] Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.319588 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z5cdw"] Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.334180 4846 scope.go:117] "RemoveContainer" containerID="15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.373845 4846 scope.go:117] "RemoveContainer" containerID="86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a" Feb 02 14:45:10 crc kubenswrapper[4846]: E0202 14:45:10.374166 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a\": container with ID starting with 86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a not found: ID does not exist" containerID="86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.374197 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a"} err="failed to get container status \"86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a\": rpc error: code = NotFound desc = could not find container \"86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a\": container with ID starting with 86216754c65a6ff57503ea41e3b9f27837d4b31373e88f03bd582a6f37d3d85a not found: ID does not exist" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.374216 4846 scope.go:117] "RemoveContainer" containerID="e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0" Feb 02 14:45:10 crc kubenswrapper[4846]: E0202 14:45:10.374626 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0\": container with ID starting with e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0 not found: ID does not exist" containerID="e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.374665 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0"} err="failed to get container status \"e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0\": rpc error: code = NotFound desc = could not find container \"e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0\": container with ID starting with e3035dd89975d54ec7dae5037c0ec2323f6c5be0038204a5f727a3e3910183a0 not found: ID does not exist" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.374681 4846 scope.go:117] "RemoveContainer" containerID="15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7" Feb 02 14:45:10 crc kubenswrapper[4846]: E0202 14:45:10.375894 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7\": container with ID starting with 15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7 not found: ID does not exist" containerID="15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7" Feb 02 14:45:10 crc kubenswrapper[4846]: I0202 14:45:10.375921 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7"} err="failed to get container status \"15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7\": rpc error: code = NotFound desc = could not find container \"15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7\": container with ID starting with 15a14b05cbd022bf85e21e5a0b797a92cea621ea7050a19a859f1c59832facb7 not found: ID does not exist" Feb 02 14:45:11 crc kubenswrapper[4846]: I0202 14:45:11.435710 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" path="/var/lib/kubelet/pods/72962c33-2fa8-4d67-a3ad-81748efcb2e7/volumes" Feb 02 14:45:22 crc kubenswrapper[4846]: I0202 14:45:22.423846 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:45:22 crc kubenswrapper[4846]: E0202 14:45:22.425108 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:45:35 crc kubenswrapper[4846]: I0202 14:45:35.423497 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:45:35 crc kubenswrapper[4846]: E0202 14:45:35.424224 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:45:46 crc kubenswrapper[4846]: I0202 14:45:46.350426 4846 scope.go:117] "RemoveContainer" containerID="4bd03e5ad50707aa0e71408b7e83edd214369278933e9d4dede1351c4ea61f87" Feb 02 14:45:46 crc kubenswrapper[4846]: I0202 14:45:46.423914 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:45:46 crc kubenswrapper[4846]: E0202 14:45:46.424259 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:46:01 crc kubenswrapper[4846]: I0202 14:46:01.428537 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:46:01 crc kubenswrapper[4846]: E0202 14:46:01.429960 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:46:16 crc kubenswrapper[4846]: I0202 14:46:16.424612 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:46:16 crc kubenswrapper[4846]: E0202 14:46:16.425817 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:46:28 crc kubenswrapper[4846]: I0202 14:46:28.423980 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:46:28 crc kubenswrapper[4846]: E0202 14:46:28.424725 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:46:41 crc kubenswrapper[4846]: I0202 14:46:41.424239 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:46:41 crc kubenswrapper[4846]: E0202 14:46:41.425122 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:46:46 crc kubenswrapper[4846]: I0202 14:46:46.418761 4846 scope.go:117] "RemoveContainer" containerID="680c8d7e76ec2bc066e8b78c1d2768031f249da9bddf734871cdef15fa790971" Feb 02 14:46:53 crc kubenswrapper[4846]: I0202 14:46:53.424146 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:46:53 crc kubenswrapper[4846]: E0202 14:46:53.425060 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:47:08 crc kubenswrapper[4846]: I0202 14:47:08.424467 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:47:09 crc kubenswrapper[4846]: I0202 14:47:09.567398 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"229e52826f90f0a1025ad49518481c25a85ee892227df94a2914ba07fbeeb68e"} Feb 02 14:49:20 crc kubenswrapper[4846]: I0202 14:49:20.554046 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ec35752b-6884-4a8e-b077-43cf63dd98c3/init-config-reloader/0.log" Feb 02 14:49:21 crc kubenswrapper[4846]: I0202 14:49:21.214432 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ec35752b-6884-4a8e-b077-43cf63dd98c3/config-reloader/0.log" Feb 02 14:49:21 crc kubenswrapper[4846]: I0202 14:49:21.224921 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ec35752b-6884-4a8e-b077-43cf63dd98c3/init-config-reloader/0.log" Feb 02 14:49:21 crc kubenswrapper[4846]: I0202 14:49:21.229789 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ec35752b-6884-4a8e-b077-43cf63dd98c3/alertmanager/0.log" Feb 02 14:49:21 crc kubenswrapper[4846]: I0202 14:49:21.442302 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_feca1cf6-5353-4280-9250-c233f7e2f888/aodh-api/0.log" Feb 02 14:49:21 crc kubenswrapper[4846]: I0202 14:49:21.479736 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_feca1cf6-5353-4280-9250-c233f7e2f888/aodh-listener/0.log" Feb 02 14:49:21 crc kubenswrapper[4846]: I0202 14:49:21.489001 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_feca1cf6-5353-4280-9250-c233f7e2f888/aodh-evaluator/0.log" Feb 02 14:49:21 crc kubenswrapper[4846]: I0202 14:49:21.641841 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_feca1cf6-5353-4280-9250-c233f7e2f888/aodh-notifier/0.log" Feb 02 14:49:21 crc kubenswrapper[4846]: I0202 14:49:21.698098 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-857b8ccfcd-55wgg_a33b00bf-aa6a-4eb8-be29-be3e943e276f/barbican-api/0.log" Feb 02 14:49:21 crc kubenswrapper[4846]: I0202 14:49:21.782652 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-857b8ccfcd-55wgg_a33b00bf-aa6a-4eb8-be29-be3e943e276f/barbican-api-log/0.log" Feb 02 14:49:21 crc kubenswrapper[4846]: I0202 14:49:21.901382 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7687b59478-hmtmb_3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb/barbican-keystone-listener/0.log" Feb 02 14:49:22 crc kubenswrapper[4846]: I0202 14:49:22.089715 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7687b59478-hmtmb_3ee8ddd0-79b7-4444-9f6b-ac2b2e0385eb/barbican-keystone-listener-log/0.log" Feb 02 14:49:22 crc kubenswrapper[4846]: I0202 14:49:22.153892 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7874cbf7b5-2kgbm_42acd0dc-ac0d-4bc9-a814-12ced76e190a/barbican-worker/0.log" Feb 02 14:49:22 crc kubenswrapper[4846]: I0202 14:49:22.253277 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7874cbf7b5-2kgbm_42acd0dc-ac0d-4bc9-a814-12ced76e190a/barbican-worker-log/0.log" Feb 02 14:49:22 crc kubenswrapper[4846]: I0202 14:49:22.534058 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f6b5c3d0-71ce-441c-8423-bc2f285a782f/ceilometer-central-agent/0.log" Feb 02 14:49:22 crc kubenswrapper[4846]: I0202 14:49:22.537967 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-92q2f_b0606b34-07f7-4eab-96e1-4601339b9bfa/bootstrap-openstack-openstack-cell1/0.log" Feb 02 14:49:22 crc kubenswrapper[4846]: I0202 14:49:22.698378 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f6b5c3d0-71ce-441c-8423-bc2f285a782f/ceilometer-notification-agent/0.log" Feb 02 14:49:22 crc kubenswrapper[4846]: I0202 14:49:22.737740 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f6b5c3d0-71ce-441c-8423-bc2f285a782f/proxy-httpd/0.log" Feb 02 14:49:22 crc kubenswrapper[4846]: I0202 14:49:22.770394 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f6b5c3d0-71ce-441c-8423-bc2f285a782f/sg-core/0.log" Feb 02 14:49:22 crc kubenswrapper[4846]: I0202 14:49:22.968205 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f4d63fe7-c925-40de-b144-8ec6f667f22f/cinder-api/0.log" Feb 02 14:49:23 crc kubenswrapper[4846]: I0202 14:49:23.203863 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f4d63fe7-c925-40de-b144-8ec6f667f22f/cinder-api-log/0.log" Feb 02 14:49:23 crc kubenswrapper[4846]: I0202 14:49:23.700978 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f67f8ace-1e94-4f90-b8cf-d5a732eed4dd/cinder-scheduler/0.log" Feb 02 14:49:23 crc kubenswrapper[4846]: I0202 14:49:23.722267 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-ktqhx_8284a56e-0573-4a1e-85a0-58e62178133d/configure-network-openstack-openstack-cell1/0.log" Feb 02 14:49:23 crc kubenswrapper[4846]: I0202 14:49:23.836450 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f67f8ace-1e94-4f90-b8cf-d5a732eed4dd/probe/0.log" Feb 02 14:49:24 crc kubenswrapper[4846]: I0202 14:49:24.016307 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-txvs7_e2556352-fb48-4800-a719-ece237df78f5/configure-os-openstack-openstack-cell1/0.log" Feb 02 14:49:24 crc kubenswrapper[4846]: I0202 14:49:24.148207 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7d4cd97679-9wtr9_eea30013-7cac-4157-b39e-da5f71999fba/init/0.log" Feb 02 14:49:24 crc kubenswrapper[4846]: I0202 14:49:24.447817 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7d4cd97679-9wtr9_eea30013-7cac-4157-b39e-da5f71999fba/init/0.log" Feb 02 14:49:24 crc kubenswrapper[4846]: I0202 14:49:24.456739 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7d4cd97679-9wtr9_eea30013-7cac-4157-b39e-da5f71999fba/dnsmasq-dns/0.log" Feb 02 14:49:24 crc kubenswrapper[4846]: I0202 14:49:24.539136 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-hlgvr_dac40561-426a-470e-8af9-05e0aea97e41/download-cache-openstack-openstack-cell1/0.log" Feb 02 14:49:24 crc kubenswrapper[4846]: I0202 14:49:24.755195 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_98a97c37-c689-484f-9f21-c3bd2fc20d32/glance-log/0.log" Feb 02 14:49:24 crc kubenswrapper[4846]: I0202 14:49:24.831245 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_98a97c37-c689-484f-9f21-c3bd2fc20d32/glance-httpd/0.log" Feb 02 14:49:25 crc kubenswrapper[4846]: I0202 14:49:25.112356 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_07db105d-693c-46f6-b19c-e2f470acb949/glance-log/0.log" Feb 02 14:49:25 crc kubenswrapper[4846]: I0202 14:49:25.122319 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_07db105d-693c-46f6-b19c-e2f470acb949/glance-httpd/0.log" Feb 02 14:49:25 crc kubenswrapper[4846]: I0202 14:49:25.575357 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-555b878bf4-mzv78_4b20035b-bca2-4bfd-86c8-e53deda9da15/heat-engine/0.log" Feb 02 14:49:25 crc kubenswrapper[4846]: I0202 14:49:25.776140 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-684d5dd44f-9r7lc_fddebaab-459c-4aba-b375-859a15797a03/heat-api/0.log" Feb 02 14:49:25 crc kubenswrapper[4846]: I0202 14:49:25.879353 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c49958db-lgw6w_46333d82-5d6b-4bae-b1ad-7a30e6a88490/horizon/0.log" Feb 02 14:49:25 crc kubenswrapper[4846]: I0202 14:49:25.949285 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-54dbdfd74b-wqbr7_bfc91827-42a5-4fc6-9720-b4dfd5e246d4/heat-cfnapi/0.log" Feb 02 14:49:26 crc kubenswrapper[4846]: I0202 14:49:26.162300 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-swnxb_392df862-6b24-4249-9a20-52d721c57a73/install-certs-openstack-openstack-cell1/0.log" Feb 02 14:49:26 crc kubenswrapper[4846]: I0202 14:49:26.311480 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-ds2qr_1dc40137-f83c-4dd8-9a91-cda128768635/install-os-openstack-openstack-cell1/0.log" Feb 02 14:49:26 crc kubenswrapper[4846]: I0202 14:49:26.365152 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c49958db-lgw6w_46333d82-5d6b-4bae-b1ad-7a30e6a88490/horizon-log/0.log" Feb 02 14:49:26 crc kubenswrapper[4846]: I0202 14:49:26.557742 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29500681-ln5z2_32950bad-c06b-4cd1-bae5-1fd20646ce59/keystone-cron/0.log" Feb 02 14:49:26 crc kubenswrapper[4846]: I0202 14:49:26.581498 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-964849ccd-5gkkw_900a578d-a369-4470-a522-ee6b4cf5019d/keystone-api/0.log" Feb 02 14:49:26 crc kubenswrapper[4846]: I0202 14:49:26.728809 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_f3d505b7-7c82-49d0-930d-8ba793ac0d4a/kube-state-metrics/0.log" Feb 02 14:49:26 crc kubenswrapper[4846]: I0202 14:49:26.807892 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-qqkcn_8042645e-7a96-4a10-ab69-ba61f5280a61/libvirt-openstack-openstack-cell1/0.log" Feb 02 14:49:27 crc kubenswrapper[4846]: I0202 14:49:27.121551 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d7f794577-lckxg_04c833b7-c2fd-4697-b77b-5a777c7678e8/neutron-httpd/0.log" Feb 02 14:49:27 crc kubenswrapper[4846]: I0202 14:49:27.279896 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d7f794577-lckxg_04c833b7-c2fd-4697-b77b-5a777c7678e8/neutron-api/0.log" Feb 02 14:49:27 crc kubenswrapper[4846]: I0202 14:49:27.356604 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-zpkt6_7bea9696-bedb-46b6-a458-51e7776a7922/neutron-dhcp-openstack-openstack-cell1/0.log" Feb 02 14:49:27 crc kubenswrapper[4846]: I0202 14:49:27.586013 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-t7g6s_fac8ff80-70d1-4d0c-974f-011c0ccb6bf9/neutron-metadata-openstack-openstack-cell1/0.log" Feb 02 14:49:27 crc kubenswrapper[4846]: I0202 14:49:27.610399 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-7n5sr_23bb6579-b1e5-4e62-a6b1-6a4ee25b8ebc/neutron-sriov-openstack-openstack-cell1/0.log" Feb 02 14:49:28 crc kubenswrapper[4846]: I0202 14:49:28.084456 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d7dd5959-c237-42de-9578-e5450cfe6092/nova-api-log/0.log" Feb 02 14:49:28 crc kubenswrapper[4846]: I0202 14:49:28.133173 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d7dd5959-c237-42de-9578-e5450cfe6092/nova-api-api/0.log" Feb 02 14:49:28 crc kubenswrapper[4846]: I0202 14:49:28.148131 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_7ab6bdf6-72fd-46d7-9c26-bb7a11063b5e/nova-cell0-conductor-conductor/0.log" Feb 02 14:49:28 crc kubenswrapper[4846]: I0202 14:49:28.503130 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c241eff7-d5c6-4a5f-ae44-a28e16a4e609/nova-cell1-conductor-conductor/0.log" Feb 02 14:49:28 crc kubenswrapper[4846]: I0202 14:49:28.577667 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e611f1f1-65e6-4ce4-898b-a16e4bf0ee61/nova-cell1-novncproxy-novncproxy/0.log" Feb 02 14:49:28 crc kubenswrapper[4846]: I0202 14:49:28.770426 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell8j46x_96fbf45d-7c8a-48e4-9d49-c6498105c3a5/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Feb 02 14:49:28 crc kubenswrapper[4846]: I0202 14:49:28.918437 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-vzlgm_a1cd8b9b-045d-422d-ad00-60bccdec0ca7/nova-cell1-openstack-openstack-cell1/0.log" Feb 02 14:49:29 crc kubenswrapper[4846]: I0202 14:49:29.132825 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f1b2b0df-c8bd-4dad-a9bc-a91edc63d246/nova-metadata-log/0.log" Feb 02 14:49:29 crc kubenswrapper[4846]: I0202 14:49:29.430121 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_fe78d901-4a4a-4db5-9ef3-c15b7501cc11/nova-scheduler-scheduler/0.log" Feb 02 14:49:29 crc kubenswrapper[4846]: I0202 14:49:29.492245 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-db4457bbc-xgwkd_543b5e7c-8a34-45bc-a19c-75f27a036804/init/0.log" Feb 02 14:49:29 crc kubenswrapper[4846]: I0202 14:49:29.713724 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-db4457bbc-xgwkd_543b5e7c-8a34-45bc-a19c-75f27a036804/init/0.log" Feb 02 14:49:29 crc kubenswrapper[4846]: I0202 14:49:29.728759 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f1b2b0df-c8bd-4dad-a9bc-a91edc63d246/nova-metadata-metadata/0.log" Feb 02 14:49:29 crc kubenswrapper[4846]: I0202 14:49:29.751923 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-db4457bbc-xgwkd_543b5e7c-8a34-45bc-a19c-75f27a036804/octavia-api-provider-agent/0.log" Feb 02 14:49:29 crc kubenswrapper[4846]: I0202 14:49:29.964991 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-db4457bbc-xgwkd_543b5e7c-8a34-45bc-a19c-75f27a036804/octavia-api/0.log" Feb 02 14:49:29 crc kubenswrapper[4846]: I0202 14:49:29.994754 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-c782n_70a6f659-55db-455b-9264-a9b4cefb6334/init/0.log" Feb 02 14:49:30 crc kubenswrapper[4846]: I0202 14:49:30.169718 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-c782n_70a6f659-55db-455b-9264-a9b4cefb6334/init/0.log" Feb 02 14:49:30 crc kubenswrapper[4846]: I0202 14:49:30.277056 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-cdp5l_70f3f665-75af-4ef6-b89f-4479385845af/init/0.log" Feb 02 14:49:30 crc kubenswrapper[4846]: I0202 14:49:30.336340 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-c782n_70a6f659-55db-455b-9264-a9b4cefb6334/octavia-healthmanager/0.log" Feb 02 14:49:30 crc kubenswrapper[4846]: I0202 14:49:30.480732 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:49:30 crc kubenswrapper[4846]: I0202 14:49:30.480795 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:49:30 crc kubenswrapper[4846]: I0202 14:49:30.505891 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-cdp5l_70f3f665-75af-4ef6-b89f-4479385845af/init/0.log" Feb 02 14:49:30 crc kubenswrapper[4846]: I0202 14:49:30.537234 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-cdp5l_70f3f665-75af-4ef6-b89f-4479385845af/octavia-housekeeping/0.log" Feb 02 14:49:30 crc kubenswrapper[4846]: I0202 14:49:30.707855 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-65dd99cb46-gjk5s_7bae16d8-7405-45de-8a7f-e3714d1fa4eb/init/0.log" Feb 02 14:49:30 crc kubenswrapper[4846]: I0202 14:49:30.855113 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-65dd99cb46-gjk5s_7bae16d8-7405-45de-8a7f-e3714d1fa4eb/init/0.log" Feb 02 14:49:30 crc kubenswrapper[4846]: I0202 14:49:30.859825 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-65dd99cb46-gjk5s_7bae16d8-7405-45de-8a7f-e3714d1fa4eb/octavia-amphora-httpd/0.log" Feb 02 14:49:31 crc kubenswrapper[4846]: I0202 14:49:31.027483 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-rddhp_88517515-ad14-4ea6-93f9-f1bde00e726a/init/0.log" Feb 02 14:49:31 crc kubenswrapper[4846]: I0202 14:49:31.256675 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-rddhp_88517515-ad14-4ea6-93f9-f1bde00e726a/octavia-rsyslog/0.log" Feb 02 14:49:31 crc kubenswrapper[4846]: I0202 14:49:31.287554 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-7jg2f_ea7dbcb7-36a7-4606-ba9f-a85ecb78640a/init/0.log" Feb 02 14:49:31 crc kubenswrapper[4846]: I0202 14:49:31.310316 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-rddhp_88517515-ad14-4ea6-93f9-f1bde00e726a/init/0.log" Feb 02 14:49:31 crc kubenswrapper[4846]: I0202 14:49:31.575540 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-7jg2f_ea7dbcb7-36a7-4606-ba9f-a85ecb78640a/init/0.log" Feb 02 14:49:31 crc kubenswrapper[4846]: I0202 14:49:31.637895 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3003f765-6322-4adf-a73d-1ded30560bf9/mysql-bootstrap/0.log" Feb 02 14:49:31 crc kubenswrapper[4846]: I0202 14:49:31.841828 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-7jg2f_ea7dbcb7-36a7-4606-ba9f-a85ecb78640a/octavia-worker/0.log" Feb 02 14:49:31 crc kubenswrapper[4846]: I0202 14:49:31.850360 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3003f765-6322-4adf-a73d-1ded30560bf9/mysql-bootstrap/0.log" Feb 02 14:49:31 crc kubenswrapper[4846]: I0202 14:49:31.896959 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3003f765-6322-4adf-a73d-1ded30560bf9/galera/0.log" Feb 02 14:49:32 crc kubenswrapper[4846]: I0202 14:49:32.105949 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bed62f97-3e65-40e9-bc60-1b0338fc4672/mysql-bootstrap/0.log" Feb 02 14:49:32 crc kubenswrapper[4846]: I0202 14:49:32.276707 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bed62f97-3e65-40e9-bc60-1b0338fc4672/mysql-bootstrap/0.log" Feb 02 14:49:32 crc kubenswrapper[4846]: I0202 14:49:32.328644 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_bed62f97-3e65-40e9-bc60-1b0338fc4672/galera/0.log" Feb 02 14:49:33 crc kubenswrapper[4846]: I0202 14:49:33.025575 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b1f840c7-4bd5-427a-9f4b-b4fec6280665/openstackclient/0.log" Feb 02 14:49:33 crc kubenswrapper[4846]: I0202 14:49:33.059918 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-hnvm7_7dd42298-2b9c-4382-b25e-9afe253d226f/openstack-network-exporter/0.log" Feb 02 14:49:33 crc kubenswrapper[4846]: I0202 14:49:33.260911 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mkdn9_f7c4abfa-9b37-4c15-accf-fcb0007b4d26/ovsdb-server-init/0.log" Feb 02 14:49:33 crc kubenswrapper[4846]: I0202 14:49:33.838480 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mkdn9_f7c4abfa-9b37-4c15-accf-fcb0007b4d26/ovs-vswitchd/0.log" Feb 02 14:49:33 crc kubenswrapper[4846]: I0202 14:49:33.859555 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mkdn9_f7c4abfa-9b37-4c15-accf-fcb0007b4d26/ovsdb-server-init/0.log" Feb 02 14:49:33 crc kubenswrapper[4846]: I0202 14:49:33.879501 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mkdn9_f7c4abfa-9b37-4c15-accf-fcb0007b4d26/ovsdb-server/0.log" Feb 02 14:49:34 crc kubenswrapper[4846]: I0202 14:49:34.042851 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-v8bjp_79522d21-4471-4cdb-8caa-41c17214b4f1/ovn-controller/0.log" Feb 02 14:49:34 crc kubenswrapper[4846]: I0202 14:49:34.066247 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e9c4be3d-913c-42f8-9847-db39fcc49afd/openstack-network-exporter/0.log" Feb 02 14:49:34 crc kubenswrapper[4846]: I0202 14:49:34.075523 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_e9c4be3d-913c-42f8-9847-db39fcc49afd/ovn-northd/0.log" Feb 02 14:49:34 crc kubenswrapper[4846]: I0202 14:49:34.363119 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-vqz97_650a9e6b-b0ff-4129-b83c-b547a072b2af/ovn-openstack-openstack-cell1/0.log" Feb 02 14:49:34 crc kubenswrapper[4846]: I0202 14:49:34.382069 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_692cf6ee-1eb8-45e5-be12-44bb709fd6ca/openstack-network-exporter/0.log" Feb 02 14:49:35 crc kubenswrapper[4846]: I0202 14:49:35.197149 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_692cf6ee-1eb8-45e5-be12-44bb709fd6ca/ovsdbserver-nb/0.log" Feb 02 14:49:35 crc kubenswrapper[4846]: I0202 14:49:35.297155 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_fdb3364f-23b6-4875-a37d-a16af9832978/ovsdbserver-nb/0.log" Feb 02 14:49:35 crc kubenswrapper[4846]: I0202 14:49:35.327391 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_fdb3364f-23b6-4875-a37d-a16af9832978/openstack-network-exporter/0.log" Feb 02 14:49:35 crc kubenswrapper[4846]: I0202 14:49:35.505035 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_396188ac-2f00-4ae5-8a3a-0273a704e60f/openstack-network-exporter/0.log" Feb 02 14:49:35 crc kubenswrapper[4846]: I0202 14:49:35.546097 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_396188ac-2f00-4ae5-8a3a-0273a704e60f/ovsdbserver-nb/0.log" Feb 02 14:49:35 crc kubenswrapper[4846]: I0202 14:49:35.699739 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_106a77c0-d654-4a19-a691-cf7bf2efd9c4/openstack-network-exporter/0.log" Feb 02 14:49:35 crc kubenswrapper[4846]: I0202 14:49:35.766485 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_106a77c0-d654-4a19-a691-cf7bf2efd9c4/ovsdbserver-sb/0.log" Feb 02 14:49:35 crc kubenswrapper[4846]: I0202 14:49:35.922317 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_230ca7a4-fc29-44a0-8986-e25db00d4ebc/openstack-network-exporter/0.log" Feb 02 14:49:35 crc kubenswrapper[4846]: I0202 14:49:35.936705 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_230ca7a4-fc29-44a0-8986-e25db00d4ebc/ovsdbserver-sb/0.log" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.091816 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_e326ede9-6300-47da-b9b8-48e8ade170a4/openstack-network-exporter/0.log" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.192787 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_e326ede9-6300-47da-b9b8-48e8ade170a4/ovsdbserver-sb/0.log" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.284832 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68f6f77966-bf8l5_f14d64ee-705c-4c20-ab89-12cbfd89f8b8/placement-api/0.log" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.417791 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68f6f77966-bf8l5_f14d64ee-705c-4c20-ab89-12cbfd89f8b8/placement-log/0.log" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.538108 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-c4scfn_1f33b8f6-a26a-4052-aa57-b0ce0d2924c6/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.647676 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fqskx"] Feb 02 14:49:36 crc kubenswrapper[4846]: E0202 14:49:36.648131 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerName="extract-content" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.648149 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerName="extract-content" Feb 02 14:49:36 crc kubenswrapper[4846]: E0202 14:49:36.648173 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerName="extract-utilities" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.648179 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerName="extract-utilities" Feb 02 14:49:36 crc kubenswrapper[4846]: E0202 14:49:36.648193 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b0837e7-a782-4c89-9f6b-3fe6df3aecb9" containerName="collect-profiles" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.648202 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b0837e7-a782-4c89-9f6b-3fe6df3aecb9" containerName="collect-profiles" Feb 02 14:49:36 crc kubenswrapper[4846]: E0202 14:49:36.648237 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerName="registry-server" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.648244 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerName="registry-server" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.648492 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="72962c33-2fa8-4d67-a3ad-81748efcb2e7" containerName="registry-server" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.648516 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b0837e7-a782-4c89-9f6b-3fe6df3aecb9" containerName="collect-profiles" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.650309 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.660930 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fqskx"] Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.662779 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_92f8c952-e300-4858-9bc5-c06350e65d6f/init-config-reloader/0.log" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.696706 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-utilities\") pod \"certified-operators-fqskx\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.696995 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-catalog-content\") pod \"certified-operators-fqskx\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.697145 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6dxl\" (UniqueName: \"kubernetes.io/projected/d742d110-c739-489c-9402-c468ff1a21f8-kube-api-access-t6dxl\") pod \"certified-operators-fqskx\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.824261 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-utilities\") pod \"certified-operators-fqskx\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.824915 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-catalog-content\") pod \"certified-operators-fqskx\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.825091 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6dxl\" (UniqueName: \"kubernetes.io/projected/d742d110-c739-489c-9402-c468ff1a21f8-kube-api-access-t6dxl\") pod \"certified-operators-fqskx\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.826142 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-catalog-content\") pod \"certified-operators-fqskx\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.827832 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-utilities\") pod \"certified-operators-fqskx\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.869093 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6dxl\" (UniqueName: \"kubernetes.io/projected/d742d110-c739-489c-9402-c468ff1a21f8-kube-api-access-t6dxl\") pod \"certified-operators-fqskx\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.985094 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_92f8c952-e300-4858-9bc5-c06350e65d6f/init-config-reloader/0.log" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.985343 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_92f8c952-e300-4858-9bc5-c06350e65d6f/config-reloader/0.log" Feb 02 14:49:36 crc kubenswrapper[4846]: I0202 14:49:36.989634 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:37 crc kubenswrapper[4846]: I0202 14:49:37.122202 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_92f8c952-e300-4858-9bc5-c06350e65d6f/prometheus/0.log" Feb 02 14:49:37 crc kubenswrapper[4846]: I0202 14:49:37.149129 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_92f8c952-e300-4858-9bc5-c06350e65d6f/thanos-sidecar/0.log" Feb 02 14:49:37 crc kubenswrapper[4846]: I0202 14:49:37.410970 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_aaa5fe94-6e26-4ef5-884c-b0731723c243/setup-container/0.log" Feb 02 14:49:37 crc kubenswrapper[4846]: I0202 14:49:37.622333 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fqskx"] Feb 02 14:49:37 crc kubenswrapper[4846]: I0202 14:49:37.743913 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3bf0590a-eb90-4b29-baaf-d4eaa8404e50/setup-container/0.log" Feb 02 14:49:37 crc kubenswrapper[4846]: I0202 14:49:37.767325 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_aaa5fe94-6e26-4ef5-884c-b0731723c243/setup-container/0.log" Feb 02 14:49:37 crc kubenswrapper[4846]: I0202 14:49:37.873881 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_aaa5fe94-6e26-4ef5-884c-b0731723c243/rabbitmq/0.log" Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.097156 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3bf0590a-eb90-4b29-baaf-d4eaa8404e50/setup-container/0.log" Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.156942 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3bf0590a-eb90-4b29-baaf-d4eaa8404e50/rabbitmq/0.log" Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.172800 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-dmq2m_27ffd62f-5922-47de-8a2a-8374a1540888/reboot-os-openstack-openstack-cell1/0.log" Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.213238 4846 generic.go:334] "Generic (PLEG): container finished" podID="d742d110-c739-489c-9402-c468ff1a21f8" containerID="b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b" exitCode=0 Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.213488 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqskx" event={"ID":"d742d110-c739-489c-9402-c468ff1a21f8","Type":"ContainerDied","Data":"b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b"} Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.213513 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqskx" event={"ID":"d742d110-c739-489c-9402-c468ff1a21f8","Type":"ContainerStarted","Data":"0d291e161ffed2210a60ce2574c975f9ea8412282fbacc83d4dc56e74b2869c6"} Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.329542 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-j9gn9_acd9db54-1e35-46b2-8e31-be31c11303a8/run-os-openstack-openstack-cell1/0.log" Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.506143 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-qd5tq_408627ad-8ac6-484a-822c-f8b1aec54895/ssh-known-hosts-openstack/0.log" Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.724744 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6977cb8458-tl6r6_37cf7ba0-1c89-417e-958b-883454800ce9/proxy-server/0.log" Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.885793 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6977cb8458-tl6r6_37cf7ba0-1c89-417e-958b-883454800ce9/proxy-httpd/0.log" Feb 02 14:49:38 crc kubenswrapper[4846]: I0202 14:49:38.973039 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-grx5c_f0c65201-a430-4889-a853-00c23cf8658e/swift-ring-rebalance/0.log" Feb 02 14:49:39 crc kubenswrapper[4846]: I0202 14:49:39.105382 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-b5j8z_7fdfa2d0-e63f-4659-8c81-7bb3dfc07545/telemetry-openstack-openstack-cell1/0.log" Feb 02 14:49:39 crc kubenswrapper[4846]: I0202 14:49:39.127818 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_705ad25e-8635-426c-8069-f09e3b16712f/memcached/0.log" Feb 02 14:49:39 crc kubenswrapper[4846]: I0202 14:49:39.274060 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-65tjw_5e22c32b-e1f8-4bc4-89e1-e27acb1117ef/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Feb 02 14:49:39 crc kubenswrapper[4846]: I0202 14:49:39.341979 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-8hgxz_55b40167-7f63-4a89-8043-2e83365fbc1b/validate-network-openstack-openstack-cell1/0.log" Feb 02 14:49:40 crc kubenswrapper[4846]: I0202 14:49:40.234427 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqskx" event={"ID":"d742d110-c739-489c-9402-c468ff1a21f8","Type":"ContainerStarted","Data":"fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8"} Feb 02 14:49:43 crc kubenswrapper[4846]: I0202 14:49:43.273357 4846 generic.go:334] "Generic (PLEG): container finished" podID="d742d110-c739-489c-9402-c468ff1a21f8" containerID="fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8" exitCode=0 Feb 02 14:49:43 crc kubenswrapper[4846]: I0202 14:49:43.273766 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqskx" event={"ID":"d742d110-c739-489c-9402-c468ff1a21f8","Type":"ContainerDied","Data":"fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8"} Feb 02 14:49:44 crc kubenswrapper[4846]: I0202 14:49:44.289378 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqskx" event={"ID":"d742d110-c739-489c-9402-c468ff1a21f8","Type":"ContainerStarted","Data":"de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d"} Feb 02 14:49:44 crc kubenswrapper[4846]: I0202 14:49:44.316816 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fqskx" podStartSLOduration=2.779410565 podStartE2EDuration="8.316799092s" podCreationTimestamp="2026-02-02 14:49:36 +0000 UTC" firstStartedPulling="2026-02-02 14:49:38.215923661 +0000 UTC m=+9609.444510524" lastFinishedPulling="2026-02-02 14:49:43.753312188 +0000 UTC m=+9614.981899051" observedRunningTime="2026-02-02 14:49:44.308730805 +0000 UTC m=+9615.537317688" watchObservedRunningTime="2026-02-02 14:49:44.316799092 +0000 UTC m=+9615.545385945" Feb 02 14:49:46 crc kubenswrapper[4846]: I0202 14:49:46.992185 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:46 crc kubenswrapper[4846]: I0202 14:49:46.992733 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:47 crc kubenswrapper[4846]: I0202 14:49:47.049234 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:57 crc kubenswrapper[4846]: I0202 14:49:57.052437 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:57 crc kubenswrapper[4846]: I0202 14:49:57.111351 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fqskx"] Feb 02 14:49:57 crc kubenswrapper[4846]: I0202 14:49:57.431646 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fqskx" podUID="d742d110-c739-489c-9402-c468ff1a21f8" containerName="registry-server" containerID="cri-o://de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d" gracePeriod=2 Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.037463 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.054710 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-utilities\") pod \"d742d110-c739-489c-9402-c468ff1a21f8\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.054971 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-catalog-content\") pod \"d742d110-c739-489c-9402-c468ff1a21f8\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.055497 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6dxl\" (UniqueName: \"kubernetes.io/projected/d742d110-c739-489c-9402-c468ff1a21f8-kube-api-access-t6dxl\") pod \"d742d110-c739-489c-9402-c468ff1a21f8\" (UID: \"d742d110-c739-489c-9402-c468ff1a21f8\") " Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.055633 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-utilities" (OuterVolumeSpecName: "utilities") pod "d742d110-c739-489c-9402-c468ff1a21f8" (UID: "d742d110-c739-489c-9402-c468ff1a21f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.056424 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.082938 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d742d110-c739-489c-9402-c468ff1a21f8-kube-api-access-t6dxl" (OuterVolumeSpecName: "kube-api-access-t6dxl") pod "d742d110-c739-489c-9402-c468ff1a21f8" (UID: "d742d110-c739-489c-9402-c468ff1a21f8"). InnerVolumeSpecName "kube-api-access-t6dxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.158728 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6dxl\" (UniqueName: \"kubernetes.io/projected/d742d110-c739-489c-9402-c468ff1a21f8-kube-api-access-t6dxl\") on node \"crc\" DevicePath \"\"" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.166693 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d742d110-c739-489c-9402-c468ff1a21f8" (UID: "d742d110-c739-489c-9402-c468ff1a21f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.261286 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d742d110-c739-489c-9402-c468ff1a21f8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.456511 4846 generic.go:334] "Generic (PLEG): container finished" podID="d742d110-c739-489c-9402-c468ff1a21f8" containerID="de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d" exitCode=0 Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.456583 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqskx" event={"ID":"d742d110-c739-489c-9402-c468ff1a21f8","Type":"ContainerDied","Data":"de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d"} Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.456619 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqskx" event={"ID":"d742d110-c739-489c-9402-c468ff1a21f8","Type":"ContainerDied","Data":"0d291e161ffed2210a60ce2574c975f9ea8412282fbacc83d4dc56e74b2869c6"} Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.456657 4846 scope.go:117] "RemoveContainer" containerID="de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.468174 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqskx" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.529823 4846 scope.go:117] "RemoveContainer" containerID="fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.572817 4846 scope.go:117] "RemoveContainer" containerID="b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.594693 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fqskx"] Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.616708 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fqskx"] Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.630827 4846 scope.go:117] "RemoveContainer" containerID="de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d" Feb 02 14:49:58 crc kubenswrapper[4846]: E0202 14:49:58.634849 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d\": container with ID starting with de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d not found: ID does not exist" containerID="de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.634909 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d"} err="failed to get container status \"de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d\": rpc error: code = NotFound desc = could not find container \"de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d\": container with ID starting with de9a045708c0c24afce163d200d6df0d7e6ae194500e7fd84b052caf9e5dc19d not found: ID does not exist" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.634944 4846 scope.go:117] "RemoveContainer" containerID="fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8" Feb 02 14:49:58 crc kubenswrapper[4846]: E0202 14:49:58.637762 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8\": container with ID starting with fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8 not found: ID does not exist" containerID="fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.637822 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8"} err="failed to get container status \"fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8\": rpc error: code = NotFound desc = could not find container \"fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8\": container with ID starting with fe8dad4063d61cb46164fc4811d53837516b215c515d4f1d4a01a6f0bc5db5e8 not found: ID does not exist" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.637850 4846 scope.go:117] "RemoveContainer" containerID="b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b" Feb 02 14:49:58 crc kubenswrapper[4846]: E0202 14:49:58.641769 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b\": container with ID starting with b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b not found: ID does not exist" containerID="b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b" Feb 02 14:49:58 crc kubenswrapper[4846]: I0202 14:49:58.641816 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b"} err="failed to get container status \"b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b\": rpc error: code = NotFound desc = could not find container \"b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b\": container with ID starting with b88ed21998f0ef0656d55a411f4590ba77e7d3c2de61b13ead9eb7d77c64f96b not found: ID does not exist" Feb 02 14:49:59 crc kubenswrapper[4846]: I0202 14:49:59.441112 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d742d110-c739-489c-9402-c468ff1a21f8" path="/var/lib/kubelet/pods/d742d110-c739-489c-9402-c468ff1a21f8/volumes" Feb 02 14:50:00 crc kubenswrapper[4846]: I0202 14:50:00.479306 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:50:00 crc kubenswrapper[4846]: I0202 14:50:00.479661 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:50:12 crc kubenswrapper[4846]: I0202 14:50:12.518731 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw_0841d2fe-5012-47c0-a591-849d4f8cc0f5/util/0.log" Feb 02 14:50:12 crc kubenswrapper[4846]: I0202 14:50:12.700832 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw_0841d2fe-5012-47c0-a591-849d4f8cc0f5/pull/0.log" Feb 02 14:50:12 crc kubenswrapper[4846]: I0202 14:50:12.739017 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw_0841d2fe-5012-47c0-a591-849d4f8cc0f5/util/0.log" Feb 02 14:50:12 crc kubenswrapper[4846]: I0202 14:50:12.764586 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw_0841d2fe-5012-47c0-a591-849d4f8cc0f5/pull/0.log" Feb 02 14:50:12 crc kubenswrapper[4846]: I0202 14:50:12.950836 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw_0841d2fe-5012-47c0-a591-849d4f8cc0f5/extract/0.log" Feb 02 14:50:12 crc kubenswrapper[4846]: I0202 14:50:12.957025 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw_0841d2fe-5012-47c0-a591-849d4f8cc0f5/util/0.log" Feb 02 14:50:12 crc kubenswrapper[4846]: I0202 14:50:12.995588 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_805de4b7130efd4fbe2290d0c51d78e6295b83d77efc3ff2d1b014ffe7pwxzw_0841d2fe-5012-47c0-a591-849d4f8cc0f5/pull/0.log" Feb 02 14:50:13 crc kubenswrapper[4846]: I0202 14:50:13.363491 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b6c4d8c5f-8lrz8_c0f1cd1a-da1b-4292-8622-87119d40ec03/manager/0.log" Feb 02 14:50:13 crc kubenswrapper[4846]: I0202 14:50:13.373388 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d874c8fc-hc6td_6ba95849-98ff-4d50-8c41-605aa99c3b2f/manager/0.log" Feb 02 14:50:13 crc kubenswrapper[4846]: I0202 14:50:13.444494 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d9697b7f4-88lrz_fced078c-1f83-492e-9f21-e9e82c1f9275/manager/0.log" Feb 02 14:50:13 crc kubenswrapper[4846]: I0202 14:50:13.728287 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8886f4c47-bklmw_b9c7230d-0e38-4540-9074-09a47500dd40/manager/0.log" Feb 02 14:50:13 crc kubenswrapper[4846]: I0202 14:50:13.769963 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69d6db494d-jscz6_478eaaaa-5884-4b31-a87e-8655bedb96f1/manager/0.log" Feb 02 14:50:13 crc kubenswrapper[4846]: I0202 14:50:13.937914 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5fb775575f-cz76f_46d6dfa1-87b3-4e53-bcfd-27c709ad7911/manager/0.log" Feb 02 14:50:14 crc kubenswrapper[4846]: I0202 14:50:14.200167 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f4b8bd54d-7lr2n_69f3343d-ca8a-4ff2-bc0c-14121a6938a8/manager/0.log" Feb 02 14:50:14 crc kubenswrapper[4846]: I0202 14:50:14.597537 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7dd968899f-5vzvg_d6f0f90b-675e-40e9-b541-7ff04333a885/manager/0.log" Feb 02 14:50:14 crc kubenswrapper[4846]: I0202 14:50:14.626404 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-84f48565d4-hfvt6_a70f0094-0775-49ea-9002-9f3da3ff87d1/manager/0.log" Feb 02 14:50:14 crc kubenswrapper[4846]: I0202 14:50:14.699239 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79955696d6-c6sk4_f75d8c59-636e-4496-bd81-e4e53f37a713/manager/0.log" Feb 02 14:50:15 crc kubenswrapper[4846]: I0202 14:50:15.176998 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf948998-nvnpj_2e0bfdb6-f6e5-41c4-ab53-f13863d687b8/manager/0.log" Feb 02 14:50:15 crc kubenswrapper[4846]: I0202 14:50:15.276126 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-585dbc889-hrnrk_f8ebd280-c5fa-4cf3-890b-93a018bcb2d2/manager/0.log" Feb 02 14:50:15 crc kubenswrapper[4846]: I0202 14:50:15.533194 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-55bff696bd-89dd7_f4ec43b3-d763-4478-8ef9-0453a2fb730a/manager/0.log" Feb 02 14:50:15 crc kubenswrapper[4846]: I0202 14:50:15.597266 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6687f8d877-vfmzp_8d9a7d64-a757-48b8-bdb9-4ec3e2921321/manager/0.log" Feb 02 14:50:15 crc kubenswrapper[4846]: I0202 14:50:15.763902 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5ff45dfdbf6hgrw_b75ce3f5-f42f-4c8f-aa7f-6e8bc0a08ee1/manager/0.log" Feb 02 14:50:15 crc kubenswrapper[4846]: I0202 14:50:15.966600 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6bf6665fd-bzplf_80b73f30-8c98-477e-9f49-f7a39a5cf559/operator/0.log" Feb 02 14:50:16 crc kubenswrapper[4846]: I0202 14:50:16.436201 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xl544_122f29f2-1a54-4f6f-a913-24fb8948af5d/registry-server/0.log" Feb 02 14:50:16 crc kubenswrapper[4846]: I0202 14:50:16.674559 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-788c46999f-tw44b_2497b300-a910-41c8-b70d-1c9acd871a90/manager/0.log" Feb 02 14:50:16 crc kubenswrapper[4846]: I0202 14:50:16.703652 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b964cf4cd-pqnqm_124b64f2-5773-4d35-ab14-f4ad0fe4f318/manager/0.log" Feb 02 14:50:17 crc kubenswrapper[4846]: I0202 14:50:17.052144 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68fc8c869-kkbb8_9c5ac153-1fd9-491b-942d-2b0b3e3e9da5/manager/0.log" Feb 02 14:50:17 crc kubenswrapper[4846]: I0202 14:50:17.113523 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-9tnhp_b45d49d2-cad0-49fa-addf-766dba1544a4/operator/0.log" Feb 02 14:50:17 crc kubenswrapper[4846]: I0202 14:50:17.424742 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56f8bfcd9f-6g958_204fa43c-5360-4282-b002-ae22e0f0595d/manager/0.log" Feb 02 14:50:17 crc kubenswrapper[4846]: I0202 14:50:17.687012 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-64b5b76f97-xk9n4_cb8f2427-7089-4ecf-8ae9-4ece48067de8/manager/0.log" Feb 02 14:50:17 crc kubenswrapper[4846]: I0202 14:50:17.734696 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-564965969-8gkkw_8ee9f8c5-8078-4ad1-b7c8-4a8e2dfd4e41/manager/0.log" Feb 02 14:50:18 crc kubenswrapper[4846]: I0202 14:50:18.830237 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-646f757d77-rlzgs_b08a9394-cfb5-4249-8f6c-a3fbe49bbeea/manager/0.log" Feb 02 14:50:30 crc kubenswrapper[4846]: I0202 14:50:30.479008 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:50:30 crc kubenswrapper[4846]: I0202 14:50:30.479546 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:50:30 crc kubenswrapper[4846]: I0202 14:50:30.479613 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 14:50:30 crc kubenswrapper[4846]: I0202 14:50:30.480358 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"229e52826f90f0a1025ad49518481c25a85ee892227df94a2914ba07fbeeb68e"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 14:50:30 crc kubenswrapper[4846]: I0202 14:50:30.480422 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://229e52826f90f0a1025ad49518481c25a85ee892227df94a2914ba07fbeeb68e" gracePeriod=600 Feb 02 14:50:31 crc kubenswrapper[4846]: I0202 14:50:31.791149 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="229e52826f90f0a1025ad49518481c25a85ee892227df94a2914ba07fbeeb68e" exitCode=0 Feb 02 14:50:31 crc kubenswrapper[4846]: I0202 14:50:31.791377 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"229e52826f90f0a1025ad49518481c25a85ee892227df94a2914ba07fbeeb68e"} Feb 02 14:50:31 crc kubenswrapper[4846]: I0202 14:50:31.791869 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerStarted","Data":"c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f"} Feb 02 14:50:31 crc kubenswrapper[4846]: I0202 14:50:31.791891 4846 scope.go:117] "RemoveContainer" containerID="7eb732be88005a8f007449073cc6a49eb144ad70a858d15c49384319df866288" Feb 02 14:50:40 crc kubenswrapper[4846]: I0202 14:50:40.593058 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6w8zm_9a4cbd38-0dde-440f-96a3-4b0bd3c2da81/control-plane-machine-set-operator/0.log" Feb 02 14:50:40 crc kubenswrapper[4846]: I0202 14:50:40.849588 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7qdsx_5256ddb9-8edb-42a8-a041-cb806d849e58/machine-api-operator/0.log" Feb 02 14:50:40 crc kubenswrapper[4846]: I0202 14:50:40.862050 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7qdsx_5256ddb9-8edb-42a8-a041-cb806d849e58/kube-rbac-proxy/0.log" Feb 02 14:50:56 crc kubenswrapper[4846]: I0202 14:50:56.624597 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-vghnr_5ccfdad1-2b84-4146-8c7f-5bfbf8009c66/cert-manager-cainjector/0.log" Feb 02 14:50:56 crc kubenswrapper[4846]: I0202 14:50:56.674817 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-nb4jl_f6b1cf52-1897-4864-8335-257ac2391d17/cert-manager-controller/0.log" Feb 02 14:50:56 crc kubenswrapper[4846]: I0202 14:50:56.826141 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-65bwf_2ce8cd4d-d1ff-4a6b-aea1-7dc2396b1ef1/cert-manager-webhook/0.log" Feb 02 14:51:11 crc kubenswrapper[4846]: I0202 14:51:11.979899 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-gpb7r_f7d3eab5-fef3-48e9-aa77-dbaf5720fb6f/nmstate-console-plugin/0.log" Feb 02 14:51:12 crc kubenswrapper[4846]: I0202 14:51:12.359820 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7t7k9_f34abdcc-0d3c-4133-8437-7e354fb8cea7/nmstate-handler/0.log" Feb 02 14:51:12 crc kubenswrapper[4846]: I0202 14:51:12.441612 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-jzshh_395d0b79-6975-4843-9a3f-f53121bfc1bc/kube-rbac-proxy/0.log" Feb 02 14:51:12 crc kubenswrapper[4846]: I0202 14:51:12.497264 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-jzshh_395d0b79-6975-4843-9a3f-f53121bfc1bc/nmstate-metrics/0.log" Feb 02 14:51:12 crc kubenswrapper[4846]: I0202 14:51:12.684768 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-dgl5b_39214593-1488-4797-8b63-a77baba63417/nmstate-operator/0.log" Feb 02 14:51:12 crc kubenswrapper[4846]: I0202 14:51:12.741786 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-qhlrg_5f0ff130-2029-4436-8be0-a49dc96987b0/nmstate-webhook/0.log" Feb 02 14:51:27 crc kubenswrapper[4846]: I0202 14:51:27.342967 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-4d59k_80118cc5-e446-4a77-9b65-8d50ac60308f/prometheus-operator/0.log" Feb 02 14:51:27 crc kubenswrapper[4846]: I0202 14:51:27.375307 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt_dd224d4b-1d4c-403e-89e9-b7fda056a631/prometheus-operator-admission-webhook/0.log" Feb 02 14:51:27 crc kubenswrapper[4846]: I0202 14:51:27.571143 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt_87119693-d4c2-4f7c-bbc1-d73409d91508/prometheus-operator-admission-webhook/0.log" Feb 02 14:51:27 crc kubenswrapper[4846]: I0202 14:51:27.612545 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-2fxhf_ca5f1eb1-41d2-4db2-8f80-953bada112f4/operator/0.log" Feb 02 14:51:27 crc kubenswrapper[4846]: I0202 14:51:27.761193 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-cb9p7_c7fb2f0a-18d3-46ff-a910-de6aa48e71eb/perses-operator/0.log" Feb 02 14:51:41 crc kubenswrapper[4846]: I0202 14:51:41.767136 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-lrmwh_529cfe45-6776-47c3-8cc4-0c5cbcb6e130/kube-rbac-proxy/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.123961 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-frr-files/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.316310 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-lrmwh_529cfe45-6776-47c3-8cc4-0c5cbcb6e130/controller/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.365212 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-reloader/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.387983 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-metrics/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.403422 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-frr-files/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.505212 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-reloader/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.707841 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-frr-files/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.768860 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-reloader/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.783143 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-metrics/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.832430 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-metrics/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.934352 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-reloader/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.948212 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-frr-files/0.log" Feb 02 14:51:42 crc kubenswrapper[4846]: I0202 14:51:42.997342 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/cp-metrics/0.log" Feb 02 14:51:43 crc kubenswrapper[4846]: I0202 14:51:43.023811 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/controller/0.log" Feb 02 14:51:43 crc kubenswrapper[4846]: I0202 14:51:43.180569 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/frr-metrics/0.log" Feb 02 14:51:43 crc kubenswrapper[4846]: I0202 14:51:43.262164 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/kube-rbac-proxy/0.log" Feb 02 14:51:43 crc kubenswrapper[4846]: I0202 14:51:43.298199 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/kube-rbac-proxy-frr/0.log" Feb 02 14:51:43 crc kubenswrapper[4846]: I0202 14:51:43.421774 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/reloader/0.log" Feb 02 14:51:43 crc kubenswrapper[4846]: I0202 14:51:43.627878 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-2r6kw_162e27cd-61d6-4f44-801c-775077032ce7/frr-k8s-webhook-server/0.log" Feb 02 14:51:43 crc kubenswrapper[4846]: I0202 14:51:43.803822 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5b7785ffb5-mxf8t_a520c385-fbf3-4941-968c-2163e685af23/manager/0.log" Feb 02 14:51:43 crc kubenswrapper[4846]: I0202 14:51:43.916188 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-76d4f8c956-swps9_4efa0217-6d2e-4857-8585-559854a1df0f/webhook-server/0.log" Feb 02 14:51:44 crc kubenswrapper[4846]: I0202 14:51:44.199246 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5lpq4_77e7c4a3-0018-438a-9163-5a155ab480d1/kube-rbac-proxy/0.log" Feb 02 14:51:45 crc kubenswrapper[4846]: I0202 14:51:45.155981 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5lpq4_77e7c4a3-0018-438a-9163-5a155ab480d1/speaker/0.log" Feb 02 14:51:46 crc kubenswrapper[4846]: I0202 14:51:46.825398 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r692g_ec7b6ba3-8411-4248-ad6e-d684293002c4/frr/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.132270 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc_e2281eb7-94f9-4e30-85c7-6c7a22385c16/util/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.289411 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc_e2281eb7-94f9-4e30-85c7-6c7a22385c16/util/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.331456 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc_e2281eb7-94f9-4e30-85c7-6c7a22385c16/pull/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.373885 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc_e2281eb7-94f9-4e30-85c7-6c7a22385c16/pull/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.519067 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc_e2281eb7-94f9-4e30-85c7-6c7a22385c16/util/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.559254 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc_e2281eb7-94f9-4e30-85c7-6c7a22385c16/pull/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.563275 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dctgdmc_e2281eb7-94f9-4e30-85c7-6c7a22385c16/extract/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.706276 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx_b7413188-295c-46c8-b396-cbfc04fc1178/util/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.857477 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx_b7413188-295c-46c8-b396-cbfc04fc1178/pull/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.878561 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx_b7413188-295c-46c8-b396-cbfc04fc1178/util/0.log" Feb 02 14:51:58 crc kubenswrapper[4846]: I0202 14:51:58.888297 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx_b7413188-295c-46c8-b396-cbfc04fc1178/pull/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.042774 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx_b7413188-295c-46c8-b396-cbfc04fc1178/pull/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.064143 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx_b7413188-295c-46c8-b396-cbfc04fc1178/extract/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.099762 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713f8mpx_b7413188-295c-46c8-b396-cbfc04fc1178/util/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.190124 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q_9aa5a1ea-0fe6-4b98-adf6-a6af93581534/util/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.417885 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q_9aa5a1ea-0fe6-4b98-adf6-a6af93581534/util/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.421138 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q_9aa5a1ea-0fe6-4b98-adf6-a6af93581534/pull/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.465591 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q_9aa5a1ea-0fe6-4b98-adf6-a6af93581534/pull/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.637449 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q_9aa5a1ea-0fe6-4b98-adf6-a6af93581534/util/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.655816 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q_9aa5a1ea-0fe6-4b98-adf6-a6af93581534/pull/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.688330 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e575b7q_9aa5a1ea-0fe6-4b98-adf6-a6af93581534/extract/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.804669 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf_816eb531-874d-4167-8351-ceef3cd2704a/util/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.971339 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf_816eb531-874d-4167-8351-ceef3cd2704a/pull/0.log" Feb 02 14:51:59 crc kubenswrapper[4846]: I0202 14:51:59.983971 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf_816eb531-874d-4167-8351-ceef3cd2704a/util/0.log" Feb 02 14:52:00 crc kubenswrapper[4846]: I0202 14:52:00.016362 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf_816eb531-874d-4167-8351-ceef3cd2704a/pull/0.log" Feb 02 14:52:00 crc kubenswrapper[4846]: I0202 14:52:00.166441 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf_816eb531-874d-4167-8351-ceef3cd2704a/util/0.log" Feb 02 14:52:00 crc kubenswrapper[4846]: I0202 14:52:00.191357 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf_816eb531-874d-4167-8351-ceef3cd2704a/extract/0.log" Feb 02 14:52:00 crc kubenswrapper[4846]: I0202 14:52:00.220837 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q29lf_816eb531-874d-4167-8351-ceef3cd2704a/pull/0.log" Feb 02 14:52:00 crc kubenswrapper[4846]: I0202 14:52:00.361756 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ks6t8_2f68e1c1-ce23-4532-9dd5-5a195a959e84/extract-utilities/0.log" Feb 02 14:52:00 crc kubenswrapper[4846]: I0202 14:52:00.564053 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ks6t8_2f68e1c1-ce23-4532-9dd5-5a195a959e84/extract-utilities/0.log" Feb 02 14:52:00 crc kubenswrapper[4846]: I0202 14:52:00.583448 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ks6t8_2f68e1c1-ce23-4532-9dd5-5a195a959e84/extract-content/0.log" Feb 02 14:52:00 crc kubenswrapper[4846]: I0202 14:52:00.588951 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ks6t8_2f68e1c1-ce23-4532-9dd5-5a195a959e84/extract-content/0.log" Feb 02 14:52:00 crc kubenswrapper[4846]: I0202 14:52:00.779585 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ks6t8_2f68e1c1-ce23-4532-9dd5-5a195a959e84/extract-content/0.log" Feb 02 14:52:00 crc kubenswrapper[4846]: I0202 14:52:00.862570 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ks6t8_2f68e1c1-ce23-4532-9dd5-5a195a959e84/extract-utilities/0.log" Feb 02 14:52:01 crc kubenswrapper[4846]: I0202 14:52:01.026959 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dwjk4_5ec82720-fb76-468d-88d2-682ab80aab08/extract-utilities/0.log" Feb 02 14:52:01 crc kubenswrapper[4846]: I0202 14:52:01.331690 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dwjk4_5ec82720-fb76-468d-88d2-682ab80aab08/extract-content/0.log" Feb 02 14:52:01 crc kubenswrapper[4846]: I0202 14:52:01.358954 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dwjk4_5ec82720-fb76-468d-88d2-682ab80aab08/extract-content/0.log" Feb 02 14:52:01 crc kubenswrapper[4846]: I0202 14:52:01.358997 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dwjk4_5ec82720-fb76-468d-88d2-682ab80aab08/extract-utilities/0.log" Feb 02 14:52:01 crc kubenswrapper[4846]: I0202 14:52:01.542995 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dwjk4_5ec82720-fb76-468d-88d2-682ab80aab08/extract-utilities/0.log" Feb 02 14:52:01 crc kubenswrapper[4846]: I0202 14:52:01.581237 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dwjk4_5ec82720-fb76-468d-88d2-682ab80aab08/extract-content/0.log" Feb 02 14:52:01 crc kubenswrapper[4846]: I0202 14:52:01.702558 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ks6t8_2f68e1c1-ce23-4532-9dd5-5a195a959e84/registry-server/0.log" Feb 02 14:52:01 crc kubenswrapper[4846]: I0202 14:52:01.811704 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kgbzq_17938097-4a7b-4e18-9e26-4ada5f07b343/marketplace-operator/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.086488 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r7jc8_de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf/extract-utilities/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.233765 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r7jc8_de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf/extract-content/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.241708 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r7jc8_de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf/extract-utilities/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.280108 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r7jc8_de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf/extract-content/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.464077 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r7jc8_de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf/extract-utilities/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.513455 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r7jc8_de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf/extract-content/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.696845 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48ctv_26a0bb46-1e22-4ff5-bc22-d54540212e99/extract-utilities/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.907713 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48ctv_26a0bb46-1e22-4ff5-bc22-d54540212e99/extract-content/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.921396 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48ctv_26a0bb46-1e22-4ff5-bc22-d54540212e99/extract-utilities/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.963124 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48ctv_26a0bb46-1e22-4ff5-bc22-d54540212e99/extract-content/0.log" Feb 02 14:52:02 crc kubenswrapper[4846]: I0202 14:52:02.990413 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-r7jc8_de0c2e42-9e1c-4d7d-9b86-106d3bbfddbf/registry-server/0.log" Feb 02 14:52:03 crc kubenswrapper[4846]: I0202 14:52:03.145741 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48ctv_26a0bb46-1e22-4ff5-bc22-d54540212e99/extract-utilities/0.log" Feb 02 14:52:03 crc kubenswrapper[4846]: I0202 14:52:03.170773 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48ctv_26a0bb46-1e22-4ff5-bc22-d54540212e99/extract-content/0.log" Feb 02 14:52:03 crc kubenswrapper[4846]: I0202 14:52:03.201074 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dwjk4_5ec82720-fb76-468d-88d2-682ab80aab08/registry-server/0.log" Feb 02 14:52:03 crc kubenswrapper[4846]: I0202 14:52:03.627332 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48ctv_26a0bb46-1e22-4ff5-bc22-d54540212e99/registry-server/0.log" Feb 02 14:52:16 crc kubenswrapper[4846]: I0202 14:52:16.901554 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-b9fdc8658-4hvkt_dd224d4b-1d4c-403e-89e9-b7fda056a631/prometheus-operator-admission-webhook/0.log" Feb 02 14:52:16 crc kubenswrapper[4846]: I0202 14:52:16.941561 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-4d59k_80118cc5-e446-4a77-9b65-8d50ac60308f/prometheus-operator/0.log" Feb 02 14:52:17 crc kubenswrapper[4846]: I0202 14:52:17.003720 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-b9fdc8658-bl8jt_87119693-d4c2-4f7c-bbc1-d73409d91508/prometheus-operator-admission-webhook/0.log" Feb 02 14:52:17 crc kubenswrapper[4846]: I0202 14:52:17.187639 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-2fxhf_ca5f1eb1-41d2-4db2-8f80-953bada112f4/operator/0.log" Feb 02 14:52:17 crc kubenswrapper[4846]: I0202 14:52:17.244429 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-cb9p7_c7fb2f0a-18d3-46ff-a910-de6aa48e71eb/perses-operator/0.log" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.229849 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6f29n"] Feb 02 14:52:35 crc kubenswrapper[4846]: E0202 14:52:35.230749 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d742d110-c739-489c-9402-c468ff1a21f8" containerName="registry-server" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.230761 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d742d110-c739-489c-9402-c468ff1a21f8" containerName="registry-server" Feb 02 14:52:35 crc kubenswrapper[4846]: E0202 14:52:35.230808 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d742d110-c739-489c-9402-c468ff1a21f8" containerName="extract-content" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.230814 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d742d110-c739-489c-9402-c468ff1a21f8" containerName="extract-content" Feb 02 14:52:35 crc kubenswrapper[4846]: E0202 14:52:35.230827 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d742d110-c739-489c-9402-c468ff1a21f8" containerName="extract-utilities" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.230833 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d742d110-c739-489c-9402-c468ff1a21f8" containerName="extract-utilities" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.231032 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d742d110-c739-489c-9402-c468ff1a21f8" containerName="registry-server" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.232633 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.265824 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6f29n"] Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.392342 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpnq2\" (UniqueName: \"kubernetes.io/projected/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-kube-api-access-tpnq2\") pod \"community-operators-6f29n\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.392579 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-catalog-content\") pod \"community-operators-6f29n\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.392809 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-utilities\") pod \"community-operators-6f29n\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.494742 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpnq2\" (UniqueName: \"kubernetes.io/projected/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-kube-api-access-tpnq2\") pod \"community-operators-6f29n\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.494897 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-catalog-content\") pod \"community-operators-6f29n\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.494969 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-utilities\") pod \"community-operators-6f29n\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.495463 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-catalog-content\") pod \"community-operators-6f29n\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.495472 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-utilities\") pod \"community-operators-6f29n\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.520042 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpnq2\" (UniqueName: \"kubernetes.io/projected/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-kube-api-access-tpnq2\") pod \"community-operators-6f29n\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:35 crc kubenswrapper[4846]: I0202 14:52:35.555582 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:36 crc kubenswrapper[4846]: I0202 14:52:36.309726 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6f29n"] Feb 02 14:52:37 crc kubenswrapper[4846]: I0202 14:52:37.144901 4846 generic.go:334] "Generic (PLEG): container finished" podID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerID="6692108895a0c9e58ce3b26eac41df221281466f5012b676335ddc806d16194c" exitCode=0 Feb 02 14:52:37 crc kubenswrapper[4846]: I0202 14:52:37.145081 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f29n" event={"ID":"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1","Type":"ContainerDied","Data":"6692108895a0c9e58ce3b26eac41df221281466f5012b676335ddc806d16194c"} Feb 02 14:52:37 crc kubenswrapper[4846]: I0202 14:52:37.145220 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f29n" event={"ID":"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1","Type":"ContainerStarted","Data":"07e23219c2f97dde20444dde44ba9d9e0034b64435cd73f086b0c61d4d0a3c85"} Feb 02 14:52:37 crc kubenswrapper[4846]: I0202 14:52:37.148198 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 14:52:39 crc kubenswrapper[4846]: I0202 14:52:39.175344 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f29n" event={"ID":"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1","Type":"ContainerStarted","Data":"19f55e54d22b070301e873d9f533bab044280ec6ceff93360233e37fb32e1695"} Feb 02 14:52:41 crc kubenswrapper[4846]: I0202 14:52:41.194686 4846 generic.go:334] "Generic (PLEG): container finished" podID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerID="19f55e54d22b070301e873d9f533bab044280ec6ceff93360233e37fb32e1695" exitCode=0 Feb 02 14:52:41 crc kubenswrapper[4846]: I0202 14:52:41.195345 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f29n" event={"ID":"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1","Type":"ContainerDied","Data":"19f55e54d22b070301e873d9f533bab044280ec6ceff93360233e37fb32e1695"} Feb 02 14:52:42 crc kubenswrapper[4846]: I0202 14:52:42.207213 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f29n" event={"ID":"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1","Type":"ContainerStarted","Data":"b0f4ccf3391a6e6afcb89e05be777cba25513dc7818ff2ecf15b5535b225930d"} Feb 02 14:52:42 crc kubenswrapper[4846]: I0202 14:52:42.241119 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6f29n" podStartSLOduration=2.7787144169999998 podStartE2EDuration="7.241093214s" podCreationTimestamp="2026-02-02 14:52:35 +0000 UTC" firstStartedPulling="2026-02-02 14:52:37.14795435 +0000 UTC m=+9788.376541203" lastFinishedPulling="2026-02-02 14:52:41.610333137 +0000 UTC m=+9792.838920000" observedRunningTime="2026-02-02 14:52:42.228757762 +0000 UTC m=+9793.457344625" watchObservedRunningTime="2026-02-02 14:52:42.241093214 +0000 UTC m=+9793.469680077" Feb 02 14:52:45 crc kubenswrapper[4846]: I0202 14:52:45.556618 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:45 crc kubenswrapper[4846]: I0202 14:52:45.557166 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:45 crc kubenswrapper[4846]: I0202 14:52:45.607507 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:46 crc kubenswrapper[4846]: I0202 14:52:46.320486 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:46 crc kubenswrapper[4846]: I0202 14:52:46.406322 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6f29n"] Feb 02 14:52:48 crc kubenswrapper[4846]: I0202 14:52:48.266413 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6f29n" podUID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerName="registry-server" containerID="cri-o://b0f4ccf3391a6e6afcb89e05be777cba25513dc7818ff2ecf15b5535b225930d" gracePeriod=2 Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.286498 4846 generic.go:334] "Generic (PLEG): container finished" podID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerID="b0f4ccf3391a6e6afcb89e05be777cba25513dc7818ff2ecf15b5535b225930d" exitCode=0 Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.286585 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f29n" event={"ID":"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1","Type":"ContainerDied","Data":"b0f4ccf3391a6e6afcb89e05be777cba25513dc7818ff2ecf15b5535b225930d"} Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.287070 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6f29n" event={"ID":"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1","Type":"ContainerDied","Data":"07e23219c2f97dde20444dde44ba9d9e0034b64435cd73f086b0c61d4d0a3c85"} Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.287090 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07e23219c2f97dde20444dde44ba9d9e0034b64435cd73f086b0c61d4d0a3c85" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.361724 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kwszx"] Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.365719 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.372426 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.391930 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kwszx"] Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.416741 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-utilities\") pod \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.416815 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpnq2\" (UniqueName: \"kubernetes.io/projected/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-kube-api-access-tpnq2\") pod \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.416873 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-catalog-content\") pod \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\" (UID: \"cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1\") " Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.417032 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-catalog-content\") pod \"redhat-operators-kwszx\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.417194 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nkqn\" (UniqueName: \"kubernetes.io/projected/624db4c3-82c4-4546-90b8-606f72444e4f-kube-api-access-2nkqn\") pod \"redhat-operators-kwszx\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.417243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-utilities\") pod \"redhat-operators-kwszx\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.417852 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-utilities" (OuterVolumeSpecName: "utilities") pod "cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" (UID: "cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.444866 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-kube-api-access-tpnq2" (OuterVolumeSpecName: "kube-api-access-tpnq2") pod "cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" (UID: "cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1"). InnerVolumeSpecName "kube-api-access-tpnq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.520561 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nkqn\" (UniqueName: \"kubernetes.io/projected/624db4c3-82c4-4546-90b8-606f72444e4f-kube-api-access-2nkqn\") pod \"redhat-operators-kwszx\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.520963 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-utilities\") pod \"redhat-operators-kwszx\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.521090 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-catalog-content\") pod \"redhat-operators-kwszx\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.521309 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.521326 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpnq2\" (UniqueName: \"kubernetes.io/projected/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-kube-api-access-tpnq2\") on node \"crc\" DevicePath \"\"" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.521665 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-utilities\") pod \"redhat-operators-kwszx\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.521952 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-catalog-content\") pod \"redhat-operators-kwszx\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.553009 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nkqn\" (UniqueName: \"kubernetes.io/projected/624db4c3-82c4-4546-90b8-606f72444e4f-kube-api-access-2nkqn\") pod \"redhat-operators-kwszx\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.693378 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.738579 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" (UID: "cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:52:49 crc kubenswrapper[4846]: I0202 14:52:49.829633 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:52:50 crc kubenswrapper[4846]: I0202 14:52:50.295261 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6f29n" Feb 02 14:52:50 crc kubenswrapper[4846]: I0202 14:52:50.347593 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kwszx"] Feb 02 14:52:50 crc kubenswrapper[4846]: I0202 14:52:50.372975 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6f29n"] Feb 02 14:52:50 crc kubenswrapper[4846]: I0202 14:52:50.399821 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6f29n"] Feb 02 14:52:51 crc kubenswrapper[4846]: I0202 14:52:51.307156 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwszx" event={"ID":"624db4c3-82c4-4546-90b8-606f72444e4f","Type":"ContainerStarted","Data":"7d115629d4bccd0a8313818d34442c77f924cc4cf96b4d3d8391e184cfa07ec3"} Feb 02 14:52:51 crc kubenswrapper[4846]: I0202 14:52:51.307549 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwszx" event={"ID":"624db4c3-82c4-4546-90b8-606f72444e4f","Type":"ContainerStarted","Data":"0b9718f1db1473d5bcca23261160bcb5cd5353c107f2fca2ad9974ecb0959309"} Feb 02 14:52:51 crc kubenswrapper[4846]: I0202 14:52:51.439803 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" path="/var/lib/kubelet/pods/cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1/volumes" Feb 02 14:52:52 crc kubenswrapper[4846]: I0202 14:52:52.321102 4846 generic.go:334] "Generic (PLEG): container finished" podID="624db4c3-82c4-4546-90b8-606f72444e4f" containerID="7d115629d4bccd0a8313818d34442c77f924cc4cf96b4d3d8391e184cfa07ec3" exitCode=0 Feb 02 14:52:52 crc kubenswrapper[4846]: I0202 14:52:52.321726 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwszx" event={"ID":"624db4c3-82c4-4546-90b8-606f72444e4f","Type":"ContainerDied","Data":"7d115629d4bccd0a8313818d34442c77f924cc4cf96b4d3d8391e184cfa07ec3"} Feb 02 14:52:54 crc kubenswrapper[4846]: I0202 14:52:54.344388 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwszx" event={"ID":"624db4c3-82c4-4546-90b8-606f72444e4f","Type":"ContainerStarted","Data":"bec962f6dd18ab9ba405faba949ef07e1d87ca13cf23d63924fb0d874fe1a7e3"} Feb 02 14:52:59 crc kubenswrapper[4846]: I0202 14:52:59.393360 4846 generic.go:334] "Generic (PLEG): container finished" podID="624db4c3-82c4-4546-90b8-606f72444e4f" containerID="bec962f6dd18ab9ba405faba949ef07e1d87ca13cf23d63924fb0d874fe1a7e3" exitCode=0 Feb 02 14:52:59 crc kubenswrapper[4846]: I0202 14:52:59.393743 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwszx" event={"ID":"624db4c3-82c4-4546-90b8-606f72444e4f","Type":"ContainerDied","Data":"bec962f6dd18ab9ba405faba949ef07e1d87ca13cf23d63924fb0d874fe1a7e3"} Feb 02 14:53:00 crc kubenswrapper[4846]: I0202 14:53:00.406986 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwszx" event={"ID":"624db4c3-82c4-4546-90b8-606f72444e4f","Type":"ContainerStarted","Data":"7ebfefff6b59a0370fa09d13daeb29cca946bd85fc737ae8db2c0f1fb24cd6c0"} Feb 02 14:53:00 crc kubenswrapper[4846]: I0202 14:53:00.434145 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kwszx" podStartSLOduration=3.802970319 podStartE2EDuration="11.434125545s" podCreationTimestamp="2026-02-02 14:52:49 +0000 UTC" firstStartedPulling="2026-02-02 14:52:52.324206693 +0000 UTC m=+9803.552793556" lastFinishedPulling="2026-02-02 14:52:59.955361919 +0000 UTC m=+9811.183948782" observedRunningTime="2026-02-02 14:53:00.430544957 +0000 UTC m=+9811.659131840" watchObservedRunningTime="2026-02-02 14:53:00.434125545 +0000 UTC m=+9811.662712418" Feb 02 14:53:00 crc kubenswrapper[4846]: I0202 14:53:00.479141 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:53:00 crc kubenswrapper[4846]: I0202 14:53:00.479207 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:53:09 crc kubenswrapper[4846]: I0202 14:53:09.694242 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:53:09 crc kubenswrapper[4846]: I0202 14:53:09.694899 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:53:09 crc kubenswrapper[4846]: I0202 14:53:09.757308 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:53:11 crc kubenswrapper[4846]: I0202 14:53:11.052476 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:53:11 crc kubenswrapper[4846]: I0202 14:53:11.109018 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kwszx"] Feb 02 14:53:12 crc kubenswrapper[4846]: I0202 14:53:12.538426 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kwszx" podUID="624db4c3-82c4-4546-90b8-606f72444e4f" containerName="registry-server" containerID="cri-o://7ebfefff6b59a0370fa09d13daeb29cca946bd85fc737ae8db2c0f1fb24cd6c0" gracePeriod=2 Feb 02 14:53:13 crc kubenswrapper[4846]: I0202 14:53:13.552702 4846 generic.go:334] "Generic (PLEG): container finished" podID="624db4c3-82c4-4546-90b8-606f72444e4f" containerID="7ebfefff6b59a0370fa09d13daeb29cca946bd85fc737ae8db2c0f1fb24cd6c0" exitCode=0 Feb 02 14:53:13 crc kubenswrapper[4846]: I0202 14:53:13.552782 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwszx" event={"ID":"624db4c3-82c4-4546-90b8-606f72444e4f","Type":"ContainerDied","Data":"7ebfefff6b59a0370fa09d13daeb29cca946bd85fc737ae8db2c0f1fb24cd6c0"} Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.127308 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.185469 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nkqn\" (UniqueName: \"kubernetes.io/projected/624db4c3-82c4-4546-90b8-606f72444e4f-kube-api-access-2nkqn\") pod \"624db4c3-82c4-4546-90b8-606f72444e4f\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.185671 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-utilities\") pod \"624db4c3-82c4-4546-90b8-606f72444e4f\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.185762 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-catalog-content\") pod \"624db4c3-82c4-4546-90b8-606f72444e4f\" (UID: \"624db4c3-82c4-4546-90b8-606f72444e4f\") " Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.186520 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-utilities" (OuterVolumeSpecName: "utilities") pod "624db4c3-82c4-4546-90b8-606f72444e4f" (UID: "624db4c3-82c4-4546-90b8-606f72444e4f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.187400 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.191034 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/624db4c3-82c4-4546-90b8-606f72444e4f-kube-api-access-2nkqn" (OuterVolumeSpecName: "kube-api-access-2nkqn") pod "624db4c3-82c4-4546-90b8-606f72444e4f" (UID: "624db4c3-82c4-4546-90b8-606f72444e4f"). InnerVolumeSpecName "kube-api-access-2nkqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.289273 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nkqn\" (UniqueName: \"kubernetes.io/projected/624db4c3-82c4-4546-90b8-606f72444e4f-kube-api-access-2nkqn\") on node \"crc\" DevicePath \"\"" Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.372052 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "624db4c3-82c4-4546-90b8-606f72444e4f" (UID: "624db4c3-82c4-4546-90b8-606f72444e4f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.390850 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/624db4c3-82c4-4546-90b8-606f72444e4f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.568058 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kwszx" event={"ID":"624db4c3-82c4-4546-90b8-606f72444e4f","Type":"ContainerDied","Data":"0b9718f1db1473d5bcca23261160bcb5cd5353c107f2fca2ad9974ecb0959309"} Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.568111 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kwszx" Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.568126 4846 scope.go:117] "RemoveContainer" containerID="7ebfefff6b59a0370fa09d13daeb29cca946bd85fc737ae8db2c0f1fb24cd6c0" Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.589084 4846 scope.go:117] "RemoveContainer" containerID="bec962f6dd18ab9ba405faba949ef07e1d87ca13cf23d63924fb0d874fe1a7e3" Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.618082 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kwszx"] Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.628512 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kwszx"] Feb 02 14:53:14 crc kubenswrapper[4846]: I0202 14:53:14.632729 4846 scope.go:117] "RemoveContainer" containerID="7d115629d4bccd0a8313818d34442c77f924cc4cf96b4d3d8391e184cfa07ec3" Feb 02 14:53:15 crc kubenswrapper[4846]: I0202 14:53:15.436249 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="624db4c3-82c4-4546-90b8-606f72444e4f" path="/var/lib/kubelet/pods/624db4c3-82c4-4546-90b8-606f72444e4f/volumes" Feb 02 14:53:30 crc kubenswrapper[4846]: I0202 14:53:30.478483 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:53:30 crc kubenswrapper[4846]: I0202 14:53:30.479884 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:54:00 crc kubenswrapper[4846]: I0202 14:54:00.478942 4846 patch_prober.go:28] interesting pod/machine-config-daemon-dhzp7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 14:54:00 crc kubenswrapper[4846]: I0202 14:54:00.479455 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 14:54:00 crc kubenswrapper[4846]: I0202 14:54:00.479498 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" Feb 02 14:54:00 crc kubenswrapper[4846]: I0202 14:54:00.480373 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f"} pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 14:54:00 crc kubenswrapper[4846]: I0202 14:54:00.480431 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerName="machine-config-daemon" containerID="cri-o://c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" gracePeriod=600 Feb 02 14:54:00 crc kubenswrapper[4846]: E0202 14:54:00.607912 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:54:01 crc kubenswrapper[4846]: I0202 14:54:01.153787 4846 generic.go:334] "Generic (PLEG): container finished" podID="05ed703a-a310-418e-8c5d-aeddb32977cf" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" exitCode=0 Feb 02 14:54:01 crc kubenswrapper[4846]: I0202 14:54:01.153839 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" event={"ID":"05ed703a-a310-418e-8c5d-aeddb32977cf","Type":"ContainerDied","Data":"c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f"} Feb 02 14:54:01 crc kubenswrapper[4846]: I0202 14:54:01.153875 4846 scope.go:117] "RemoveContainer" containerID="229e52826f90f0a1025ad49518481c25a85ee892227df94a2914ba07fbeeb68e" Feb 02 14:54:01 crc kubenswrapper[4846]: I0202 14:54:01.154339 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:54:01 crc kubenswrapper[4846]: E0202 14:54:01.154593 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:54:12 crc kubenswrapper[4846]: I0202 14:54:12.423385 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:54:12 crc kubenswrapper[4846]: E0202 14:54:12.424434 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:54:27 crc kubenswrapper[4846]: I0202 14:54:27.435198 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:54:27 crc kubenswrapper[4846]: E0202 14:54:27.436365 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:54:40 crc kubenswrapper[4846]: I0202 14:54:40.424484 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:54:40 crc kubenswrapper[4846]: E0202 14:54:40.425530 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:54:44 crc kubenswrapper[4846]: I0202 14:54:44.596199 4846 generic.go:334] "Generic (PLEG): container finished" podID="bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" containerID="5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24" exitCode=0 Feb 02 14:54:44 crc kubenswrapper[4846]: I0202 14:54:44.596250 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bbhbc/must-gather-ps8nn" event={"ID":"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a","Type":"ContainerDied","Data":"5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24"} Feb 02 14:54:44 crc kubenswrapper[4846]: I0202 14:54:44.597311 4846 scope.go:117] "RemoveContainer" containerID="5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24" Feb 02 14:54:45 crc kubenswrapper[4846]: I0202 14:54:45.019162 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bbhbc_must-gather-ps8nn_bf24498c-a1a4-44a0-9d40-3d0ca8b1941a/gather/0.log" Feb 02 14:54:52 crc kubenswrapper[4846]: I0202 14:54:52.423810 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:54:52 crc kubenswrapper[4846]: E0202 14:54:52.424551 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.111236 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bbhbc/must-gather-ps8nn"] Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.111995 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-bbhbc/must-gather-ps8nn" podUID="bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" containerName="copy" containerID="cri-o://568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc" gracePeriod=2 Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.123450 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bbhbc/must-gather-ps8nn"] Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.669049 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bbhbc_must-gather-ps8nn_bf24498c-a1a4-44a0-9d40-3d0ca8b1941a/copy/0.log" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.669945 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/must-gather-ps8nn" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.702415 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bbhbc_must-gather-ps8nn_bf24498c-a1a4-44a0-9d40-3d0ca8b1941a/copy/0.log" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.702905 4846 generic.go:334] "Generic (PLEG): container finished" podID="bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" containerID="568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc" exitCode=143 Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.702973 4846 scope.go:117] "RemoveContainer" containerID="568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.703094 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bbhbc/must-gather-ps8nn" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.737777 4846 scope.go:117] "RemoveContainer" containerID="5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.784562 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-must-gather-output\") pod \"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a\" (UID: \"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a\") " Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.784775 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8hl9\" (UniqueName: \"kubernetes.io/projected/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-kube-api-access-s8hl9\") pod \"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a\" (UID: \"bf24498c-a1a4-44a0-9d40-3d0ca8b1941a\") " Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.805477 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-kube-api-access-s8hl9" (OuterVolumeSpecName: "kube-api-access-s8hl9") pod "bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" (UID: "bf24498c-a1a4-44a0-9d40-3d0ca8b1941a"). InnerVolumeSpecName "kube-api-access-s8hl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.812706 4846 scope.go:117] "RemoveContainer" containerID="568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc" Feb 02 14:54:55 crc kubenswrapper[4846]: E0202 14:54:55.813092 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc\": container with ID starting with 568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc not found: ID does not exist" containerID="568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.813128 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc"} err="failed to get container status \"568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc\": rpc error: code = NotFound desc = could not find container \"568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc\": container with ID starting with 568ccf53cf611de71b9635065417a36dd648d8a246b25222296485924c3c46dc not found: ID does not exist" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.813149 4846 scope.go:117] "RemoveContainer" containerID="5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24" Feb 02 14:54:55 crc kubenswrapper[4846]: E0202 14:54:55.813444 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24\": container with ID starting with 5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24 not found: ID does not exist" containerID="5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.813499 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24"} err="failed to get container status \"5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24\": rpc error: code = NotFound desc = could not find container \"5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24\": container with ID starting with 5eb34420f9eb70bd986b28f401a489cabf055e838bc1633f21b09a3419652b24 not found: ID does not exist" Feb 02 14:54:55 crc kubenswrapper[4846]: I0202 14:54:55.888003 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8hl9\" (UniqueName: \"kubernetes.io/projected/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-kube-api-access-s8hl9\") on node \"crc\" DevicePath \"\"" Feb 02 14:54:56 crc kubenswrapper[4846]: I0202 14:54:56.036366 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" (UID: "bf24498c-a1a4-44a0-9d40-3d0ca8b1941a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:54:56 crc kubenswrapper[4846]: I0202 14:54:56.091292 4846 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.438930 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" path="/var/lib/kubelet/pods/bf24498c-a1a4-44a0-9d40-3d0ca8b1941a/volumes" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.517158 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-km5zp"] Feb 02 14:54:57 crc kubenswrapper[4846]: E0202 14:54:57.517910 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" containerName="gather" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.517934 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" containerName="gather" Feb 02 14:54:57 crc kubenswrapper[4846]: E0202 14:54:57.517963 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerName="registry-server" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.517974 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerName="registry-server" Feb 02 14:54:57 crc kubenswrapper[4846]: E0202 14:54:57.517991 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624db4c3-82c4-4546-90b8-606f72444e4f" containerName="extract-content" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.518001 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="624db4c3-82c4-4546-90b8-606f72444e4f" containerName="extract-content" Feb 02 14:54:57 crc kubenswrapper[4846]: E0202 14:54:57.518015 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624db4c3-82c4-4546-90b8-606f72444e4f" containerName="registry-server" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.518023 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="624db4c3-82c4-4546-90b8-606f72444e4f" containerName="registry-server" Feb 02 14:54:57 crc kubenswrapper[4846]: E0202 14:54:57.518037 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerName="extract-content" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.518046 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerName="extract-content" Feb 02 14:54:57 crc kubenswrapper[4846]: E0202 14:54:57.518076 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624db4c3-82c4-4546-90b8-606f72444e4f" containerName="extract-utilities" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.518085 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="624db4c3-82c4-4546-90b8-606f72444e4f" containerName="extract-utilities" Feb 02 14:54:57 crc kubenswrapper[4846]: E0202 14:54:57.518103 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerName="extract-utilities" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.518111 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerName="extract-utilities" Feb 02 14:54:57 crc kubenswrapper[4846]: E0202 14:54:57.518124 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" containerName="copy" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.518131 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" containerName="copy" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.518403 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" containerName="copy" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.518422 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf24498c-a1a4-44a0-9d40-3d0ca8b1941a" containerName="gather" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.518437 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd33bb9a-31dd-4d43-a4b5-7a2c84eaadb1" containerName="registry-server" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.518461 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="624db4c3-82c4-4546-90b8-606f72444e4f" containerName="registry-server" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.520796 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.532905 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-km5zp"] Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.648354 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-catalog-content\") pod \"redhat-marketplace-km5zp\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.648995 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9wzt\" (UniqueName: \"kubernetes.io/projected/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-kube-api-access-s9wzt\") pod \"redhat-marketplace-km5zp\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.649425 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-utilities\") pod \"redhat-marketplace-km5zp\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.751799 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-utilities\") pod \"redhat-marketplace-km5zp\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.751903 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-catalog-content\") pod \"redhat-marketplace-km5zp\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.751979 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9wzt\" (UniqueName: \"kubernetes.io/projected/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-kube-api-access-s9wzt\") pod \"redhat-marketplace-km5zp\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.752861 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-catalog-content\") pod \"redhat-marketplace-km5zp\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:57 crc kubenswrapper[4846]: I0202 14:54:57.753172 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-utilities\") pod \"redhat-marketplace-km5zp\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:58 crc kubenswrapper[4846]: I0202 14:54:58.403394 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9wzt\" (UniqueName: \"kubernetes.io/projected/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-kube-api-access-s9wzt\") pod \"redhat-marketplace-km5zp\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:58 crc kubenswrapper[4846]: I0202 14:54:58.456557 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:54:58 crc kubenswrapper[4846]: I0202 14:54:58.962551 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-km5zp"] Feb 02 14:54:59 crc kubenswrapper[4846]: I0202 14:54:59.781786 4846 generic.go:334] "Generic (PLEG): container finished" podID="e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff" containerID="d9df5a08fcba20f44c4a6696d7fdfa34639d80c4cc486bb707b48f32dc343cd3" exitCode=0 Feb 02 14:54:59 crc kubenswrapper[4846]: I0202 14:54:59.782745 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km5zp" event={"ID":"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff","Type":"ContainerDied","Data":"d9df5a08fcba20f44c4a6696d7fdfa34639d80c4cc486bb707b48f32dc343cd3"} Feb 02 14:54:59 crc kubenswrapper[4846]: I0202 14:54:59.782866 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km5zp" event={"ID":"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff","Type":"ContainerStarted","Data":"4c57aa62d7ee075d9118788088dd95d60fedb228aa7429ed9c48a315389ce71e"} Feb 02 14:55:01 crc kubenswrapper[4846]: I0202 14:55:01.804222 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km5zp" event={"ID":"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff","Type":"ContainerStarted","Data":"02e25c31fd5d968f51a9b87ad086f7fde4d9cd98e582d65547743453b0aeb86d"} Feb 02 14:55:02 crc kubenswrapper[4846]: I0202 14:55:02.816338 4846 generic.go:334] "Generic (PLEG): container finished" podID="e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff" containerID="02e25c31fd5d968f51a9b87ad086f7fde4d9cd98e582d65547743453b0aeb86d" exitCode=0 Feb 02 14:55:02 crc kubenswrapper[4846]: I0202 14:55:02.816384 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km5zp" event={"ID":"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff","Type":"ContainerDied","Data":"02e25c31fd5d968f51a9b87ad086f7fde4d9cd98e582d65547743453b0aeb86d"} Feb 02 14:55:03 crc kubenswrapper[4846]: I0202 14:55:03.827944 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km5zp" event={"ID":"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff","Type":"ContainerStarted","Data":"eeeadeb86aa169dfc0fe79625a4d4ab23284f761210cb97648bb9636419bd8fa"} Feb 02 14:55:03 crc kubenswrapper[4846]: I0202 14:55:03.851317 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-km5zp" podStartSLOduration=3.113677166 podStartE2EDuration="6.85129791s" podCreationTimestamp="2026-02-02 14:54:57 +0000 UTC" firstStartedPulling="2026-02-02 14:54:59.785283599 +0000 UTC m=+9931.013870462" lastFinishedPulling="2026-02-02 14:55:03.522904343 +0000 UTC m=+9934.751491206" observedRunningTime="2026-02-02 14:55:03.843221012 +0000 UTC m=+9935.071807895" watchObservedRunningTime="2026-02-02 14:55:03.85129791 +0000 UTC m=+9935.079884793" Feb 02 14:55:07 crc kubenswrapper[4846]: I0202 14:55:07.423309 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:55:07 crc kubenswrapper[4846]: E0202 14:55:07.424222 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:55:08 crc kubenswrapper[4846]: I0202 14:55:08.456993 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:55:08 crc kubenswrapper[4846]: I0202 14:55:08.457038 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:55:08 crc kubenswrapper[4846]: I0202 14:55:08.520335 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:55:08 crc kubenswrapper[4846]: I0202 14:55:08.930761 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:55:08 crc kubenswrapper[4846]: I0202 14:55:08.979991 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-km5zp"] Feb 02 14:55:10 crc kubenswrapper[4846]: I0202 14:55:10.970648 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-km5zp" podUID="e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff" containerName="registry-server" containerID="cri-o://eeeadeb86aa169dfc0fe79625a4d4ab23284f761210cb97648bb9636419bd8fa" gracePeriod=2 Feb 02 14:55:11 crc kubenswrapper[4846]: I0202 14:55:11.988307 4846 generic.go:334] "Generic (PLEG): container finished" podID="e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff" containerID="eeeadeb86aa169dfc0fe79625a4d4ab23284f761210cb97648bb9636419bd8fa" exitCode=0 Feb 02 14:55:11 crc kubenswrapper[4846]: I0202 14:55:11.989015 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km5zp" event={"ID":"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff","Type":"ContainerDied","Data":"eeeadeb86aa169dfc0fe79625a4d4ab23284f761210cb97648bb9636419bd8fa"} Feb 02 14:55:11 crc kubenswrapper[4846]: I0202 14:55:11.989045 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km5zp" event={"ID":"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff","Type":"ContainerDied","Data":"4c57aa62d7ee075d9118788088dd95d60fedb228aa7429ed9c48a315389ce71e"} Feb 02 14:55:11 crc kubenswrapper[4846]: I0202 14:55:11.989076 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c57aa62d7ee075d9118788088dd95d60fedb228aa7429ed9c48a315389ce71e" Feb 02 14:55:12 crc kubenswrapper[4846]: I0202 14:55:12.059929 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:55:12 crc kubenswrapper[4846]: I0202 14:55:12.164510 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-catalog-content\") pod \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " Feb 02 14:55:12 crc kubenswrapper[4846]: I0202 14:55:12.164785 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-utilities\") pod \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " Feb 02 14:55:12 crc kubenswrapper[4846]: I0202 14:55:12.164893 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9wzt\" (UniqueName: \"kubernetes.io/projected/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-kube-api-access-s9wzt\") pod \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\" (UID: \"e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff\") " Feb 02 14:55:12 crc kubenswrapper[4846]: I0202 14:55:12.167285 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-utilities" (OuterVolumeSpecName: "utilities") pod "e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff" (UID: "e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:55:12 crc kubenswrapper[4846]: I0202 14:55:12.178883 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-kube-api-access-s9wzt" (OuterVolumeSpecName: "kube-api-access-s9wzt") pod "e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff" (UID: "e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff"). InnerVolumeSpecName "kube-api-access-s9wzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 14:55:12 crc kubenswrapper[4846]: I0202 14:55:12.209078 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff" (UID: "e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 14:55:12 crc kubenswrapper[4846]: I0202 14:55:12.268306 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 14:55:12 crc kubenswrapper[4846]: I0202 14:55:12.268376 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 14:55:12 crc kubenswrapper[4846]: I0202 14:55:12.268393 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9wzt\" (UniqueName: \"kubernetes.io/projected/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff-kube-api-access-s9wzt\") on node \"crc\" DevicePath \"\"" Feb 02 14:55:13 crc kubenswrapper[4846]: I0202 14:55:13.001489 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-km5zp" Feb 02 14:55:13 crc kubenswrapper[4846]: I0202 14:55:13.049146 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-km5zp"] Feb 02 14:55:13 crc kubenswrapper[4846]: I0202 14:55:13.064582 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-km5zp"] Feb 02 14:55:13 crc kubenswrapper[4846]: I0202 14:55:13.438805 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff" path="/var/lib/kubelet/pods/e80f6d1f-d3d4-45f7-aefd-d1f9e66be7ff/volumes" Feb 02 14:55:22 crc kubenswrapper[4846]: I0202 14:55:22.423881 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:55:22 crc kubenswrapper[4846]: E0202 14:55:22.424803 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:55:37 crc kubenswrapper[4846]: I0202 14:55:37.423028 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:55:37 crc kubenswrapper[4846]: E0202 14:55:37.423756 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:55:48 crc kubenswrapper[4846]: I0202 14:55:48.423811 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:55:48 crc kubenswrapper[4846]: E0202 14:55:48.424742 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:56:00 crc kubenswrapper[4846]: I0202 14:56:00.424119 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:56:00 crc kubenswrapper[4846]: E0202 14:56:00.425388 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:56:12 crc kubenswrapper[4846]: I0202 14:56:12.424681 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:56:12 crc kubenswrapper[4846]: E0202 14:56:12.426010 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:56:27 crc kubenswrapper[4846]: I0202 14:56:27.423193 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:56:27 crc kubenswrapper[4846]: E0202 14:56:27.424030 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:56:38 crc kubenswrapper[4846]: I0202 14:56:38.424488 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:56:38 crc kubenswrapper[4846]: E0202 14:56:38.425206 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:56:49 crc kubenswrapper[4846]: I0202 14:56:49.435146 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:56:49 crc kubenswrapper[4846]: E0202 14:56:49.435857 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:57:03 crc kubenswrapper[4846]: I0202 14:57:03.424838 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:57:03 crc kubenswrapper[4846]: E0202 14:57:03.425717 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:57:15 crc kubenswrapper[4846]: I0202 14:57:15.423654 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:57:15 crc kubenswrapper[4846]: E0202 14:57:15.424465 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:57:30 crc kubenswrapper[4846]: I0202 14:57:30.424119 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:57:30 crc kubenswrapper[4846]: E0202 14:57:30.425401 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:57:44 crc kubenswrapper[4846]: I0202 14:57:44.423946 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:57:44 crc kubenswrapper[4846]: E0202 14:57:44.424822 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:57:55 crc kubenswrapper[4846]: I0202 14:57:55.424122 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:57:55 crc kubenswrapper[4846]: E0202 14:57:55.424979 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" Feb 02 14:58:10 crc kubenswrapper[4846]: I0202 14:58:10.424252 4846 scope.go:117] "RemoveContainer" containerID="c722defe61285f3e3b42810ef5831da450389f1e3d489d2c47b02455051d1c0f" Feb 02 14:58:10 crc kubenswrapper[4846]: E0202 14:58:10.425023 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dhzp7_openshift-machine-config-operator(05ed703a-a310-418e-8c5d-aeddb32977cf)\"" pod="openshift-machine-config-operator/machine-config-daemon-dhzp7" podUID="05ed703a-a310-418e-8c5d-aeddb32977cf" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515140135627024451 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015140135627017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015140111446016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015140111447015453 5ustar corecore